Please use this identifier to cite or link to this item:
https://doi.org/10.1145/2388676.2388793
DC Field | Value | |
---|---|---|
dc.title | Speak-As-You-Swipe (SAYS): A multimodal interface combining speech and gesture keyboard synchronously for continuous mobile text entry | |
dc.contributor.author | Sim, K.C. | |
dc.date.accessioned | 2013-07-04T08:41:36Z | |
dc.date.available | 2013-07-04T08:41:36Z | |
dc.date.issued | 2012 | |
dc.identifier.citation | Sim, K.C. (2012). Speak-As-You-Swipe (SAYS): A multimodal interface combining speech and gesture keyboard synchronously for continuous mobile text entry. ICMI'12 - Proceedings of the ACM International Conference on Multimodal Interaction : 555-560. ScholarBank@NUS Repository. <a href="https://doi.org/10.1145/2388676.2388793" target="_blank">https://doi.org/10.1145/2388676.2388793</a> | |
dc.identifier.isbn | 9781450314671 | |
dc.identifier.uri | http://scholarbank.nus.edu.sg/handle/10635/42028 | |
dc.description.abstract | Modern mobile devices, such as the smartphones and tablets, are becoming increasingly popular amongst users of all ages. Text entry is one of the most important modes of interaction between human and their mobile devices. Although typing on a touchscreen display using a soft keyboard remains the most common text input method for many users, the process can be frustratingly slow, especially on smartphones with a much smaller screen. Voice input offers an attractive alternative that completely eliminates the need for typing. However, voice input relies on automatic speech recognition technology whose performance degrades significantly in noisy environment or for non-native users. This paper presents Speak-As-You-Swipe (SAYS), a novel multimodal interface that enables efficient continuous text entry on mobile devices. SAYS integrates a gesture keyboard with speech recognition to improve the efficiency and accuracy of text entry. The swipe gesture and voice inputs provide complementary information that can be very effective in disambiguating confusions in word predictions. The word prediction hypotheses from a gesture keyboard are directly incorporated into the speech recognition process so that the SAYS interface can handle continuous input. Experimental results show that for a 20k vocabulary, the proposed SAYS interface can achieve prediction accuracy of 96.4% in clean condition and about 94.0% in noisy environment, compared to 92.2% using a gesture keyboard alone. Copyright 2012 ACM. | |
dc.description.uri | http://libproxy1.nus.edu.sg/login?url=http://dx.doi.org/10.1145/2388676.2388793 | |
dc.source | Scopus | |
dc.subject | Gesture keyboard | |
dc.subject | Mobile text input | |
dc.subject | Multimodal interface | |
dc.subject | Voice input | |
dc.type | Conference Paper | |
dc.contributor.department | COMPUTER SCIENCE | |
dc.description.doi | 10.1145/2388676.2388793 | |
dc.description.sourcetitle | ICMI'12 - Proceedings of the ACM International Conference on Multimodal Interaction | |
dc.description.page | 555-560 | |
dc.identifier.isiut | NOT_IN_WOS | |
Appears in Collections: | Staff Publications |
Show simple item record
Files in This Item:
There are no files associated with this item.
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.