Privacy Overview
This website uses cookies to improve your experience while you navigate through the website. Out of these, the cookies that are categorized as necessary are stored on your browser as they are essential for the working of basic functionalities of the website. We also use third-party cookies that help us analyze and understand how you use this website. These cookies will be stored in your browser only with your consent. You also have the option to opt-out of these cookies. But opting out of some of these cookies may affect your browsing experience.
Always Active
Necessary cookies are absolutely essential for the website to function properly. These cookies ensure basic functionalities and security features of the website, anonymously.
Functional cookies help to perform certain functionalities like sharing the content of the website on social media platforms, collect feedbacks, and other third-party features.

No cookies to display.

Performance cookies are used to understand and analyze the key performance indexes of the website which helps in delivering a better user experience for the visitors.

No cookies to display.

Analytical cookies are used to understand how visitors interact with the website. These cookies help provide information on metrics the number of visitors, bounce rate, traffic source, etc.

No cookies to display.

Advertisement cookies are used to provide visitors with relevant ads and marketing campaigns. These cookies track visitors across websites and collect information to provide customized ads.

No cookies to display.

Other uncategorized cookies are those that are being analyzed and have not been classified into a category as yet.

No cookies to display.

How Speech Recognition Dictation Technology Affects Non-Native English Speakers

To investigate the interaction between speech recognition dictation technology and the use of language by non-native speakers of English, several language minority students from diverse ethnolinguistic backgrounds were asked to use speech recognition dictation software to orally compose a series of college level writing tasks. The interactions were videotaped and transcribed; the transcriptions include the participants’ verbal interactions, their non-verbal actions, and the actions of the computer.

The analysis of the transcripts draws on Halliday’s conceptual framework of register (field, tenor, and mode), emphasizing the relationship between the forms and the features of the context. Collectively, these three components serve to predict the register used with speech recognition dictation technology and to establish a general principle governing the ways in which the context of production was projected onto the text.

The findings indicate the manner in which the technology influenced the development of text and patterns of interaction. Many of the participants adopted a strategy of perfecting the composition as much as possible during the speech sessions. This strategy included making changes in the surface features of their talk in order to accommodate the software. Speech adjustments, associated with what I call “speech recognition talk,” included changes in lexical choice, utterance length, and pronunciation. In addition to these modifications in speech, many of the participants gained more control over oral phenomena that did not translate well into writing, such as filled pauses, false starts, meta-comments, and imprecise wording of oral punctuation commands. As part of the two-way interaction, the software improved in its recognition rate as these same participants updated their voice templates during the editing sessions.

The conclusions of this study offer a better understanding as to the influences of the situational features associated with speech recognition technology on spoken and written English. The recommendations include the expansion of possibilities associated with the transfer of modes offered by this technology, as well as insights into the advantages and constraints that the technology imposes on communication for language minority users.

Author Bio

Elizabeth Meddeb is Assistant Professor of ESL and Humanities at York College/CUNY. She is also coordinating the ESL program. She teaches advanced composition courses to non-native speakers of English, as well as introductory courses in linguistics and humanities. Her research interests include the interaction between technology and language use. She is currently working on a research grant that investigates how speech recognition dictation technology shapes both spoken and written language use for non-native speakers of English. This study is an outgrowth of her dissertation research at Columbia University and her work experience at IBM's TJ Watson Research Center.