Speech Recognition Market S-89.3680 Speech Recognition Seminar First name Family name Date Guzman Alvarez 18/12/2014 Speech Recognition Market 1. Introduction The aim of this work is to give an overview of what the status of speech recognition is from the commercial point of view, and try to follow the events that have driven its commercial development throughout the years. In the first place, the progression of the market of speech recognition systems is reviewed by looking at its main milestones: from the first commercial system to the revolution of the cloud. The next section gives a brief overview of the current status of the market based on financial data of some of the main players (Nuance, Microsoft, Google, Apple) as well as on other source’s analysis and prediction of its future. Finally, how those companies tackle and use speech recognition studied. In addition, a brief Q&A with the Finnish company Lingsoft is included, as an interesting example of “small” but successful company that is involved in speech recognition and closely related to Aalto University. 2 Speech Recognition Market 2. History 2.1. Early stages and first commercial ASR product Research activity on speech recognition started in the decade of 1950, with developments by Bell Laboratories based on previous advances on the field on phonetics, such as the VODER; in 1952 they had a system that could recognize single digits from a single speaker. Other advances followed, from institutions such as the Massachusets Institute of Technology, RCA Laboratories or Kyoto University, who designed systems that recognized from basic units of speech up to few whole words, and that already applied the concepts of non-uniform time scale and statistical syntax [1]. The first commercial automatic speech recognition (ASR) product was the VIP-100 by Threshold Technology Inc., first sold in 1972 [2]. It consisted of a hardware preprocessor, a feature extractor and a classifier function performed by dedicated hardware running in a “minicomputer”. The user chose the vocabulary and trained the system with 10 utterances of each word, which the recognizer compared to the input in order to determine which word was being spoken. The system allowed for rates of up to 125 words per minute, and achieved accuracies of 98.5 % for a 35-word vocabulary in a noisy environment, or 99.5 % for a 15-word vocabulary in a quiet environment. [2] 2.2. DARPA Although only around 100 units of the VIP-100 were sold (mainly for industrial activities such as material handling and quality control) [2], it had a major role in the history of speech recognition, since it influenced the Defense Advanced Research Projects Agency (DARPA) to fund the Speech Understanding Research (SUR) project in order to develop speech recognition technology with military purposes. Its goals, taken from [3], were: “Accept connected speech, from many cooperative speakers, in a quiet room, using a good microphone, with slight tuning/speaker, accepting 1000 words, using an artificial syntax, in a constraining task, yielding < 10% semantic error, in a few times real time, on a 100 MIPS machine.” Among the results of the SUR program was a recognizer by the Carnegie Mellon University (CMU), called Harpy, which could recognize more than 1000 words and was the first to use the concepts of graph search and finite-state networks [4]. Also from the CMU was the Hearsay-II system, which performed several analysis in parallel over the data in order to obtain different types of evidence, that were combined to obtain a hypothesis. This idea, which consisted effectively on using a blackboard system for speech recognition, was novel and unique in the field. [1] Bolt Beranek and Newman (BBN) created the Hear What I Mean (HWIM) system, which introduced word verification tests (at the parametric level), the concept of lexical decoding networks, and a lattice of possible hypotheses that reduced segmentation ambiguity [3]. 3 Speech Recognition Market Another system worth mentioning, created in 1975 by James and Janet Baker, is DRAGON [5], which will be discussed in a subsequent section. 2.3. Research in IBM and AT&T Bell Laboratories Besides the activities funded by DARPA, IBM and AT&T invested heavily in speech recognition research, each company following a different approach. AT&T wanted a system that would allow carrying out certain tasks such as routing calls or dialing automatically by asking the user to speak the parameters; this implied recognizing a small set of mostly single words, but at the same time a speaker-independent and robust system was needed, since it would get input from people from across the whole country. It led to the development of templates and statistical models for sound reference patterns, and to the concept of spectral distance measures. They also introduced keyword spotting, in order to detect key utterances irrespective to the preceding or following speech, which allowed to detect answers spoken in natural language: for example, to the question “How old are you?” the keywords would be numbers, and the system would recognize both “twenty-four” or “I am twenty-four years old” as valid answers. [4] On the contrary, the system from IBM was aimed at being an alternative form of general text input to other systems, a “voice typewriter”. Thus, the target was to have a large vocabulary and make as few mistakes as possible. This led to the development of a system called “Tangora”, that used the concepts of n-grams, language models and found the maximum-likelihood sentence for the input speech. This system needed a training period for each new speaker in order to adapt the models to his/her way of speaking. [3] Neither of these efforts resulted in commercial successes, but settled the grounds for further research on speech recognition. 2.4. DRAGON, Dragon Systems, NaturallySpeaking This section summarizes the history of DRAGON and Dragon Systems, and has been sourced mostly from [6] unless noted otherwise. As mentioned before, one of the notable systems born under DARPA’s SUR framework was DRAGON. It was conceived by Janet and James Baker, a couple of doctoral students at the CMU during the time the SUR program was being carried on. They pioneered the application of the Hidden Markov Models to speech recognition [6], but their description of the DRAGON’s main features did not stress that point, reading instead: “delayed decisions, generative form of model, hierarchical system, integrated representation, general theoretical framework” [5]. After graduating from the CMU in 1975 they joined IBM’s research center, where they developed a system with a vocabulary of 1000 words that took one hour to decode a single sentence of continuous-speech. However, IBM did not favor getting simpler-but-usable products to the market, which the Bakers saw as a more positive approach. For this reason, they started working for Verbex in 1979, company that was selling a recognizer of digits over phone connections; their aim was to improve the system until continuous speech recognition was possible. 4 Speech Recognition Market In 1982, when Verbex stopped their speech recognition activities, James and Janet founded their own company, called “Dragon Systems”. During the first decade they produced several successful custom discrete-recognition products, including the recognizer used by the world’s first personal computers that let users execute commands via voice instructions, made by Apricot Computers PLC [7]. In 1990 they released DragonDictate, which allowed controlling any computer and dictating text, and was successful among the disabled as an alternative to the keyboard. Still, the general market did not adopt speech recognition solutions, as all of them were discrete (required a pause between words) and thus not comfortable for natural speech. In 1994, the Bakers predicted that the computing capabilities of general-purpose PCs would allow continuous speech recognition within a few years, and restructured their company in order to be ready when the time came. In fact, when in 1997 they released the first version of NaturallySpeaking, it became an immediate success. Although IBM counterattacked some months later, selling their “ViaVoice” product at a lower price, which produced similar results, Dragon System’s technology has since then dominated the market until now. However, in the year 2000 the company was acquired by the corporation Lernout & Hauspie, which went bankrupt after just a few months [8]. The rights of NaturallySpeaking and its speech recognition technology were put on sale, and the Bakers could not compete with Scansoft (later Nuance Communications) in order to buy them back. The Bakers not only lost a handful of money, but the technology they had been working on for more than 25 years, stating that “part of them died” after the sale [9]. James Baker is today a Distinguished Professor at the CMU [10], and most likely an Apple employee (see section 3.5). 2.5. Scansoft, Nuance, Nuance Communications The current leader in the speech recognition market, Nuance Communications, was born in 2005 from the merge of Scansoft and Nuance [11]. In this section, the history of both former companies and that of the resulting corporation is briefly reviewed. 2.5.1. Scansoft Scansoft had its origins in Kurzweil Computer Products Inc., founded in 1974, which specialized in Optical Character Recognition (OCR) products. In 1980, Kurzweil was bought by Xerox Inc., which ran it as Xerox Imaging Solutions (XIS), and later as Scansoft. In 1999, Visioneer, a company that developed scanner hardware and software, bought Scansoft from Xerox, but continued to use the same commercial name. In 2001 Scansoft made a key acquisition, buying the speech technology assets of Lernout and Hauspie [12], after a financial turmoil described in the previous section. A major part of these assets were the former technology from Dragon Systems, although Lernout and Hauspie had previously acquired other companies involved in speech as well. This operation was considered a “bargain”, taking into account that while Dragon Systems was sold to L&H for $580 M, Scansoft bought the technology for a total price of approximately $40 M [12]. 5 Speech Recognition Market The acquisition meant that Scansoft moved into the speech recognition market, and started competing with Nuance. In 2003, Scansoft bought SpeechWorks, a direct competitor in the field of speech recognition, in a $132 M-deal [13], becoming a global player in the market. During the rest of 2003, 2004 and the beginning of 2005 Scansoft continued to acquire smaller speech-related companies, which contributed to increase their market share and product portfolio. 2.5.2. Nuance Nuance was born in 1994 as a spin-off of the Speech Technology and Research (STAR) laboratory of the Stanford Research Institute (SRI), with the purpose of commercializing the solutions that the institute had developed for the US Government [14]. The first commercial product was released in 1996, which consisted in a call center automation solution based on speech recognition. Through the following years they became the global leader in the market of speech recognition solutions, selling mainly a platform on top of which other vendors could build applications. This platform featured Speaker-Independent Natural Language Speech Recognition (SI-NLSR), which at the time was a key advantage over their competitors [15]. The difference between Nuance and Dragon Systems was that while the former sold a speaker-independent recognizer that other companies could use in their products, generally involving telephone network operators, the latter sold end solutions, either personalized for a specific customer or to the general consumer market. Nuance only acquired the company SpeechFront in the year 2000 [15], continuing to compete with Philips and SpeechWorks [16] [17]. 2.5.3. Nuance Communications As stated before, Nuance Communications was born in 2005, when Scansoft acquired Nuance, and started to operate under the new name, which was perhaps more adequate given the fact that after Scansoft’s latest acquisitions their focus was clearly on the speech recognition market. Nuance Communications started from a position of leadership both in the consumer market, selling mainly the NaturallySpeaking product and in the enterprise market, having the portfolio of all the previously acquired companies. During the following years, until today, the strategy of Nuance has been to control the speech recognition market and expand their range of products by taking over potential competitors and startups. In Section 3.2, an overview of the products that Nuance offers currently is given. 2.6. Siri Although the word “Siri” is nowadays associated with the digital personal assistant application present in Apple smartphones, it was initially a company founded in 2007 by Adam Cheyer and Dag Kittlausas as a spin-off from the Stanford Research Institute [18]. During the previous years, the SRI had participated in an artificial intelligence (AI) research project called “Cognitive Assistant that Learns and Organizes” (CALO); this project was in turn part of the “Personalized 6 Speech Recognition Market Assistant that Learns” (PAL) program funded by the DARPA jointly with the Swiss institute of technology (EPFL), which was the largest (known) AI project in the history of the U.S. [19]. After the foundation, Siri spent around three years developing their product until February 2010, when they released the “Siri” application in Apple’s App Store, available to the iPhone smartphones [20]. The groundbreaking feature was the ability to understand natural language, adapting to that of each user, as well as integration with other services that allowed returning a non-generic response; for example, if a user asked about the weather, instead of recognizing the question and querying a search provider, the application would connect to a weather services provider and display the information directly. Specialized publications recognized the breakthrough character of the application (for example, in [21], the author states “this is the future of the web”), and so did Apple Inc., who purchased the company a few months after the initial release (the operation was made public in April 2010 [22]). As a result, Siri stopped developing the initial application, canceling their plans to make it available in other platforms, and worked with Apple to integrate the functionality into the iPhone’s operating system. In October 2011 Siri was released as a feature of the iPhone 4S, being the protagonist during the presentation of the phone to the public. Since then the interest spread globally, and it became clear for Apple’s competitors that their products needed similar features. Siri as a feature of the iPhone was a clear breakthrough in speech recognition, not because of the technology, which was developed years before, but because it started the interest of the general public in the field, which in turn resulted in bigger research efforts and investments. As an example, a quote from [23]: “Todd Mozer, the CEO of a company that works with device manufacturers to create voice interfaces called Sensory, remembers the attitude toward speech interfaces seeming to change overnight. "You wouldn’t believe the number of companies that suddenly were knocking on our doors and saying, ‘we want to add speech recognition, we don’t exactly know how, but we know we need to do it now,’" he says”. 2.7. Cloud Although the concept behind the term “cloud computing” started to form in the 1950s, and it applies to general computing, it meant a true breakthrough for speech recognition when it became a feasible and known enabler. From this point of view the former Siri application was a pioneer in using the cloud to perform speech recognition. The main idea behind cloud computing is to run tasks in servers whose physical location is unknown, using the Internet to send the input from a client towards these servers (the cloud) and to send the result of the task back to the client. In the case of speech recognition this is particularly useful, since the use of advanced methods and models requires heavy computation. In the past, a compromise solution needed to be taken in order to balance the available processing power with the desirable recognizer quality. Using cloud computing, the end user needs only a device capable of recording his/her voice and sending 7 Speech Recognition Market fragments to the servers of the speech recognition provider. The recognition is run in powerful machines that can deliver results almost in real time, and the resulting text is sent back to the user device, which can use it appropriately. This effectively enables any device that can connect to the Internet to recognize speech using state-of-the-art systems, eliminating the processing power limitation. On top of this, the use of cloud computing allows for small developers to implement speech recognition in their applications without having to develop the recognizer themselves. For example, Nuance offers the Dragon Mobile software development kit (SDK), which allows mobile applications to use Nuance’s speech recognition engine, using different pricing schemes [24]. Giving access to these technologies to every developer who wants to use them will promote new ways of taking advantage of speech recognition, which in turn will drive the market and promote faster research and development in the field. In addition, according to ABI Research (see [25]), these solutions are a considerable factor influencing the growth that mobile speech recognition is currently experiencing. 8 Speech Recognition Market 3. Current status This section provides an overview of the current situation of the speech recognition market. First, its approximate size and distribution are studied, as well as the financial evolution of four of the most active companies in the field. Then, the approach of these companies towards speech recognition is very briefly looked at, mainly by means of their current lines of products that include speech recognition. Last, the contents of a “question and answer” session with a manager of Lingsoft Oy (a language management Finnish company) are reproduced, which give a good insight in the role of speech recognition in the company. 3.1. Market size, evolution According to BCC Research [26], the voice recognition market had a size of $47 billion in 2011, $53 billion in 2012, and will reach $113 billion in 2017, which means a total compound annual growth rate of 16.2 %. The distribution between consumer, enterprise and health care sections is shown in Figure 1; it is clear that the consumer market will grow the fastest, whereas no significant changes are predicted for the medical field. This could be reflecting the increasing interest from the end users who want to use voice-enabled applications in their smart devices, which promotes the development of such programs both for the consumer and the enterprise markets. Figure 1. Distribution of the speech recognition market in 2011, 2012 and 2017 (prediction) [26] Some of the most important companies that either sell or use speech-recognition related services or devices are at the moment Nuance, Microsoft, Google and Apple. The approach of each of these towards speech recognition is summarized in the following sections; however, in order to give a general overview of the market status from a different point of view, the evolution of the shares price of these four companies during the last 10 years has been collected in Figure 2. 9 Speech Recognition Market Figure 2 Price evolution of the shares of Nuance, Apple, Microsoft and Google in the last 10 years [Combined data from Yahoo Finance] Rather than studying the details of each graph, the conclusion to be drawn from the general picture is that all of the companies have a history of growth, and it is safe to predict that the same trend will continue, at least in the upcoming years. Since these are the companies that drive the largest part of the speech recognition market, it can be expected that it will continue growing in the following years, which agrees with the previously mentioned prediction by BCC Research [26]. 3.2. Nuance Communications Nuance is the global leader in the speech recognition market; this is a widely known and accepted fact, but no direct sources have been found (a statement from the own company can be read from some of their job offers, such as [27]). This leadership position was already present in the roots of the company, with Scansoft and Nuance being leaders in their respective markets prior to the merge, and has been maintained through the years. Nuance uses an aggressive competing strategy, which consists on acquiring well-established companies that are already competitors, start-ups that might be in the future, or companies who are already leaders in different markets they plan to get into. In this way, they can both control the competition and continuously extend their product offering. A detailed list of the companies acquired by Nuance Communications can be found in [15]. The majority of Nuance’s benefits come from the enterprise market, in which they sell or license their recognition engines to other companies, and from the health care market, where they offer ready-made solutions that can be used directly by the institutions. In the consumer market they sell the Dragon recognizer for personal computers, a personal assistant called “Dragon Asssistant” (Sirilike) for personal computers, and other products not related to speech recognition [28]. 10 Speech Recognition Market 3.3. Google Google addresses speech recognition as a philanthropic issue, where their goals are the following [29]: − To make speaking to your phones and computers ubiquitous and seamless. − To help make videos on the web accessible and searchable. Furthermore, also in [29] it is stated that users diversity is a “huge commitment”, imagining an ideal situation where every person in the world can use a speech recognizer. Quoting the mentioned text: “The payoff is immense: imagine making every lecture on the web accessible to every language; this is the kind of impact we are striving for”. Google has a unique position that allows for collection of immense amount of data, being the global search provider leader, providing the default speech recognition engine for Android phones, and owning the leading video-hosting website “Youtube”. Their research is consequently heavily focused on finding optimal ways of treating the data, from the points of view of both efficiency and output (improved models and methods). They don’t sell to the general public any products that involve speech recognition. Rather, it is included in the Android operating system and the Chrome browser. 3.4. Microsoft Even though not as actively as specialized companies, Microsoft has traditionally conducted heavy research in speech recognition [30]. This continues today, when they have an extensive team of researchers exclusively dedicated to speech recognition, working in a variety of topics that can be checked from [31]. They have made possible speech recognition in their main product (the Windows operating System) since 1995, when the Microsoft Speech Application Programming Interface (SAPI) was released. Apart from the operating system, they set an important milestone being the first company to use deep neural networks (DNN) in a commercial product, the Microsoft Audio Video Indexing Service. Nowadays, their mainstream products using speech recognition are the Windows operating system, Cortana (the virtual assistant present in smartphones with Windows Phone), and Skype, which with the new service “Skype Translate” is setting a new milestone, allowing to recognize, translate, and synthetize the words of one party in the language of the other, and vice-versa, in almost real time. 3.5. Apple Given the secrecy Apple maintains about their company and processes, it is not possible to be certain about how involved they are in speech recognition. Some sources maintain that they made an important investment prior to including it in their operating system (Mac OS, later OSX) for the first time, in 1993 [32]; however, the lack of popularity of speech recognition in personal computers 11 Speech Recognition Market during the last decade, together with the fact that they used a recognizer provided by Nuance for Siri [33], suggest that at least in these last years they have not been very active in the field. Nowadays the situation might be changing, most likely due to the increased interest of the customers in voice interfaces; besides many of the customary Apple-related “rumors” about Apple hiring engineers from Nuance, in the end of 2013 they acquired Novauris [34], a company founded by the James Baker (co-founder of Dragon Systems, see section 2.4) in 2001 [35]. 3.6. Lingsoft Lingsoft Oy is a relatively small Finnish company specialized in language management, which uses speech recognition mainly internally (as a part of processes that they sell to other companies) or as a part of a “wider” product. In order to get an idea of how speech recognition is managed at a smaller scale, a set of questions were asked to Dr. Vesa Siivola, who is a Development Manager at Lingsoft, which have been reproduced below: − Strategy, role of speech recognition in Lingsoft + “Our strategy is to help the customer improve their language processes. Sometimes, a speech recognizer will help to improve some part of the process and thus ends up being a part of the solution that is offered” − Common products or customized projects? + “All of our current projects involve customization for the client” − Impact of speech recognition + “Most speech recognition clients that we have are using grammar (command) speech recognition at phone portals (eg., Tallink, Palvelualojen Ammattiliitto, Lindorff). Continuous speech recognition is currently used mostly by our internal language services (subtitling of Finnish programs for YLE, transcription of medical patient dictations)”. − In which fields do you think the demand for improving speech recognition is at the moment? + “Since our business revolves around processes, we tend to see cases where there is a high volume of similar speech transcription tasks and improving the efficiency of the whole process is the key. Thus, it tends to be big organizations like YLE and hospitals. We have not concentrated on consumer space”. − Any special considerations applying speech recognition to the medical field? + “For the basic speech recognition engine, it does not differ terribly from any other customization case. The protection and laws regarding patient medical records pose extra challenges for getting the corpora required for model customization, though” − Competitors in the field of speech recognition? + “Nuance” 12 Speech Recognition Market 4. Conclusion It can be observed that during roughly 40 years of history of speech recognition, there has been a clear shift in the interests that drive the research. In the beginning the interest was purely academic; then the governments realized the potential (for example, DARPA and the SUR program); and at the same time private companies started to get involved. However, until the beginning of 90’s speech recognition was mostly a research topic, rather than a commercial product. Afterwards a slow move into the consumer market started to happen, moved by software solutions for personal computers, and it “exploded” in 2011 with the introduction of an actually usable voice interface in a phone. Nowadays the end consumers constitute the largest part of the speech recognition market. From the point of view of the author, what has been seen until now in terms of voice interaction with machines is just the tip of the iceberg: using their voice is the natural way most humans communicate, thus it is logical that as more flexible and reliable systems are developed a larger amount of people will start to use them; that will lead to more development efforts, and in this way an always-faster cycle will be created which will bring great advances and even possibly push the current keyboards into the background. 13 Speech Recognition Market 5. Bibliography [1] Fang Chen and Kristiina Jokinen, Speech Technology, Theory and Applications. New York, USA: Springer, 2010. [2] M. B. Herscher and R. B. Cox, "Source data entry using voice input," in IEEE International Conference on Acoustics, Speech, and Signal Processing, 1976, pp. 190-193. [3] Jacob Benesty, M. Mohan Sondhi, and Yiteng Arden Huang, Springer handbook of speech processing.: Springer, 2008. [4] B. H. Juang and Lawrence R Rabiner, "Automatic speech recognition–a brief history of the technology development," Georgia Institute of Technology. Atlanta Rutgers University and the University of California. Santa Barbara, 2005. [5] James K. Baker, "The DRAGON System - An Overview," in IEEE Transactions on Acoustics, Speech, and Signal Processing, vol. 23, 1975, pp. 24-29. [6] Simson L. Garfinkel, "Enter the dragon," MIT's Technology Review, vol. 101, pp. 58-64, 1998. [7] Jonathan Marsters. (2004) http://actapricot.org/history/apricot_history.html Apricot History. [Online]. [8] The New York Times. (2001, May) Dragon Systems Sputters After Belgian Suitor Fails. [Online]. http://www.nytimes.com/2001/05/07/business/dragon-systems-sputters-after-belgiansuitor-fails.html?pagewanted=all [9] Bloomberg. (2012, December) Dragon Systems Founder Says ‘Part of Me Died’ After Sale. [Online]. http://www.bloomberg.com/news/2012-12-12/dragon-systems-founder-says-part-ofme-died-at-trial.html [10] Carnegie Mellon University. Carnegie Mellon http://www.cs.cmu.edu/directory/James-Baker University - Directory. [Online]. [11] Nuance Communications. (2005, May) ScanSoft and Nuance to Merge, Creating Comprehensive Portfolio of Enterprise Speech Solutions and Expertise. [Online]. http://www.nuance.com/news/pressreleases/2005/20050509_nuance.asp [12] The New York Times. (2001, November) Buyers of Units Of Lernout Are Disclosed. [Online]. http://www.nytimes.com/2001/11/29/business/buyers-of-units-of-lernout-are-disclosed.html [13] Nuance Communications. (2003, April) ScanSoft and SpeechWorks to Merge. [Online]. http://www.nuance.com/news/pressreleases/2003/20030424_speechworks.asp 14 Speech Recognition Market [14] Standord Research Institute. Speech Technology and Research (STAR) Laboratory. [Online]. http://www-speech.sri.com [15] Wikipedia. Nuance Communications. http://en.wikipedia.org/wiki/Nuance_Communications [Online]. [16] Talking Technology International, Inc. (2002) CTADE Computer Telephony Application Development Environment. [Online]. http://www.tti.net/computer-telephony-toolkits/CTADE-Computer-Telephony-Toolkit.html [17] Nuance Communications. (2008, October) Nuance Acquires Philips Speech Recognition Systems. [Online]. http://www.nuance.com/news/pressreleases/2008/20081001_philips.asp [18] MIT Technology Review. (2009, February) TR10: Intelligent Software Assistant. [Online]. http://www.technologyreview.com/video/412282/tr10-intelligent-software-assistant/ [19] SRI International. (2007) Siri. [Online]. http://www.sri.com/work/timelineinnovation/timeline.php?timeline=computing-digital#!&innovation=siri [20] Siri. (2010, February) Siri is Born. https://web.archive.org/web/20100617003733/http://siri.com/blog/siri_is_born [Online]. [21] Robert Scoble. (2010, February) Why if you miss Siri you'll miss the future of the web. [Online]. http://scobleizer.com/?p=6299 [22] Business Insider. (2010, April) Apple Buys Siri, A Mobile Assistant App. [Online]. http://www.businessinsider.com/apple-buys-siri-a-mobile-assistant-app-as-war-with-googleheats-up-2010-4?IR=T [23] Fast Company. (2014, June) Why Siri is Apple's most important footnote. [Online]. http://www.fastcompany.com/3031335/most-innovative-companies/why-siri-is-apples-mostimportant-footnote [24] Nuance Communications. Integrate Speech Technology in your App with our SDK! [Online]. http://dragonmobile.nuancemobiledeveloper.com/public/index.php?task=home [25] ABI Research. (2012, August) Cloud Based Solutions Provide Impetus for Mobile Speech Recognition. [Online]. https://www.abiresearch.com/press/cloud-based-solutions-provideimpetus-for-mobile-s [26] BCC Research. Global Voice Recognition Market To Reach $113 Billion In 2017. [Online]. http://www.bccresearch.com/pressroom/ift/global-voice-recognition-market-reach-$113billion-2017 [27] Nuance Communications. Our Next BIG Innovation Begins with YOU. [Online]. 15 Speech Recognition Market http://www.nuance.com/company/careers/natural-language/index.htm [28] Nuance Communications. Nuance. [Online]. http://www.nuance.com/ [29] Google. Speech Processing. [Online]. http://research.google.com/pubs/SpeechProcessing.html [30] The Seattle Times. (1997, June) Talking Face To Interface. [Online]. http://community.seattletimes.nwsource.com/archive/?date=19970601&slug=2542162 [31] Microsoft. Speech Recognition, Synthesis, and Dialog Systems. [Online]. http://research.microsoft.com/en-us/about/our-research/speech-recognition-synthesis.aspx [32] TidBITS. (2000, August) http://tidbits.com/?tbart=06090 Bossing Your Mac with PlainTalk. [Online]. [33] Juli Clover. (2013, May) Nuance CEO Confirms Siri Partnership with Apple. [Online]. http://www.macrumors.com/2013/05/30/nuance-ceo-confirms-siri-partnership-with-apple/ [34] The Wall Street Journal. (2014, April) Apple Acquires Speech-Recognition Specialist Novauris. [Online]. http://blogs.wsj.com/digits/2014/04/04/apple-acquires-speech-recognitionspecialist-novauris/ [35] Novauris Laboratories, Inc. (2002) Company Overview. https://web.archive.org/web/20020803202742/http://www.novauris.com/ 16 [Online].