- Speech Application Programming Interface
The Speech Application Programming Interface or SAPI is an API developed by
Microsoft to allow the use ofspeech recognition andspeech synthesis within Windows applications. To date a number of versions of the API have been released, which have shipped either as part of a SpeechSDK , or as part of the Windows OS itself. Applications that use SAPI includeMicrosoft Office ,Microsoft Agent andMicrosoft Speech Server .In general all versions of the API have been designed such that a software developer can write an application to perform speech recognition and synthesis by using a standard set of interfaces, accessible from a variety of programming languages. In addition, it is possible for a 3rd-party company to produce their own Speech Recognition and Text-To-Speech engines or adapt existing engines to work with SAPI. In principle, as long as these engines conform to the defined interfaces they can be used instead of the Microsoft-supplied engines.
In general the Speech API is freely-redistributable component which can be shipped with any Windows application that wishes to use speech technology. Many versions (although not all) of the speech recognition and synthesis engines are also freely redistributable.
There have been two main 'families' of the Microsoft Speech API. SAPI versions 1 through 4 are all similar to each other, with extra features in each newer version. SAPI 5 however was a completely new interface, released in
2000 . Since then several sub-versions of this API have been released.Basic architecture
Broadly the Speech API can be viewed as an interface or piece of middleware which sits between "applications" and speech "engines" (recognition and synthesis). In SAPI versions 1 to 4, applications could directly communicate with engines. The API included an abstract "interface definition" which applications and engines conformed to. Applications could also use simplified higher-level objects rather than directly call methods on the engines.
In SAPI 5 however, applications and engines do not directly communicate with each other. Instead each talk to a
runtime component (sapi.dll). There is an API implemented by this component which applications use, and another set of interfaces for engines.Typically in SAPI 5 applications issue calls through the API (for example to load a recognition grammar; start recognition; or provide text to be synthesized). The sapi.dll runtime component interprets these commands and processes them, where necessary calling on the engine through the engine interfaces (for example, the loading of a grammar from a file is done in the runtime, but then the grammar data is passed to the recognition engine to actually use in recognition). The recognition and synthesis engines also generate events while processing (for example, to indicate an utterance has been recognized or to indicate word boundaries in the synthesized speech). These pass in the reverse direction, from the engines, through the runtime dll, and on to an "event sink" in the application.
In addition to the actual API definition and runtime dll, other components are shipped with all versions of SAPI to make a complete Speech
Software Development Kit . The following components are among those included in most versions of the Speech SDK:
*"API definition files" - inMIDL and as C or C++ header files.
*"Runtime components" - e.g. sapi.dll.
*"Control Panel applet" - to select and configure default speech recognizer and synthesizer.
*"Text-To-Speech engines" in multiple languages.
*"Speech Recognition engines" in multiple languages.
*"Redistributable components" to allow developers to package the engines and runtime with their application code to produce a single installable application.
*"Sample application code".
*"Sample engines" - implementations of the necessary engine interfaces but with no true speech processing which could be used as a sample for those porting an engine to SAPI.
*"Documentation".Versions
Xuedong Huang was a key person who led Microsoft's early SAPI efforts.API 1-4 API family
API 1
The first version of SAPI was released in 1995, and was supported on
Windows 95 andWindows NT 3.51 . This version included low-level Direct Speech Recognition and Direct Text To Speech APIs which applications could use to directly control engines, as well as simplified 'higher-level' Voice Command and Voice Talk APIs.API 2
SAPI 2.0 was released in 1996.
API 3
SAPI 3.0 was released in 1997. It added limited support for dictation speech recognition (discrete speech, not continuous), and additional sample applications and audio sources.
API 4
SAPI 4.0 was released in 1998. This version of SAPI included both the core COM API; together with
C++ wrapper classes to make programming from C++ easier; andActiveX controls to allow drag-and-dropVisual Basic development. This was shipped as part of an SDK that included recognition and synthesis engines. It also shipped (with synthesis engines only) inWindows 2000 .The main components of the SAPI 4 API (which were all available in C++, COM, and ActiveX flavors) were:
*Voice Command - high-level objects for command & control speech recognition
*Voice Dictation - high-level objects for continuous dictation speech recognition
*Voice Talk - high-level objects for speech synthesis
*Voice Telephony - objects for writing telephone speech applications
*Direct Speech Recognition - objects for direct control of recognition engine
*Direct Text To Speech - objects for direct control of synthesis engine
*Audio objects - for reading to and from an audio device or fileAPI 5 API family
The Speech SDK version 5.0, incorporating the SAPI 5.0 runtime was released in 2000. This was a complete redesign from previous versions and neither engines nor applications which used older versions of SAPI could use the new version without considerable modification.
The design of the new API included the concept of strictly separating the application and engine so all calls were routed through the runtime sapi.dll. This change was intended to make the API more 'engine-independent', preventing applications from inadvertently depending on features of a specific engine. In addition this change was aimed at making it much easier to incorporate speech technology into an application by moving some management and initialization code into the runtime.
The new API was initially a pure COM API and could be used easily only from C/C++. Support for VB and scripting languages were added later. Operating systems from
Windows 98 andNT 4.0 upwards were supported.Major features of the API include:
*Shared Recognizer. For desktop speech recognition applications, a recognizer object can be used that runs in a separate process (sapisvr.exe). All applications using the shared recognizer communicate with this single instance. This allows sharing of resources, removes contention for the microphone and allows for a global UI for control of all speech applications.
*In-proc recognizer. For applications that require explicit control of the recognition process the in-proc recognizer object can be used instead of the shared one.
*Grammar objects. Speech grammars are used to specify the words that the recognizer is listening for. SAPI 5 defines anXML markup for specifying a grammar, as well as mechanisms to create them dynamically in code. Methods also exist for instructing the recognizer to load a built-in dictation language model.
*Voice object. This performs speech synthesis, producing an audio stream from text. A markup language (similar to XML, but not strictly XML) can be used for controlling the synthesis process.
*Audio interfaces. The runtime includes objects for performing speech input from the microphone or speech output to speakers (or any sound device); as well as to and from wave files. It is also possible to write a custom audio object to stream audio to or from a non-standard location.
*User lexicon object. This allows custom words and pronunciations to be added by a user or application. These are added to the recognition or synthesis engine's built-in lexicons.
*Object tokens. This is a concept allowing recognition and TTS engines, audio objects, lexicons and other categories of object to be registered, enumerated and instantiated in a common way.API 5.0
This version shipped in late 2000 as part of the Speech SDK version 5.0, together with version 5.0 recognition and synthesis engines. The recognition engines supported continuous dictation and command & control and were released in U.S. English, Japanese and Simplified Chinese versions. In the U.S. English system, special acoustic models were available for children's speech and telephony speech. The synthesis engine was available in English and Chinese. This version of the API and recognition engines also shipped in Microsoft Office XP in 2001.
API 5.1
This version shipped in late 2001 as part of the Speech SDK version 5.1. Automation-compliant interfaces were added to the API to allow use from Visual Basic, scripting languages such as
JScript , andmanaged code . This version of the API and TTS engines was shipped inWindows XP . This API was also shipped, together with a substantially improved version 6 recognition engine in Office 2003 andWindows XP Tablet PC Edition ._API 5.2
This was a special version of the API for use only in the
Microsoft Speech Server which shipped in 2004. It added support for SRGS and SSML mark-up languages, as well as additional server features and performance improvements. The Speech Server also shipped with the version 6 desktop recognition engine and the version 7 server recognition engine.API 5.3
This is the version of the API that ships in
Windows Vista together with new recognition and synthesis engines. AsWindows Speech Recognition is now integrated into the operating system, the Speech SDK and APIs are a part of theWindows SDK . SAPI 5.3 includes the following new features:* Support for W3C XML speech grammars for recognition and synthesis. The
Speech Synthesis Markup Language (SSML) version 1.0 provides the ability to mark up voice characteristics, speed, volume, pitch, emphasis, and pronunciation.
* TheSpeech Recognition Grammar Specification (SRGS) supports the definition of context-free grammars, with two limitations:
** It does not support the use of SRGS to specify dual-tone modulated-frequency (touch-tone) grammars.
** It does not supportAugmented Backus–Naur form (ABNF).
* Support for semantic interpretation script within grammars. SAPI 5.3 enables an SRGS grammar to be annotated withJavaScript for semantic interpretation to supplement the recognized text.
* User-Specified shortcuts in lexicons, which is the ability to add a string to the lexicon and associate it with a shortcut word. When dictating, the user can say the shortcut word and the recognizer will return the expanded string.
* Additional functionality and ease-of-programming provided by new types.
* Performance improvements, improved reliability and security.
* Version 8 of the speech recognition engine ("Microsoft Speech Recognizer")API 5 Voices
Microsoft Sam (Speech Articulation Module) is a commonly-shipped SAPI 5 voice. In addition,
Microsoft Office XP and Office 2003 installed L&H Michael and Michelle voices. The SAPI 5.1 SDK installs 2 more voices, "Mike" and "Mary".Windows Vista includesMicrosoft Anna which replaces Microsoft Sam. Anna is designed to sound more natural and offer greater intelligibility. Several multilingual voices are also included in localized versions of Windows Vista. Microsoft Anna is also installed on Windows XP byMicrosoft Streets & Trips 2006 and later versions.Managed code Speech API
A
managed code API ships as part of the.NET Framework 3.0 [ [http://reddevnews.com/techbriefs/article.aspx?editorialsid=182 Speech synthesis and recognition in .NET - Give applications a voice: Redmond Developer News] ] . It has similar functionality to SAPI 5 but is more suitable to be used by managed code applications. The new API is available onWindows XP ,Windows Server 2003 ,Windows Vista , andWindows Server 2008 .The existing SAPI 5 API can also be used from managed code to a limited extent by creating COM Interop code (helper code designed to assist in accessing COM interfaces and classes). This works well in some scenarios however the new API should provide a more seamless experience equivalent to using any other managed code library.Lopsided|date=September 2008
peech functionality in Windows Vista
Windows Vista includes a number of new speech-related features including:
* Speech control of the full Windows GUI and applications
* New tutorial, microphone wizard, and UI for controlling speech recognition
* New version of the Speech API runtime: SAPI 5.3
* Built-in updated Speech Recognition engine (Version 8)
* New Speech Synthesis engine and SAPI voiceMicrosoft Anna
*Managed code speech API (codenamed SpeechFX)
* Speech recognition support for 8 languages at release time: U.S. English, U.K. English, traditional Chinese, simplified Chinese, Japanese, German, French and Spanish, with more language to be released later.
*Microsoft Agent most notably, and all other Microsoft speech applications use SAPI 5.Compatibility
The Speech API is compatible with the following operating systems: cite web
author=Microsoft Corporation
title=SAPI System Requirements
publisher=MSDN
url=http://msdn.microsoft.com/library/default.asp?url=/library/en-us/SAPI51sr/html/system_requirements.asp
accessdate=2006-04-12]
*Microsoft Windows Vista
*Microsoft Windows XP
*Microsoft Windows 2000
*Microsoft Windows Millennium Edition
*Microsoft Windows 98
*Microsoft Windows NT 4.0, Service Pack 6a, in English, Japanese and Simplified Chinese.Major applications using SAPI
*MicrosoftWindows XP Tablet PC Edition includes SAPI 5.1 and speech recognition engines 6.1 for English, Japanese, and Chinese (simplified and traditional)
*Windows Speech Recognition inWindows Vista
*Microsoft Narrator in Windows 2000 and later Windows operating systems
*Microsoft Office XP and Office 2003
*Microsoft Excel 2002, Microsoft Excel 2003, and Microsoft Excel 2007 for speaking spreadsheet data
*Microsoft Voice Command for Windows Pocket PC and Windows Mobile
*Microsoft Plus! Voice Command for Windows Media Player
*Dragon Naturally Speaking
*Adobe Reader uses voice output to read document content
*Window-Eyes screen reader
*JAWS screen readeree also
*
List of speech recognition software
* Microsoft Speech Application SDK (SASDK)
*Speech recognition
*Speech synthesis
*Windows Vista
*Windows Speech Recognition
*Windows Speech Recognition Macros External links
* [http://www.microsoft.com/speech/speech2007/default.mspx Microsoft site for SAPI]
* [http://www.microsoft.com/downloads/details.aspx?FamilyID=5e86ec97-40a7-453f-b0ee-6583171b4530 Microsoft download site for Speech API Software Developers Kit version 5.1]
* [http://www.microsoft.com/msj/archive/s233.aspx Microsoft Systems Journal Whitepaper by Mike Rozak on the first version of SAPI]
* [http://blogs.msdn.com/speech Microsoft Speech Team blog]References
Wikimedia Foundation. 2010.