EP0945851A2 - Extending the vocabulary of a client-server speech recognition system - Google Patents
Extending the vocabulary of a client-server speech recognition system Download PDFInfo
- Publication number
- EP0945851A2 EP0945851A2 EP99301097A EP99301097A EP0945851A2 EP 0945851 A2 EP0945851 A2 EP 0945851A2 EP 99301097 A EP99301097 A EP 99301097A EP 99301097 A EP99301097 A EP 99301097A EP 0945851 A2 EP0945851 A2 EP 0945851A2
- Authority
- EP
- European Patent Office
- Prior art keywords
- server
- words
- communicated
- vocabulary list
- network
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000015572 biosynthetic process Effects 0.000 description 8
- 238000003786 synthesis reaction Methods 0.000 description 7
- 238000000034 method Methods 0.000 description 6
- 239000007858 starting material Substances 0.000 description 6
- 238000012937 correction Methods 0.000 description 4
- 238000007792 addition Methods 0.000 description 2
- 230000003190 augmentative effect Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000005540 biological transmission Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000002708 enhancing effect Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/28—Constructional details of speech recognition systems
- G10L15/30—Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
Definitions
- This invention relates generally to the field of speech applications, and in particular, to a method and apparatus for automatically storing, tracking and distributing new word pronunciations to speech application clients on a network.
- spoken language with computers typically associated with speech recognition and speech synthesis, involves storing and retrieving not only word spellings but other data associated with words, for example phonemes, alternate pronunciations, associations with other words and parts of speech, such as noun, verb, adjective and the like.
- Defining the work to be done by computers to support spoken language is aided by comparing spoken language to written language. In the intersection of these two forms of communication there are words. Outside of the intersection words are represented differently, as spoken sounds or written letters. Written language is also augmented outside of the intersection by punctuation, or even font variations such as bold for emphasis. Spoken language is augmented differently outside of the intersection, for example by volume, pitch, prosody (speed), and inflection.
- spoken language as typically spoken is converted to text form by way of speech recognition and converted back to spoken form by way of speech synthesis.
- the differences between written and spoken words outside of the intersection create a number of problems for speech applications.
- Speech recognition engines as supplied in speech applications just aren't accurate enough and are sometimes slow in throughput due to misrecognitions and the time needed to correct misrecognitions.
- Correcting misrecognized words is a major factor in measures of speech recognition effectiveness including words-per-minute and usability. Large vocabularies are provided to limit the number of corrections resulting from out-of-vocabulary words.
- a starter set of words including pronunciations, is installed on a computer.
- Pronunciations are stored as base forms, which represent instructions as to the way words are pronounced, or sound.
- the number of words in a starter set is usually determined by balancing considerations such as amount of storage space required, frequency of word use, degree of common use, and recognition accuracy against all other words in the starter set.
- Developers of speech recognition systems typically install a large vocabulary of word-pronunciation data as required for recognition. It is a rather tedious and time consuming task to produce such a vocabulary.
- a network universal spoken language vocabulary developed and managed in accordance with the inventive arrangements overcomes all of the problems noted in the prior art.
- the inventive arrangements automate and optimize storage, tracking, sharing, and delivery of spoken language vocabulary data. Word information regarding all known words can be distributed from a network connected server. New words added by any client user are made available to all users, eliminating the need for subsequent users to each provide data regarding that same new word.
- a speech recognition and/or speech synthesis system or application according to the inventive arrangements eliminates the need for more than one user to add a new word. Only the very first user of a new word needs to add that new word.
- the system can, while recognizing or determining alternatives, query a universal vocabulary server for a new word and will not bother the user to provide the data.
- a network spoken language vocabulary system for a speech application comprises: a network server and a plurality of network clients communicating with one another over a network; a central vocabulary list in the server for recognizable words; a speech recognition engine and a local vocabulary list cache in each of the clients, the cache having therein previously recognized words communicated from the central vocabulary list in the server and new words recognized by the speech application; and, each of the new words being communicated to the server and added to the central vocabulary list and each of the new words added to the central vocabulary list being communicated to at least one other of the clients for addition to the cache.
- the new words can be automatically communicated from the clients to the server and can also be automatically communicated from the server to one or more clients.
- the server can further comprise a user profile table having data therein for each of the users indicative of a respective subset of the central vocabulary list for use in speech recognition data and/or indicative of membership in a work group.
- a user profile table having data therein for each of the users indicative of a respective subset of the central vocabulary list for use in speech recognition data and/or indicative of membership in a work group.
- Each of the clients can further comprise an add word pronunciation tool for identifying the new words to be communicated to the server, and for adding the new words to the cache.
- the add word pronunciation tool can add the new words as recognized by the speech recognition engine in the client and as communicated from the server.
- the central vocabulary list can comprise a plurality of data records for each word, the data records including: spelling; phonemes; usage domains; client of origin; and, workgroups requiring the word.
- a network spoken language vocabulary system for a speech application comprises: a network server having a permanent central vocabulary list for recognizable words and a user profile data base; a plurality of network clients, each having a speech recognition engine and a local vocabulary list cache; a network over which the server and the clients can communicate; the cache having stored therein a subset of previously recognized words communicated from the central vocabulary list in the server and new words as recognized by the speech application during the session; each of the new words being communicated to the server and added to the central vocabulary list; and, each of the new words added to the central vocabulary list being communicated to other ones of the clients in accordance with data in the user profile data base.
- the new words can be automatically communicated to and from the server.
- the user profile data base has data therein for each of the users indicative of a respective subset of the central vocabulary list for use in speech recognition and/or indicative of membership in a work group.
- each new word communicated to the server from any user in one of the workgroups can be automatically communicated back to all other users in the work group.
- each user in the work group can receive the same subset of the central vocabulary list for use in speech recognition.
- Each of the clients can further comprise an add word pronunciation tool for identifying the new words to be communicated to the server, and for adding the new words to the cache.
- the add word pronunciation tool can add the new words as recognized by the speech recognition engine in the client and as communicated from the server.
- the central vocabulary list can comprise a plurality of data records for each word, the data records including: spelling; phonemes; usage domains; client of origin; and, workgroups requiring the word.
- the sole Figure is a block diagram of a network universal spoken language vocabulary system in accordance with the inventive arrangements.
- a comprehensive system 10 which covers all facets of vocabulary support for spoken language technologies such as speech recognition and speech synthesis is shown in the sole Figure.
- the system comprises a word pronunciation or vocabulary server 12 connected to a plurality of vocabulary clients 14, 16 and 18 over a network 20.
- the network 20 can be a local area network (LAN) or a wide area network (WAN).
- the word pronunciation or vocabulary server 12 is a universal repository for all known words and associated spoken language attributes. New words are acquired from and distributed to network clients as needed.
- the word pronunciation server 12 comprises a universal vocabulary server program 22, a universal vocabulary list, or data base 24 and a user profile table 26.
- the word pronunciation server 12 is a network connected computer system which hosts the universal vocabulary database and the universal vocabulary server software program 22 which responds to requests by client computer systems which are running speech recognition or speech synthesis programs.
- the universal vocabulary list 24 includes all words known to the system including the spelling and data indicating how the word is pronounced.
- the term universal denotes throughout the network. In this sense, the universal vocabulary list can also be thought of as a central or master vocabulary list.
- Data records for each word entry include the following information: spelling; phonemes, that is, pronunciation data; list of usage domains; origin, that is, who added the word; and, workgroup, that is, which users are likely to use the word.
- the user vocabulary profile 26 is a database which contains information required to support the universal vocabulary client program. Data regarding which word groups a particular user uses, and the user groups which the user belongs to is maintained on the server and transmitted to each client when they begin a session with speech recognition or TTS software.
- the term client is used in the sense of a work station served by the word pronunciation server. Users are persons working at the clients.
- the user profile table or data base 26 can be keyed to users, who can be identified for example by name or by log in identification, so that any user can be provided with the appropriate subset of vocabulary words irrespective of the particular work station or client being utilized. If the user of a particular client or work station is not in the user profile table, the user profile table can also be keyed on the client or word station identification. Alternatively, a default subset for unidentified users can also be defined. As a further and presently preferred alternative, an unidentified user can be prompted to supply sufficient information to be identified in the user profile table, and thereafter, be supplied with an appropriate subset vocabulary list and appropriate new words.
- the proper vocabulary list can even be transferred to a remote work station by modem, satellite transmission, or the Internet, although immediate and automatic updates would be available only for so long as the remote link was maintained.
- the network 20 also represents remote links.
- Each of the clients 14, 16 and 18 comprises a universal vocabulary client 30, a speech recognition engine 32, an add word pronunciation tool 34 and a local vocabulary list cache 36, for storing a personalized subset of the universal vocabulary list 24.
- clients 14 and 16 are members of a work group associated with a user profile designated A.
- Client 18 is a member of a work group associated with a user profile designated B.
- the add word pronunciation tool 34 transmits word newly added to each client over the network 20 to the word pronunciation server 12, where the new words are added to the universal vocabulary list 24.
- the add word pronunciation tool 34 also receives new words from the word pronunciation server 12 over the network 20, and routes the new words to the local list cache 36.
- the universal vocabulary client program 30 is an application which runs on a client system maintaining the local cached subset 36 of the universal vocabulary list 24. Words are added to the cache from the user and from the universal vocabulary list, while words which have expired in terms of cache-worthiness are deleted.
- the cache 36 is a dynamically optimized replacement for the traditional stand-alone starter-set vocabulary.
- the programs which maintain the cache 36 employ algorithms which can consider network traffic, available local storage, frequency of use, system processor speed, current end-user task, and end-user settings to determine the parameters for the cache.
- the cache can be thought of as an optimal local vocabulary.
- the cache is optimal because specialized words known to be spoken by the user will be included, enhancing recognition, and at the same time, words known not to be spoken by the user will not be included, reducing processing time needed to search the cache for matches.
- the client application sends data regarding newly added words to the universal vocabulary server program.
- all words can be added as public, that is to be shared, but a user can choose to have added words held as private, if desired for any reason.
- Client 14 has added a new word.
- the new word is transmitted over the network 20 from the add word pronunciation tool 34 to the word pronunciation server 12, where the new word is added to the universal vocabulary list 24, as indicated by dashed line arrow 1.
- the universal vocabulary server program 22 recalls the user profile for client 16 from the user profile table 26, as indicated by dashed line arrow 2.
- the profile reveals that client 16 belongs to the same workgroup as client 14, represented by the use of the same profile A.
- client 16 is already dictating, the new word is delivered to client 16 and added to the local cache 36.
- client 16 begins dictating, a personalized vocabulary list is delivered to client 16 and stored in the local cache 36.
- the new word added earlier by client 14 is included, because both user 14 and user 16 use profile A. Both instances are indicated by dashed line arrow 3.
- a comprehensive database is dynamically and incrementally enhanced by users and automatically optimized over time to increase overall system performance and to balance computer system resources, such as network bandwidth and file and data storage, using caching techniques.
- the central vocabulary list may comprise a plurality of data records for each word, said data records including: spelling; phonemes; usage domains; client of origin; and, workgroups requiring said word.
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Computer And Data Communications (AREA)
- Machine Translation (AREA)
- Document Processing Apparatus (AREA)
Abstract
Description
- This invention relates generally to the field of speech applications, and in particular, to a method and apparatus for automatically storing, tracking and distributing new word pronunciations to speech application clients on a network.
- Use of spoken language with computers, typically associated with speech recognition and speech synthesis, involves storing and retrieving not only word spellings but other data associated with words, for example phonemes, alternate pronunciations, associations with other words and parts of speech, such as noun, verb, adjective and the like.
- Computer systems at first were developed to deal exclusively with written language. Written language is useful for many things, and is much simpler to interpret, represent within, and reproduce from a computer system than is spoken language. Today, computer systems are taking on spoken language in the form of two technologies, namely, speech recognition and speech synthesis. Speech synthesis is also referred to as text-to-speech (TTS).
- Defining the work to be done by computers to support spoken language is aided by comparing spoken language to written language. In the intersection of these two forms of communication there are words. Outside of the intersection words are represented differently, as spoken sounds or written letters. Written language is also augmented outside of the intersection by punctuation, or even font variations such as bold for emphasis. Spoken language is augmented differently outside of the intersection, for example by volume, pitch, prosody (speed), and inflection.
- As computers tackle support for spoken language, spoken language as typically spoken is converted to text form by way of speech recognition and converted back to spoken form by way of speech synthesis. This takes advantage of the significantly reduced requirements on system resources to store or transmit a written representation as compared to an audible representation. The differences between written and spoken words outside of the intersection create a number of problems for speech applications.
- End-users are greatly inconvenienced by the need to add word pronunciations which are not included in the starter set of vocabulary words which can be recognized. By design, the user encounters this problem as a special case of a word which was recognized incorrectly. Recognition can be viewed as a best-guess by the recognition engine as to which word was spoken by the user. When the user speaks a word which is not known to the recognition engine, the engine simply guesses wrong. The user must then initiate correction of the word and choose a new word from a short list of appropriate alternatives. If the spoken word is not listed as an alternate choice, the user typically is required to type the word and perhaps pronounce it again. This inconvenience can encourage users to bypass the proper correction procedures and simply type in the corrected text. Unfortunately, although this procedure is quicker in the short run, it is important that the speech recognition system learn about the correction with the proper procedures because this information is the only way to add and correct words, thereby improving future recognition performance.
- Speech recognition engines as supplied in speech applications just aren't accurate enough and are sometimes slow in throughput due to misrecognitions and the time needed to correct misrecognitions.
- Correcting misrecognized words is a major factor in measures of speech recognition effectiveness including words-per-minute and usability. Large vocabularies are provided to limit the number of corrections resulting from out-of-vocabulary words.
- In a stand alone system, user added words need to be backed up and migrated from system to system as a user moves around or switches between systems (for example, between home and office computers and between portable and desktop computers) or upgrades a computer or a speech recognition application, or both. This task is time consuming, tedious, unobvious, and subsequently, not generally done.
- Typically, along with speech recognition or synthesis software, a starter set of words, including pronunciations, is installed on a computer. Pronunciations are stored as base forms, which represent instructions as to the way words are pronounced, or sound. Many factors must be considered in order to produce an ideal starter set for a given application. The number of words in a starter set is usually determined by balancing considerations such as amount of storage space required, frequency of word use, degree of common use, and recognition accuracy against all other words in the starter set. Developers of speech recognition systems typically install a large vocabulary of word-pronunciation data as required for recognition. It is a rather tedious and time consuming task to produce such a vocabulary.
- A network universal spoken language vocabulary developed and managed in accordance with the inventive arrangements overcomes all of the problems noted in the prior art. The inventive arrangements automate and optimize storage, tracking, sharing, and delivery of spoken language vocabulary data. Word information regarding all known words can be distributed from a network connected server. New words added by any client user are made available to all users, eliminating the need for subsequent users to each provide data regarding that same new word.
- Fewer manual additions of pronunciations are required from each user, so that the users are more likely to follow those procedures for correcting misrecognitions in a manner which teaches the speech application new words, thus improving future recognition performance for all users.
- A speech recognition and/or speech synthesis system or application according to the inventive arrangements eliminates the need for more than one user to add a new word. Only the very first user of a new word needs to add that new word. The system can, while recognizing or determining alternatives, query a universal vocabulary server for a new word and will not bother the user to provide the data.
- Increased accuracy and throughput of speech recognition are among the advantages of a highly optimized, cached universal vocabulary simply by virtue of a higher hit-rate between words spoken and words known and accurately recognized.
- The requirements for installing and maintaining a large vocabulary on each system is advantageously reduced. It is transparent to the user that some words are stored on a server, and as the system gets used, caching will increase the hit-rate, reducing the number of words stored on a client as well as generally speeding things up.
- The need to maintain a users added words, and the need to individual back up and migrate updated vocabulary lists is advantageously eliminated by centralized data backup.
- The burden of carefully selecting the most appropriate starter set in advance is advantageously eliminated because specialized words will be added efficiently and in a shared-burden manner by the network of users. On the contrary, the vocabulary for all users is automatically updated as needed, without requiring user interaction.
- A network spoken language vocabulary system for a speech application, in accordance with an inventive arrangement, comprises: a network server and a plurality of network clients communicating with one another over a network; a central vocabulary list in the server for recognizable words; a speech recognition engine and a local vocabulary list cache in each of the clients, the cache having therein previously recognized words communicated from the central vocabulary list in the server and new words recognized by the speech application; and, each of the new words being communicated to the server and added to the central vocabulary list and each of the new words added to the central vocabulary list being communicated to at least one other of the clients for addition to the cache.
- The new words can be automatically communicated from the clients to the server and can also be automatically communicated from the server to one or more clients.
- The server can further comprise a user profile table having data therein for each of the users indicative of a respective subset of the central vocabulary list for use in speech recognition data and/or indicative of membership in a work group. When this data is available, each new word communicated to the server from any user in one of the workgroups can be automatically communicated back to all other users in the work group. Moreover, each user in the work group can receive the same subset of the central vocabulary list for use in speech recognition.
- Each of the clients can further comprise an add word pronunciation tool for identifying the new words to be communicated to the server, and for adding the new words to the cache. The add word pronunciation tool can add the new words as recognized by the speech recognition engine in the client and as communicated from the server.
- The central vocabulary list can comprise a plurality of data records for each word, the data records including: spelling; phonemes; usage domains; client of origin; and, workgroups requiring the word.
- A network spoken language vocabulary system for a speech application, in accordance with another inventive arrangement, comprises: a network server having a permanent central vocabulary list for recognizable words and a user profile data base; a plurality of network clients, each having a speech recognition engine and a local vocabulary list cache; a network over which the server and the clients can communicate; the cache having stored therein a subset of previously recognized words communicated from the central vocabulary list in the server and new words as recognized by the speech application during the session; each of the new words being communicated to the server and added to the central vocabulary list; and, each of the new words added to the central vocabulary list being communicated to other ones of the clients in accordance with data in the user profile data base.
- The new words can be automatically communicated to and from the server.
- The user profile data base has data therein for each of the users indicative of a respective subset of the central vocabulary list for use in speech recognition and/or indicative of membership in a work group. When this information is available, each new word communicated to the server from any user in one of the workgroups can be automatically communicated back to all other users in the work group. Moreover, each user in the work group can receive the same subset of the central vocabulary list for use in speech recognition.
- Each of the clients can further comprise an add word pronunciation tool for identifying the new words to be communicated to the server, and for adding the new words to the cache. The add word pronunciation tool can add the new words as recognized by the speech recognition engine in the client and as communicated from the server.
- The central vocabulary list can comprise a plurality of data records for each word, the data records including: spelling; phonemes; usage domains; client of origin; and, workgroups requiring the word.
- The sole Figure is a block diagram of a network universal spoken language vocabulary system in accordance with the inventive arrangements.
- A
comprehensive system 10 which covers all facets of vocabulary support for spoken language technologies such as speech recognition and speech synthesis is shown in the sole Figure. The system comprises a word pronunciation orvocabulary server 12 connected to a plurality ofvocabulary clients network 20. Thenetwork 20 can be a local area network (LAN) or a wide area network (WAN). In accordance with the inventive arrangements, the word pronunciation orvocabulary server 12 is a universal repository for all known words and associated spoken language attributes. New words are acquired from and distributed to network clients as needed. - The
word pronunciation server 12 comprises a universalvocabulary server program 22, a universal vocabulary list, ordata base 24 and a user profile table 26. Theword pronunciation server 12 is a network connected computer system which hosts the universal vocabulary database and the universal vocabularyserver software program 22 which responds to requests by client computer systems which are running speech recognition or speech synthesis programs. - The
universal vocabulary list 24 includes all words known to the system including the spelling and data indicating how the word is pronounced. The term universal denotes throughout the network. In this sense, the universal vocabulary list can also be thought of as a central or master vocabulary list. Data records for each word entry include the following information: spelling; phonemes, that is, pronunciation data; list of usage domains; origin, that is, who added the word; and, workgroup, that is, which users are likely to use the word. - The
user vocabulary profile 26 is a database which contains information required to support the universal vocabulary client program. Data regarding which word groups a particular user uses, and the user groups which the user belongs to is maintained on the server and transmitted to each client when they begin a session with speech recognition or TTS software. - The term client is used in the sense of a work station served by the word pronunciation server. Users are persons working at the clients. The user profile table or
data base 26 can be keyed to users, who can be identified for example by name or by log in identification, so that any user can be provided with the appropriate subset of vocabulary words irrespective of the particular work station or client being utilized. If the user of a particular client or work station is not in the user profile table, the user profile table can also be keyed on the client or word station identification. Alternatively, a default subset for unidentified users can also be defined. As a further and presently preferred alternative, an unidentified user can be prompted to supply sufficient information to be identified in the user profile table, and thereafter, be supplied with an appropriate subset vocabulary list and appropriate new words. - The proper vocabulary list can even be transferred to a remote work station by modem, satellite transmission, or the Internet, although immediate and automatic updates would be available only for so long as the remote link was maintained. In this respect, the
network 20 also represents remote links. - Each of the
clients universal vocabulary client 30, aspeech recognition engine 32, an addword pronunciation tool 34 and a localvocabulary list cache 36, for storing a personalized subset of theuniversal vocabulary list 24. For purposes of illustration,clients A. Client 18 is a member of a work group associated with a user profile designated B. - The add
word pronunciation tool 34 transmits word newly added to each client over thenetwork 20 to theword pronunciation server 12, where the new words are added to theuniversal vocabulary list 24. The addword pronunciation tool 34 also receives new words from theword pronunciation server 12 over thenetwork 20, and routes the new words to thelocal list cache 36. - The universal
vocabulary client program 30 is an application which runs on a client system maintaining the localcached subset 36 of theuniversal vocabulary list 24. Words are added to the cache from the user and from the universal vocabulary list, while words which have expired in terms of cache-worthiness are deleted. Thecache 36 is a dynamically optimized replacement for the traditional stand-alone starter-set vocabulary. - The programs which maintain the
cache 36 employ algorithms which can consider network traffic, available local storage, frequency of use, system processor speed, current end-user task, and end-user settings to determine the parameters for the cache. The cache can be thought of as an optimal local vocabulary. The cache is optimal because specialized words known to be spoken by the user will be included, enhancing recognition, and at the same time, words known not to be spoken by the user will not be included, reducing processing time needed to search the cache for matches. - The client application sends data regarding newly added words to the universal vocabulary server program. By default, all words can be added as public, that is to be shared, but a user can choose to have added words held as private, if desired for any reason.
- A simple example has been charted in the Figure.
Client 14 has added a new word. The new word is transmitted over thenetwork 20 from the addword pronunciation tool 34 to theword pronunciation server 12, where the new word is added to theuniversal vocabulary list 24, as indicated by dashed line arrow 1. Ifclient 16 is dictating, or sometime thereafter begins dictating, the universalvocabulary server program 22 recalls the user profile forclient 16 from the user profile table 26, as indicated by dashedline arrow 2. The profile reveals thatclient 16 belongs to the same workgroup asclient 14, represented by the use of the same profile A. Ifclient 16 is already dictating, the new word is delivered toclient 16 and added to thelocal cache 36. Ifclient 16 begins dictating, a personalized vocabulary list is delivered toclient 16 and stored in thelocal cache 36. The new word added earlier byclient 14 is included, because bothuser 14 anduser 16 use profile A. Both instances are indicated by dashedline arrow 3. - In accordance with the inventive arrangements, a comprehensive database is dynamically and incrementally enhanced by users and automatically optimized over time to increase overall system performance and to balance computer system resources, such as network bandwidth and file and data storage, using caching techniques.
- In the present invention the central vocabulary list may comprise a plurality of data records for each word, said data records including: spelling; phonemes; usage domains; client of origin; and, workgroups requiring said word.
Claims (10)
- A network spoken language vocabulary system for a speech application, characterised by a network server (12) and a plurality of network clients (14, 16, 18) communicating with one another over a network; a central vocabulary list (24) in said server for recognizable words; and a speech recognition engine (32) and a local vocabulary list cache (36) in each of said clients, said cache having therein previously recognized words communicated from said central vocabulary list in said server and new words recognized by said speech application; each of said new words being communicated to said server and added to said central vocabulary list and each of said new words added to said central vocabulary list being communicated to at least one other of said clients for addition to said cache.
- A system of claim 1, characterised in that said new words are automatically communicated back to said server.
- A network spoken language vocabulary system for a speech application, characterised by a network server (12) having a permanent central vocabulary list (24) for recognizable words and a user profile data base (26); a plurality of network clients (14, 16, 18), each having a speech recognition engine (32) and a local vocabulary list cache (36); and a network (20) over which said server and said clients can communicate; said cache (36) having stored therein a subset of previously recognized words communicated from said central vocabulary list (24) in said server (12) and new words as recognized by said speech application during said session; each of said new words being communicated to said server (12) and added to said central vocabulary list (24); and each of said new words added to said central vocabulary list (24) being communicated to other ones of said clients in accordance with data in said user profile data base (26).
- A system of claim 1 or 3, characterised in that said new words are automatically communicated to and from said server.
- A system of claim 1 or 3, characterised by a user profile data base having data therein for each of said users indicative of a respective subset of said central vocabulary list for use in speech recognition.
- A system of claim 1 or 3, characterised in that said user profile data base has data therein for each of said users indicative of membership in a work group.
- A system of claim 6, characterised in that each new word communicated to said server from any user in one of said workgroups is automatically communicated back to all other users in said work group.
- A system of claim 1 or 3, characterised in that said user profile data base has data therein for each of said users indicative of membership in a work group, each user in said work group receiving the same subset of said central vocabulary list for use in speech recognition.
- A system of claim 1 or 3, characterised in that each of said clients further comprises an add word pronunciation tool for identifying said new words to be communicated to said server, and for adding said new words to said cache.
- The system of claim 9, characterised in that said add word pronunciation tool adds said new words as recognized by said speech recognition engine in said client and as communicated from said server.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/049,736 US6195641B1 (en) | 1998-03-27 | 1998-03-27 | Network universal spoken language vocabulary |
US49736 | 1998-03-27 |
Publications (3)
Publication Number | Publication Date |
---|---|
EP0945851A2 true EP0945851A2 (en) | 1999-09-29 |
EP0945851A3 EP0945851A3 (en) | 2000-01-12 |
EP0945851B1 EP0945851B1 (en) | 2004-05-12 |
Family
ID=21961423
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP99301097A Expired - Lifetime EP0945851B1 (en) | 1998-03-27 | 1999-02-15 | Extending the vocabulary of a client-server speech recognition system |
Country Status (6)
Country | Link |
---|---|
US (1) | US6195641B1 (en) |
EP (1) | EP0945851B1 (en) |
JP (1) | JP3519015B2 (en) |
KR (1) | KR100324079B1 (en) |
DE (1) | DE69917112T2 (en) |
TW (1) | TW405109B (en) |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2356318A (en) * | 1999-09-04 | 2001-05-16 | Marconi Comm Ltd | Server with voice recognition |
EP1209662A2 (en) * | 2000-11-27 | 2002-05-29 | Canon Kabushiki Kaisha | Client-server based speech recognition |
EP1215661A1 (en) * | 2000-12-14 | 2002-06-19 | TELEFONAKTIEBOLAGET L M ERICSSON (publ) | Mobile terminal controllable by spoken utterances |
WO2002059874A2 (en) * | 2001-01-05 | 2002-08-01 | Qualcomm Incorporated | System and method for voice recognition in a distributed voice recognition system |
FR2820872A1 (en) * | 2001-02-13 | 2002-08-16 | Thomson Multimedia Sa | VOICE RECOGNITION METHOD, MODULE, DEVICE AND SERVER |
EP1665229A2 (en) * | 2003-09-17 | 2006-06-07 | Motorola, Inc. | Speech synthesis |
WO2006097598A1 (en) * | 2005-03-16 | 2006-09-21 | France Telecom | Method for automatically producing voice labels in an address book |
WO2006084144A3 (en) * | 2005-02-03 | 2006-11-30 | Voice Signal Technologies Inc | Methods and apparatus for automatically extending the voice-recognizer vocabulary of mobile communications devices |
WO2007024427A1 (en) | 2005-08-19 | 2007-03-01 | Cisco Technology, Inc. | System and method for distributing a speech-recognition grammar |
WO2016039945A1 (en) * | 2014-09-08 | 2016-03-17 | Qualcomm Incorporated | Keyword detection using speaker-independent keyword models for user-designated keywords |
Families Citing this family (222)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6412011B1 (en) * | 1998-09-14 | 2002-06-25 | At&T Corp. | Method and apparatus to enhance a multicast information stream in a communication network |
US7003463B1 (en) * | 1998-10-02 | 2006-02-21 | International Business Machines Corporation | System and method for providing network coordinated conversational services |
DE69919842T2 (en) * | 1998-12-21 | 2005-09-01 | Philips Intellectual Property & Standards Gmbh | LANGUAGE MODEL BASED ON THE LANGUAGE RECOGNITION HISTORY |
DE19918382B4 (en) * | 1999-04-22 | 2004-02-05 | Siemens Ag | Creation of a reference model directory for a voice-controlled communication device |
US8645137B2 (en) | 2000-03-16 | 2014-02-04 | Apple Inc. | Fast, language-independent method for user authentication by voice |
US6741963B1 (en) * | 2000-06-21 | 2004-05-25 | International Business Machines Corporation | Method of managing a speech cache |
US7024454B1 (en) * | 2000-08-25 | 2006-04-04 | Practicefirst.Com L.L.C. | Work sharing and communicating in a web site system |
JP4109414B2 (en) * | 2000-12-18 | 2008-07-02 | セイコーエプソン株式会社 | Device control method using speech recognition, device control system using speech recognition, and recording medium recording device control program using speech recognition |
JP3838029B2 (en) * | 2000-12-18 | 2006-10-25 | セイコーエプソン株式会社 | Device control method using speech recognition and device control system using speech recognition |
US7970648B2 (en) * | 2001-04-27 | 2011-06-28 | Accenture Global Services Limited | Advertising campaign and business listing management for a location-based services system |
US7437295B2 (en) * | 2001-04-27 | 2008-10-14 | Accenture Llp | Natural language processing for a location-based services system |
US6848542B2 (en) * | 2001-04-27 | 2005-02-01 | Accenture Llp | Method for passive mining of usage information in a location-based services system |
US6944447B2 (en) * | 2001-04-27 | 2005-09-13 | Accenture Llp | Location-based services |
US7698228B2 (en) * | 2001-04-27 | 2010-04-13 | Accenture Llp | Tracking purchases in a location-based services system |
US7366673B2 (en) * | 2001-06-15 | 2008-04-29 | International Business Machines Corporation | Selective enablement of speech recognition grammars |
US20030009334A1 (en) * | 2001-07-03 | 2003-01-09 | International Business Machines Corporation | Speech processing board for high volume speech processing applications |
US7920682B2 (en) * | 2001-08-21 | 2011-04-05 | Byrne William J | Dynamic interactive voice interface |
ITFI20010199A1 (en) | 2001-10-22 | 2003-04-22 | Riccardo Vieri | SYSTEM AND METHOD TO TRANSFORM TEXTUAL COMMUNICATIONS INTO VOICE AND SEND THEM WITH AN INTERNET CONNECTION TO ANY TELEPHONE SYSTEM |
US7133829B2 (en) * | 2001-10-31 | 2006-11-07 | Dictaphone Corporation | Dynamic insertion of a speech recognition engine within a distributed speech recognition system |
US7146321B2 (en) * | 2001-10-31 | 2006-12-05 | Dictaphone Corporation | Distributed speech recognition system |
US6785654B2 (en) | 2001-11-30 | 2004-08-31 | Dictaphone Corporation | Distributed speech recognition system with speech recognition engines offering multiple functionalities |
US7013275B2 (en) * | 2001-12-28 | 2006-03-14 | Sri International | Method and apparatus for providing a dynamic speech-driven control and remote service access system |
JP4017887B2 (en) * | 2002-02-28 | 2007-12-05 | 富士通株式会社 | Voice recognition system and voice file recording system |
US7016849B2 (en) * | 2002-03-25 | 2006-03-21 | Sri International | Method and apparatus for providing speech-driven routing between spoken language applications |
US7236931B2 (en) | 2002-05-01 | 2007-06-26 | Usb Ag, Stamford Branch | Systems and methods for automatic acoustic speaker adaptation in computer-assisted transcription systems |
US7292975B2 (en) * | 2002-05-01 | 2007-11-06 | Nuance Communications, Inc. | Systems and methods for evaluating speaker suitability for automatic speech recognition aided transcription |
US8959019B2 (en) | 2002-10-31 | 2015-02-17 | Promptu Systems Corporation | Efficient empirical determination, computation, and use of acoustic confusability measures |
JP2004212641A (en) * | 2002-12-27 | 2004-07-29 | Toshiba Corp | Voice input system and terminal device equipped with voice input system |
US8150697B2 (en) * | 2003-09-30 | 2012-04-03 | Nuance Communications, Inc. | Autonomous systems and network management using speech |
US8954325B1 (en) * | 2004-03-22 | 2015-02-10 | Rockstar Consortium Us Lp | Speech recognition in automated information services systems |
DE602004024172D1 (en) * | 2004-05-21 | 2009-12-31 | Harman Becker Automotive Sys | Automatic generation of a word pronunciation for speech recognition |
JP4705023B2 (en) * | 2004-06-10 | 2011-06-22 | パナソニック株式会社 | Speech recognition apparatus, speech recognition method, and program |
US20060095266A1 (en) * | 2004-11-01 | 2006-05-04 | Mca Nulty Megan | Roaming user profiles for speech recognition |
US8024194B2 (en) * | 2004-12-08 | 2011-09-20 | Nuance Communications, Inc. | Dynamic switching between local and remote speech rendering |
US8677377B2 (en) | 2005-09-08 | 2014-03-18 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US8032372B1 (en) | 2005-09-13 | 2011-10-04 | Escription, Inc. | Dictation selection |
US8635073B2 (en) * | 2005-09-14 | 2014-01-21 | At&T Intellectual Property I, L.P. | Wireless multimodal voice browser for wireline-based IPTV services |
US7633076B2 (en) | 2005-09-30 | 2009-12-15 | Apple Inc. | Automated response to and sensing of user activity in portable devices |
EP1858005A1 (en) * | 2006-05-19 | 2007-11-21 | Texthelp Systems Limited | Streaming speech with synchronized highlighting generated by a server |
US20090204392A1 (en) * | 2006-07-13 | 2009-08-13 | Nec Corporation | Communication terminal having speech recognition function, update support device for speech recognition dictionary thereof, and update method |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US8214208B2 (en) * | 2006-09-28 | 2012-07-03 | Reqall, Inc. | Method and system for sharing portable voice profiles |
JP4715704B2 (en) * | 2006-09-29 | 2011-07-06 | 富士通株式会社 | Speech recognition apparatus and speech recognition program |
US20080082316A1 (en) * | 2006-09-30 | 2008-04-03 | Ms. Chun Yu Tsui | Method and System for Generating, Rating, and Storing a Pronunciation Corpus |
JP4859642B2 (en) * | 2006-11-30 | 2012-01-25 | 富士通株式会社 | Voice information management device |
US20080129520A1 (en) * | 2006-12-01 | 2008-06-05 | Apple Computer, Inc. | Electronic device with enhanced audio feedback |
JP5233989B2 (en) * | 2007-03-14 | 2013-07-10 | 日本電気株式会社 | Speech recognition system, speech recognition method, and speech recognition processing program |
US8977255B2 (en) | 2007-04-03 | 2015-03-10 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
TWI336048B (en) * | 2007-05-11 | 2011-01-11 | Delta Electronics Inc | Input system for mobile search and method therefor |
US9053089B2 (en) | 2007-10-02 | 2015-06-09 | Apple Inc. | Part-of-speech tagging using latent analogy |
US8620662B2 (en) | 2007-11-20 | 2013-12-31 | Apple Inc. | Context-aware unit selection |
US10002189B2 (en) | 2007-12-20 | 2018-06-19 | Apple Inc. | Method and apparatus for searching using an active ontology |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US8099289B2 (en) * | 2008-02-13 | 2012-01-17 | Sensory, Inc. | Voice interface and search for electronic devices including bluetooth headsets and remote systems |
US8065143B2 (en) | 2008-02-22 | 2011-11-22 | Apple Inc. | Providing text input using speech data and non-speech data |
US8996376B2 (en) | 2008-04-05 | 2015-03-31 | Apple Inc. | Intelligent text-to-speech conversion |
US10496753B2 (en) | 2010-01-18 | 2019-12-03 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
JP4962416B2 (en) * | 2008-06-03 | 2012-06-27 | 日本電気株式会社 | Speech recognition system |
US8464150B2 (en) | 2008-06-07 | 2013-06-11 | Apple Inc. | Automatic language identification for dynamic text processing |
US20100030549A1 (en) | 2008-07-31 | 2010-02-04 | Lee Michael M | Mobile device having human language translation capability with positional feedback |
US8768702B2 (en) | 2008-09-05 | 2014-07-01 | Apple Inc. | Multi-tiered voice feedback in an electronic device |
US8898568B2 (en) * | 2008-09-09 | 2014-11-25 | Apple Inc. | Audio user interface |
US8583418B2 (en) | 2008-09-29 | 2013-11-12 | Apple Inc. | Systems and methods of detecting language and natural language strings for text to speech synthesis |
US8712776B2 (en) * | 2008-09-29 | 2014-04-29 | Apple Inc. | Systems and methods for selective text to speech synthesis |
US8676904B2 (en) | 2008-10-02 | 2014-03-18 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US9959870B2 (en) | 2008-12-11 | 2018-05-01 | Apple Inc. | Speech recognition involving a mobile device |
US8862252B2 (en) | 2009-01-30 | 2014-10-14 | Apple Inc. | Audio user interface for displayless electronic device |
US8380507B2 (en) | 2009-03-09 | 2013-02-19 | Apple Inc. | Systems and methods for determining the language to use for speech generated by a text to speech engine |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US10540976B2 (en) * | 2009-06-05 | 2020-01-21 | Apple Inc. | Contextual voice commands |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
US9858925B2 (en) | 2009-06-05 | 2018-01-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US20120309363A1 (en) | 2011-06-03 | 2012-12-06 | Apple Inc. | Triggering notifications associated with tasks items that represent tasks to perform |
US9431006B2 (en) | 2009-07-02 | 2016-08-30 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
US8682649B2 (en) * | 2009-11-12 | 2014-03-25 | Apple Inc. | Sentiment prediction from textual data |
US8600743B2 (en) * | 2010-01-06 | 2013-12-03 | Apple Inc. | Noise profile determination for voice-related feature |
US8311838B2 (en) | 2010-01-13 | 2012-11-13 | Apple Inc. | Devices and methods for identifying a prompt corresponding to a voice input in a sequence of prompts |
US8381107B2 (en) | 2010-01-13 | 2013-02-19 | Apple Inc. | Adaptive audio feedback system and method |
US10679605B2 (en) | 2010-01-18 | 2020-06-09 | Apple Inc. | Hands-free list-reading by intelligent automated assistant |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US10553209B2 (en) | 2010-01-18 | 2020-02-04 | Apple Inc. | Systems and methods for hands-free notification summaries |
US10705794B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
DE202011111062U1 (en) | 2010-01-25 | 2019-02-19 | Newvaluexchange Ltd. | Device and system for a digital conversation management platform |
US8682667B2 (en) | 2010-02-25 | 2014-03-25 | Apple Inc. | User profiling for selecting user specific voice input processing information |
US8713021B2 (en) | 2010-07-07 | 2014-04-29 | Apple Inc. | Unsupervised document clustering using latent semantic density analysis |
US8719006B2 (en) | 2010-08-27 | 2014-05-06 | Apple Inc. | Combined statistical and rule-based part-of-speech tagging for text-to-speech synthesis |
JP5558284B2 (en) * | 2010-09-15 | 2014-07-23 | 株式会社Nttドコモ | Speech recognition system, speech recognition method, and speech recognition program |
US20120078635A1 (en) * | 2010-09-24 | 2012-03-29 | Apple Inc. | Voice control system |
US8719014B2 (en) | 2010-09-27 | 2014-05-06 | Apple Inc. | Electronic device with text error correction based on voice recognition data |
US10762293B2 (en) | 2010-12-22 | 2020-09-01 | Apple Inc. | Using parts-of-speech tagging and named entity recognition for spelling correction |
US10515147B2 (en) | 2010-12-22 | 2019-12-24 | Apple Inc. | Using statistical language models for contextual lookup |
US9953653B2 (en) | 2011-01-07 | 2018-04-24 | Nuance Communications, Inc. | Configurable speech recognition system using multiple recognizers |
JP6317111B2 (en) | 2011-02-22 | 2018-04-25 | スピーク・ウィズ・ミー・インコーポレイテッドSpeak With Me,Inc. | Hybrid client / server speech recognition |
US8781836B2 (en) | 2011-02-22 | 2014-07-15 | Apple Inc. | Hearing assistance system for providing consistent human speech |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US10672399B2 (en) | 2011-06-03 | 2020-06-02 | Apple Inc. | Switching between text data and audio data based on a mapping |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US8812294B2 (en) | 2011-06-21 | 2014-08-19 | Apple Inc. | Translating phrases from one language into another using an order-based set of declarative rules |
US8706472B2 (en) | 2011-08-11 | 2014-04-22 | Apple Inc. | Method for disambiguating multiple readings in language conversion |
US8994660B2 (en) | 2011-08-29 | 2015-03-31 | Apple Inc. | Text correction processing |
US8762156B2 (en) | 2011-09-28 | 2014-06-24 | Apple Inc. | Speech recognition repair using contextual information |
US10134385B2 (en) | 2012-03-02 | 2018-11-20 | Apple Inc. | Systems and methods for name pronunciation |
US9483461B2 (en) | 2012-03-06 | 2016-11-01 | Apple Inc. | Handling speech synthesis of content for multiple languages |
US9280610B2 (en) | 2012-05-14 | 2016-03-08 | Apple Inc. | Crowd sourcing information to fulfill user requests |
US8775442B2 (en) | 2012-05-15 | 2014-07-08 | Apple Inc. | Semantic search using a single-source semantic model |
US10417037B2 (en) | 2012-05-15 | 2019-09-17 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
WO2013185109A2 (en) | 2012-06-08 | 2013-12-12 | Apple Inc. | Systems and methods for recognizing textual identifiers within a plurality of words |
US9721563B2 (en) | 2012-06-08 | 2017-08-01 | Apple Inc. | Name recognition system |
US9495129B2 (en) | 2012-06-29 | 2016-11-15 | Apple Inc. | Device, method, and user interface for voice-activated navigation and browsing of a document |
US9576574B2 (en) | 2012-09-10 | 2017-02-21 | Apple Inc. | Context-sensitive handling of interruptions by intelligent digital assistant |
US9547647B2 (en) | 2012-09-19 | 2017-01-17 | Apple Inc. | Voice-based media searching |
US8935167B2 (en) | 2012-09-25 | 2015-01-13 | Apple Inc. | Exemplar-based latent perceptual modeling for automatic speech recognition |
KR20150063423A (en) | 2012-10-04 | 2015-06-09 | 뉘앙스 커뮤니케이션즈, 인코포레이티드 | Improved hybrid controller for asr |
US20140136210A1 (en) * | 2012-11-14 | 2014-05-15 | At&T Intellectual Property I, L.P. | System and method for robust personalization of speech recognition |
DE112014000709B4 (en) | 2013-02-07 | 2021-12-30 | Apple Inc. | METHOD AND DEVICE FOR OPERATING A VOICE TRIGGER FOR A DIGITAL ASSISTANT |
US10642574B2 (en) | 2013-03-14 | 2020-05-05 | Apple Inc. | Device, method, and graphical user interface for outputting captions |
US10572476B2 (en) | 2013-03-14 | 2020-02-25 | Apple Inc. | Refining a search based on schedule items |
US9977779B2 (en) | 2013-03-14 | 2018-05-22 | Apple Inc. | Automatic supplementation of word correction dictionaries |
US9368114B2 (en) | 2013-03-14 | 2016-06-14 | Apple Inc. | Context-sensitive handling of interruptions |
US10652394B2 (en) | 2013-03-14 | 2020-05-12 | Apple Inc. | System and method for processing voicemail |
US9733821B2 (en) | 2013-03-14 | 2017-08-15 | Apple Inc. | Voice control to diagnose inadvertent activation of accessibility features |
CN105190607B (en) | 2013-03-15 | 2018-11-30 | 苹果公司 | Pass through the user training of intelligent digital assistant |
WO2014144579A1 (en) | 2013-03-15 | 2014-09-18 | Apple Inc. | System and method for updating an adaptive speech recognition model |
US9922642B2 (en) | 2013-03-15 | 2018-03-20 | Apple Inc. | Training an at least partial voice command system |
US10078487B2 (en) | 2013-03-15 | 2018-09-18 | Apple Inc. | Context-sensitive handling of interruptions |
US10748529B1 (en) | 2013-03-15 | 2020-08-18 | Apple Inc. | Voice activated device for use with a voice-based digital assistant |
WO2014197334A2 (en) | 2013-06-07 | 2014-12-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
WO2014197336A1 (en) | 2013-06-07 | 2014-12-11 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
WO2014197335A1 (en) | 2013-06-08 | 2014-12-11 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
EP3008641A1 (en) | 2013-06-09 | 2016-04-20 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
WO2014200731A1 (en) | 2013-06-13 | 2014-12-18 | Apple Inc. | System and method for emergency calls initiated by voice command |
KR101749009B1 (en) | 2013-08-06 | 2017-06-19 | 애플 인크. | Auto-activating smart responses based on activities from remote devices |
DE102013216427B4 (en) * | 2013-08-20 | 2023-02-02 | Bayerische Motoren Werke Aktiengesellschaft | Device and method for means of transport-based speech processing |
US9218804B2 (en) * | 2013-09-12 | 2015-12-22 | At&T Intellectual Property I, L.P. | System and method for distributed voice models across cloud and device for embedded text-to-speech |
DE102013219649A1 (en) * | 2013-09-27 | 2015-04-02 | Continental Automotive Gmbh | Method and system for creating or supplementing a user-specific language model in a local data memory connectable to a terminal |
US10296160B2 (en) | 2013-12-06 | 2019-05-21 | Apple Inc. | Method for extracting salient dialog usage from live data |
US9620105B2 (en) | 2014-05-15 | 2017-04-11 | Apple Inc. | Analyzing audio input for efficient speech and music recognition |
US10592095B2 (en) | 2014-05-23 | 2020-03-17 | Apple Inc. | Instantaneous speaking of content on touch devices |
US9502031B2 (en) | 2014-05-27 | 2016-11-22 | Apple Inc. | Method for supporting dynamic grammars in WFST-based ASR |
US10078631B2 (en) | 2014-05-30 | 2018-09-18 | Apple Inc. | Entropy-guided text prediction using combined word and character n-gram language models |
US9842101B2 (en) | 2014-05-30 | 2017-12-12 | Apple Inc. | Predictive conversion of language input |
US9734193B2 (en) | 2014-05-30 | 2017-08-15 | Apple Inc. | Determining domain salience ranking from ambiguous words in natural speech |
AU2015266863B2 (en) | 2014-05-30 | 2018-03-15 | Apple Inc. | Multi-command single utterance input method |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
US10289433B2 (en) | 2014-05-30 | 2019-05-14 | Apple Inc. | Domain specific language for encoding assistant dialog |
US9785630B2 (en) | 2014-05-30 | 2017-10-10 | Apple Inc. | Text prediction using combined word N-gram and unigram language models |
US9760559B2 (en) | 2014-05-30 | 2017-09-12 | Apple Inc. | Predictive text input |
US10659851B2 (en) | 2014-06-30 | 2020-05-19 | Apple Inc. | Real-time digital assistant knowledge updates |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10446141B2 (en) | 2014-08-28 | 2019-10-15 | Apple Inc. | Automatic speech recognition based on user feedback |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US10789041B2 (en) | 2014-09-12 | 2020-09-29 | Apple Inc. | Dynamic thresholds for always listening speech trigger |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US9646609B2 (en) | 2014-09-30 | 2017-05-09 | Apple Inc. | Caching apparatus for serving phonetic pronunciations |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US9886432B2 (en) | 2014-09-30 | 2018-02-06 | Apple Inc. | Parsimonious handling of word inflection via categorical stem + suffix N-gram language models |
US10552013B2 (en) | 2014-12-02 | 2020-02-04 | Apple Inc. | Data detection |
US9711141B2 (en) | 2014-12-09 | 2017-07-18 | Apple Inc. | Disambiguating heteronyms in speech synthesis |
KR102325724B1 (en) * | 2015-02-28 | 2021-11-15 | 삼성전자주식회사 | Synchronization of Text Data among a plurality of Devices |
US9865280B2 (en) | 2015-03-06 | 2018-01-09 | Apple Inc. | Structured dictation using intelligent automated assistants |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US9899019B2 (en) | 2015-03-18 | 2018-02-20 | Apple Inc. | Systems and methods for structured stem and suffix language models |
US9842105B2 (en) | 2015-04-16 | 2017-12-12 | Apple Inc. | Parsimonious continuous-space phrase representations for natural language processing |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10127220B2 (en) | 2015-06-04 | 2018-11-13 | Apple Inc. | Language identification from short strings |
US9578173B2 (en) | 2015-06-05 | 2017-02-21 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US10101822B2 (en) | 2015-06-05 | 2018-10-16 | Apple Inc. | Language input correction |
US10255907B2 (en) | 2015-06-07 | 2019-04-09 | Apple Inc. | Automatic accent detection using acoustic models |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US10186254B2 (en) | 2015-06-07 | 2019-01-22 | Apple Inc. | Context-based endpoint detection |
DE102015211101A1 (en) * | 2015-06-17 | 2016-12-22 | Volkswagen Aktiengesellschaft | Speech recognition system and method for operating a speech recognition system with a mobile unit and an external server |
GB201511887D0 (en) | 2015-07-07 | 2015-08-19 | Touchtype Ltd | Improved artificial neural network for language modelling and prediction |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
CN108604179A (en) | 2016-05-10 | 2018-09-28 | 谷歌有限责任公司 | The realization of voice assistant in equipment |
KR102177786B1 (en) | 2016-05-13 | 2020-11-12 | 구글 엘엘씨 | Media transfer between media output devices |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
DK179309B1 (en) | 2016-06-09 | 2018-04-23 | Apple Inc | Intelligent automated assistant in a home environment |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US10586535B2 (en) | 2016-06-10 | 2020-03-10 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
DK179343B1 (en) | 2016-06-11 | 2018-05-14 | Apple Inc | Intelligent task discovery |
DK179415B1 (en) | 2016-06-11 | 2018-06-14 | Apple Inc | Intelligent device arbitration and control |
DK179049B1 (en) | 2016-06-11 | 2017-09-18 | Apple Inc | Data driven natural language event detection and classification |
DK201670540A1 (en) | 2016-06-11 | 2018-01-08 | Apple Inc | Application integration with a digital assistant |
US10043516B2 (en) | 2016-09-23 | 2018-08-07 | Apple Inc. | Intelligent automated assistant |
US11205110B2 (en) | 2016-10-24 | 2021-12-21 | Microsoft Technology Licensing, Llc | Device/server deployment of neural network data entry system |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
US10971157B2 (en) | 2017-01-11 | 2021-04-06 | Nuance Communications, Inc. | Methods and apparatus for hybrid speech recognition processing |
JP6833203B2 (en) * | 2017-02-15 | 2021-02-24 | フォルシアクラリオン・エレクトロニクス株式会社 | Voice recognition system, voice recognition server, terminal device, and phrase management method |
DK201770439A1 (en) | 2017-05-11 | 2018-12-13 | Apple Inc. | Offline personal assistant |
JP6883471B2 (en) * | 2017-05-11 | 2021-06-09 | オリンパス株式会社 | Sound collecting device, sound collecting method, sound collecting program, dictation method and information processing device |
DK179496B1 (en) | 2017-05-12 | 2019-01-15 | Apple Inc. | USER-SPECIFIC Acoustic Models |
DK179745B1 (en) | 2017-05-12 | 2019-05-01 | Apple Inc. | SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT |
DK201770432A1 (en) | 2017-05-15 | 2018-12-21 | Apple Inc. | Hierarchical belief states for digital assistants |
DK201770431A1 (en) | 2017-05-15 | 2018-12-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
DK179560B1 (en) | 2017-05-16 | 2019-02-18 | Apple Inc. | Far-field extension for digital assistant services |
US10515637B1 (en) * | 2017-09-19 | 2019-12-24 | Amazon Technologies, Inc. | Dynamic speech processing |
US11170762B2 (en) * | 2018-01-04 | 2021-11-09 | Google Llc | Learning offline voice commands based on usage of online voice commands |
JP2021156907A (en) * | 2018-06-15 | 2021-10-07 | ソニーグループ株式会社 | Information processor and information processing method |
US11552966B2 (en) | 2020-09-25 | 2023-01-10 | International Business Machines Corporation | Generating and mutually maturing a knowledge corpus |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB8702910D0 (en) * | 1987-02-10 | 1987-03-18 | British Telecomm | Multi-user speech recognition system |
JP2989211B2 (en) * | 1990-03-26 | 1999-12-13 | 株式会社リコー | Dictionary control method for speech recognition device |
US5165095A (en) | 1990-09-28 | 1992-11-17 | Texas Instruments Incorporated | Voice telephone dialing |
US5425128A (en) * | 1992-05-29 | 1995-06-13 | Sunquest Information Systems, Inc. | Automatic management system for speech recognition processes |
US5613036A (en) * | 1992-12-31 | 1997-03-18 | Apple Computer, Inc. | Dynamic categories for a speech recognition system |
US5774628A (en) * | 1995-04-10 | 1998-06-30 | Texas Instruments Incorporated | Speaker-independent dynamic vocabulary and grammar in speech recognition |
US5765132A (en) * | 1995-10-26 | 1998-06-09 | Dragon Systems, Inc. | Building speech models for new words in a multi-word utterance |
US5905773A (en) * | 1996-03-28 | 1999-05-18 | Northern Telecom Limited | Apparatus and method for reducing speech recognition vocabulary perplexity and dynamically selecting acoustic models |
US5819220A (en) * | 1996-09-30 | 1998-10-06 | Hewlett-Packard Company | Web triggered word set boosting for speech interfaces to the world wide web |
-
1998
- 1998-03-27 US US09/049,736 patent/US6195641B1/en not_active Expired - Lifetime
-
1999
- 1999-02-09 TW TW088101967A patent/TW405109B/en active
- 1999-02-15 EP EP99301097A patent/EP0945851B1/en not_active Expired - Lifetime
- 1999-02-15 DE DE69917112T patent/DE69917112T2/en not_active Expired - Lifetime
- 1999-03-15 KR KR1019990008566A patent/KR100324079B1/en not_active IP Right Cessation
- 1999-03-24 JP JP07907299A patent/JP3519015B2/en not_active Expired - Fee Related
Non-Patent Citations (2)
Title |
---|
"CLIENT-SERVER MODEL FOR SPEECH RECOGNITION" IBM TECHNICAL DISCLOSURE BULLETIN,US,IBM CORP. NEW YORK, vol. 36, no. 3, page 25-26 XP000354688 ISSN: 0018-8689 * |
"SPECIALIZED LANGUAGE MODELS FOR SPEECH RECOGNITION" IBM TECHNICAL DISCLOSURE BULLETIN,US,IBM CORP. NEW YORK, vol. 38, no. 2, page 155-157 XP000502428 ISSN: 0018-8689 * |
Cited By (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2356318A (en) * | 1999-09-04 | 2001-05-16 | Marconi Comm Ltd | Server with voice recognition |
EP1209662A3 (en) * | 2000-11-27 | 2004-01-28 | Canon Kabushiki Kaisha | Client-server based speech recognition |
EP1209662A2 (en) * | 2000-11-27 | 2002-05-29 | Canon Kabushiki Kaisha | Client-server based speech recognition |
US7099824B2 (en) | 2000-11-27 | 2006-08-29 | Canon Kabushiki Kaisha | Speech recognition system, speech recognition server, speech recognition client, their control method, and computer readable memory |
EP1215661A1 (en) * | 2000-12-14 | 2002-06-19 | TELEFONAKTIEBOLAGET L M ERICSSON (publ) | Mobile terminal controllable by spoken utterances |
WO2002049005A2 (en) * | 2000-12-14 | 2002-06-20 | Telefonaktiebolaget Lm Ericsson (Publ) | Mobile terminal controllable by spoken utterances |
WO2002049005A3 (en) * | 2000-12-14 | 2002-08-15 | Ericsson Telefon Ab L M | Mobile terminal controllable by spoken utterances |
WO2002059874A2 (en) * | 2001-01-05 | 2002-08-01 | Qualcomm Incorporated | System and method for voice recognition in a distributed voice recognition system |
WO2002059874A3 (en) * | 2001-01-05 | 2002-12-19 | Qualcomm Inc | System and method for voice recognition in a distributed voice recognition system |
WO2002065454A1 (en) * | 2001-02-13 | 2002-08-22 | Thomson Licensing Sa | Method, module, device and server for voice recognition |
FR2820872A1 (en) * | 2001-02-13 | 2002-08-16 | Thomson Multimedia Sa | VOICE RECOGNITION METHOD, MODULE, DEVICE AND SERVER |
US7983911B2 (en) | 2001-02-13 | 2011-07-19 | Thomson Licensing | Method, module, device and server for voice recognition |
EP1665229A2 (en) * | 2003-09-17 | 2006-06-07 | Motorola, Inc. | Speech synthesis |
EP1665229A4 (en) * | 2003-09-17 | 2007-04-25 | Motorola Inc | Speech synthesis |
WO2006084144A3 (en) * | 2005-02-03 | 2006-11-30 | Voice Signal Technologies Inc | Methods and apparatus for automatically extending the voice-recognizer vocabulary of mobile communications devices |
US8160884B2 (en) | 2005-02-03 | 2012-04-17 | Voice Signal Technologies, Inc. | Methods and apparatus for automatically extending the voice vocabulary of mobile communications devices |
WO2006097598A1 (en) * | 2005-03-16 | 2006-09-21 | France Telecom | Method for automatically producing voice labels in an address book |
US7542904B2 (en) | 2005-08-19 | 2009-06-02 | Cisco Technology, Inc. | System and method for maintaining a speech-recognition grammar |
CN101194305B (en) * | 2005-08-19 | 2012-02-29 | 思科技术公司 | System and method for maintaining a speech-recognition grammar |
WO2007024427A1 (en) | 2005-08-19 | 2007-03-01 | Cisco Technology, Inc. | System and method for distributing a speech-recognition grammar |
WO2016039945A1 (en) * | 2014-09-08 | 2016-03-17 | Qualcomm Incorporated | Keyword detection using speaker-independent keyword models for user-designated keywords |
CN106663430A (en) * | 2014-09-08 | 2017-05-10 | 高通股份有限公司 | Keyword detection using speaker-independent keyword models for user-designated keywords |
US9959863B2 (en) | 2014-09-08 | 2018-05-01 | Qualcomm Incorporated | Keyword detection using speaker-independent keyword models for user-designated keywords |
CN106663430B (en) * | 2014-09-08 | 2021-02-26 | 高通股份有限公司 | Keyword detection for speaker-independent keyword models using user-specified keywords |
Also Published As
Publication number | Publication date |
---|---|
DE69917112D1 (en) | 2004-06-17 |
KR100324079B1 (en) | 2002-02-16 |
JPH11327583A (en) | 1999-11-26 |
EP0945851B1 (en) | 2004-05-12 |
TW405109B (en) | 2000-09-11 |
KR19990077872A (en) | 1999-10-25 |
DE69917112T2 (en) | 2006-03-30 |
JP3519015B2 (en) | 2004-04-12 |
EP0945851A3 (en) | 2000-01-12 |
US6195641B1 (en) | 2001-02-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6195641B1 (en) | Network universal spoken language vocabulary | |
US10410627B2 (en) | Automatic language model update | |
JP4267081B2 (en) | Pattern recognition registration in distributed systems | |
US9317501B2 (en) | Data security system for natural language translation | |
US20190370398A1 (en) | Method and apparatus for searching historical data | |
US6937986B2 (en) | Automatic dynamic speech recognition vocabulary based on external sources of information | |
US10650821B1 (en) | Tailoring an interactive dialog application based on creator provided content | |
US7386449B2 (en) | Knowledge-based flexible natural speech dialogue system | |
US20050171944A1 (en) | Translator database | |
US20080114747A1 (en) | Speech interface for search engines | |
CN105190614A (en) | Search results using intonation nuances | |
US20230177272A1 (en) | Location-Based Mode(s) For Biasing Provisioning Of Content When An Automated Assistant Is Responding To Condensed Natural Language Inputs | |
US20200342018A1 (en) | Determining responsive content for a compound query based on a set of generated sub-queries | |
RU2628202C1 (en) | Adaptive context-thematic machine translation | |
IL147276A (en) | Telephone network-based method and system for automatic insertion of enhanced personal address book contact data | |
US20190116260A1 (en) | Voice recognition-based dialing | |
US11961507B2 (en) | Systems and methods for improving content discovery in response to a voice query using a recognition rate which depends on detected trigger terms | |
WO2010142422A1 (en) | A method for inter-lingual electronic communication | |
JP2023162822A (en) | Information processing method, information processing system, and program | |
US20060053100A1 (en) | Rapid search system | |
Hastie et al. | An Adaptive Speech Recognizer that Learns using Short and Long Term Memory |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
AK | Designated contracting states |
Kind code of ref document: A2 Designated state(s): DE FR GB NL |
|
AX | Request for extension of the european patent |
Free format text: AL;LT;LV;MK;RO;SI |
|
PUAL | Search report despatched |
Free format text: ORIGINAL CODE: 0009013 |
|
AK | Designated contracting states |
Kind code of ref document: A3 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE |
|
AX | Request for extension of the european patent |
Free format text: AL;LT;LV;MK;RO;SI |
|
17P | Request for examination filed |
Effective date: 20000628 |
|
AKX | Designation fees paid |
Free format text: DE FR GB NL |
|
RIC1 | Information provided on ipc code assigned before grant |
Ipc: 7G 10L 15/28 A |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
RIC1 | Information provided on ipc code assigned before grant |
Ipc: 7G 10L 15/28 A |
|
GRAS | Grant fee paid |
Free format text: ORIGINAL CODE: EPIDOSNIGR3 |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): DE FR GB NL |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: NL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20040512 |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: FG4D |
|
REF | Corresponds to: |
Ref document number: 69917112 Country of ref document: DE Date of ref document: 20040617 Kind code of ref document: P |
|
NLV1 | Nl: lapsed or annulled due to failure to fulfill the requirements of art. 29p and 29m of the patents act | ||
ET | Fr: translation filed | ||
PLBE | No opposition filed within time limit |
Free format text: ORIGINAL CODE: 0009261 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT |
|
26N | No opposition filed |
Effective date: 20050215 |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: 746 Effective date: 20080117 |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: TP |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: 732E Free format text: REGISTERED BETWEEN 20100617 AND 20100623 |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: S47 Free format text: CANCELLATION OF ENTRY; APPLICATION BY FILING PATENTS FORM 15 WITHIN 4 WEEKS FROM THE DATE OF PUBLICATION OF THIS JOURNAL NUANCE COMMUNICATIONS, INC EXTENDING THE VOCABULARY OF A CLIENT-SERVER SPEECH RECOGNITION SYSTEM |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: S47 Free format text: ENTRY CANCELLED; NOTICE IS HEREBY GIVEN THAT THE ENTRY ON THE REGISTER 'LICENCES OF RIGHT' UPON THE UNDER MENTIONED PATENT WAS CANCELLED ON 9 FEBRUARY 2011EXTENDING THE VOCABULARY OF A CLIENT-SERVER SPEECH RECOGNITION SYSTEM |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: PLFP Year of fee payment: 18 |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: PLFP Year of fee payment: 19 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R409 Ref document number: 69917112 Country of ref document: DE Ref country code: DE Ref legal event code: R119 Ref document number: 69917112 Country of ref document: DE |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R409 Ref document number: 69917112 Country of ref document: DE |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: PLFP Year of fee payment: 20 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: GB Payment date: 20180227 Year of fee payment: 20 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: FR Payment date: 20180226 Year of fee payment: 20 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: DE Payment date: 20180430 Year of fee payment: 20 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R071 Ref document number: 69917112 Country of ref document: DE |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: PE20 Expiry date: 20190214 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GB Free format text: LAPSE BECAUSE OF EXPIRATION OF PROTECTION Effective date: 20190214 |