JPH1031497A - Voice conversation control method and voice conversation system - Google Patents

Voice conversation control method and voice conversation system

Info

Publication number
JPH1031497A
JPH1031497A JP8189060A JP18906096A JPH1031497A JP H1031497 A JPH1031497 A JP H1031497A JP 8189060 A JP8189060 A JP 8189060A JP 18906096 A JP18906096 A JP 18906096A JP H1031497 A JPH1031497 A JP H1031497A
Authority
JP
Japan
Prior art keywords
dialogue
keyword
response
recognition
word
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
JP8189060A
Other languages
Japanese (ja)
Other versions
JP3700266B2 (en
Inventor
Toshiyuki Odaka
俊之 小高
Zuhaeru Toraberushi
ズハエル トラベルシ
Akio Amano
明雄 天野
Nobuo Hataoka
信夫 畑岡
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hitachi Ltd
Original Assignee
Hitachi Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hitachi Ltd filed Critical Hitachi Ltd
Priority to JP18906096A priority Critical patent/JP3700266B2/en
Publication of JPH1031497A publication Critical patent/JPH1031497A/en
Application granted granted Critical
Publication of JP3700266B2 publication Critical patent/JP3700266B2/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Abstract

PROBLEM TO BE SOLVED: To efficiently conduct voice conversation between a user and a system by eliminating the system output which is for confirmation only and providing the recognition results including a neat guidance sentence. SOLUTION: A conversation control means 5 requests the guidance contents which urge the next action to a task control means 10. The response to the request is a question for a name. Then, the means 5 asks for a latest keyword to a keyword holding means and obtains the 'Material Section' which has been just POPed. Then, the means 5 transmits the next guidance contents and the latest keyword to an response generating means 7 and instructs it to generate a response sentence. Then, the means 7 generates a question sentence, such as 'What is your name in the Material Section?' and transmits the sentence to a voice output means 5. Then, the means 5 sends the recognition vocabulary to a recognition vocabulary supplementing means 3 in order to confirm the name.

Description

【発明の詳細な説明】DETAILED DESCRIPTION OF THE INVENTION

【0001】[0001]

【発明の属する技術分野】本発明は、情報検索などを行
なうために利用する計算機システムに係り、特に、マイ
クとスピーカあるいは電話などの音声入出力インタフェ
ースを備え、誰でも容易に利用することができる音声対
話システムに関するものである。
BACKGROUND OF THE INVENTION 1. Field of the Invention The present invention relates to a computer system used for performing information retrieval and the like, and more particularly, to a computer system provided with a microphone and a voice input / output interface such as a speaker or a telephone so that anyone can use it easily. The present invention relates to a voice interaction system.

【0002】[0002]

【従来の技術】音声対話システムでは一般に、利用者の
システムに対するデータ入力において音声認識の技術を
用いている。音声によるデータ入力は、電話を通して使
う場合、あるいは車の運転中などのいわゆるハンズアイ
ビジーの状況下使う場合、非常に有効である。
2. Description of the Related Art Generally, a speech dialogue system uses a speech recognition technique for data input to a user's system. Voice data input is very effective when used over a telephone or when used in a so-called hands-eye busy situation such as driving a car.

【0003】普通は複数の項目(例えば、電話番号検索
のアプリケーションでは相手の所属と名前、チケットの
予約のアプリケーションでは月日、時間、大人または子
供、枚数等)に対してデータを入力する。しかし一方、
音声認識において100%の認識率を実現することは不
可能であり、利用者の入力に対して、入力項目毎にシス
テムの認識結果を確認することが必須である。例えば、
文または文節単位の音声認識が可能な音声対話システム
での対話例は、 システム:「相手の所属と名前をどうぞ」 利用者 :「資材課の佐藤さん」 システム:「資材課の佐藤ですね」 利用者 :「はい」 となる。この例のように、文または文節単位の音声認識
が可能なシステムで認識が正しく行われた場合は、入力
時間も短く利用効率が良くなる。しかし、文または文節
単位の認識の場合、複数のキーワード(ここでの例では
相手の所属と名前)の組合せ分だけ認識対象の種類が多
くなるばかりでなく、助詞も含めた表現のバラエティが
多様になるという点で、単語認識に比較して認識性能が
低くなる。さらに、認識を間違えた場合は、次のような
問題も発生する。
Normally, data is input for a plurality of items (for example, the affiliation and name of a partner in a telephone number search application, the date and time, adults or children, the number of sheets, etc. in a ticket reservation application). But on the other hand,
It is impossible to achieve a recognition rate of 100% in voice recognition, and it is essential to check the recognition result of the system for each input item for a user input. For example,
An example of dialogue in a voice dialogue system that can recognize sentences or phrases is as follows: System: "Please select the affiliation and name of the other party" User: "Mr. Sato of the Materials Division" System: "Sato of the Materials Division" User: Yes. As in this example, when recognition is correctly performed in a system capable of performing speech recognition in units of sentences or phrases, the input time is short and the use efficiency is improved. However, in the case of recognition by sentence or phrase, not only the number of types of recognition targets increases by the combination of a plurality of keywords (in this example, the affiliation and name of the partner), but also the variety of expressions, including particles, varies. , The recognition performance is lower than that of word recognition. Further, if the recognition is wrong, the following problem also occurs.

【0004】 システム:「相手の所属と名前をどうぞ」 利用者 :「資材課の佐藤さん」 システム:「資材課の加藤ですね」 利用者 :「いいえ」 システム:「相手の所属と名前をどうぞ」 最初のシステムの質問に、所属と名前という2つのキー
ワードが含まれている。この場合、システム側でこのや
りとりだけからでは、利用者が否定したことが所属と名
前のどちらか片方だけを誤認識したことを意味するの
か、あるいは両方誤認識したことを意味するのか、が特
定できない。したがって、所属と名前の両方のキーワー
ドが正しく認識されるまで、同じ質問を繰り返すことに
なる。この場合は、時間がかかるという問題がある。ま
た別な対処方法として、利用者に誤認識された方だけ再
入力してもらう方法も考えられる。この場合は、所属か
名前かわからない発声を認識しなければならないので、
音声認識の性能に対する要求が現状で対応可能なレベル
より高くなるという問題がある上、そのように利用者の
発声を誘導するにはどうすれば良いかという、解決が非
常に困難な課題が生じる。
[0004] System: "Please choose the affiliation and name of the other party" User: "Mr. Sato of the Materials Division" System: "It's Kato of the Materials Division" User: "No" System: "Please choose the affiliation and name of the other party The first system question included two keywords: affiliation and name. In this case, the system alone can determine whether the user's rejection means that either the affiliation or the name has been misrecognized, or that both have been misrecognized. Can not. Thus, the same question is repeated until both the affiliation and name keywords are correctly recognized. In this case, there is a problem that it takes time. As another countermeasure, a method in which only the user who has been erroneously recognized may input again is also conceivable. In this case, you need to recognize utterances that do not know your affiliation or name,
In addition to the problem that the demand for the performance of speech recognition is higher than the level that can be currently dealt with, there is a very difficult problem of how to guide the utterance of the user.

【0005】これに対して、単語認識の場合は認識対象
の表現のバラエティが抑えられ、現状レベルでもほぼ満
足のいく認識性能が得られる。また、項目を1つずつ質
問し、確認するために、確実に1項目ずつのデータ入力
が行える。例えば、次のようになる。
On the other hand, in the case of word recognition, the variety of expressions to be recognized is suppressed, and almost satisfactory recognition performance can be obtained even at the current level. In addition, in order to ask and confirm items one by one, data can be input one item at a time. For example,

【0006】 システム:「相手の所属をどうぞ」 利用者 :「資材課」 システム:「資材課ですか」 利用者 :「はい」 システム:「相手の名前をどうぞ」 利用者 :「佐藤」 システム:「佐藤ですか」 利用者 :「はい」 ただし、ここに示したようにやりとりが長くなり全体に
時間がかかる傾向になるという大きな問題が残る。
[0006] System: "Please add the other party's affiliation" User: "Materials Section" System: "Is the Materials Section?" User: "Yes" System: "Please enter the name of the other party" User: "Sato" System: "Is Sato?" User: "Yes" However, as shown here, there remains a major problem that the exchange becomes longer and the whole tends to take longer.

【0007】[0007]

【発明が解決しようとする課題】上記のような従来の音
声対話システムにおいて、現状の音声認識性能と時間的
な利用効率は相反するパラメータであった。
In the above-mentioned conventional speech dialogue system, the current speech recognition performance and the temporal utilization efficiency are incompatible parameters.

【0008】本発明の目的は、現状で可能な音声認識性
能の範囲で、最も効率良くデータ入力が行え、利用者と
システムとの間で円滑な対話を実現できる音声対話シス
テムを提供することにある。
[0008] It is an object of the present invention to provide a speech dialogue system capable of inputting data most efficiently within the range of currently available speech recognition performance and realizing a smooth dialogue between a user and the system. is there.

【0009】[0009]

【課題を解決するための手段】本発明によれば、以下の
ような手段による対話制御方法および以下のような手段
を設けた音声対話システムが提供される。
According to the present invention, there is provided a dialogue control method using the following means and a voice dialogue system provided with the following means.

【0010】タスク管理手段に要求を出しその返答結果
に応じて、応答生成手段、認識語彙補足手段、キーワー
ド判定手段、およびキーワード保持手段を制御し、また
認識語彙補足手段を介して音声認識手段を、さらに応答
生成手段を介して音声合成手段を制御し、システムと利
用者の間の対話を進行させる対話制御手段の元で、前記
キーワード保持手段は、前記対話制御手段の要求に基づ
き、キーワードの保持、削除、最新のキーワードの通知
を行い、前記対話制御手段は、タスク管理手段に対話の
進行上における次のアクションを促すガイダンス内容を
要求して該ガイダンス内容を受け取り、また前記キーワ
ード保持手段に最新のキーワードを要求して該最新のキ
ーワードを受け取り、前記ガイダンス内容と前記最新の
キーワード、およびそれらを用いて応答文を生成させる
指示を応答生成手段へ通知し、前記応答生成手段は、前
記対話制御手段から受け取った指示に従い、 対話の進
行上における次のアクションを促すガイダンス文の中
に、同時に受け取った前段階の認識結果でもある最新の
キーワードを含めた応答文を生成し、音声合成手段へ出
力し、次に前記対話制御手段は、タスク管理手段に対話
の進行上における次の認識語彙を要求して該認識語彙を
受け取り、タスクに依存したキーワードからなる該認識
語彙を認識語彙補足手段へ送り、前記認識語彙補足手段
は、前記対話制御手段より受け取った認識語彙に、「取
消」「ヘルプ」「ストップ」「もう一度」「わからな
い」「任意」などのタスクから独立なコマンドを表す単
語を補足して、音声認識手段およびキーワード判定手段
に渡し、前記キーワード判定手段は、前記認識語彙補足
手段より得た補足後の認識語彙と前記音声認識手段より
得た認識結果とを比較し、該認識結果が前記タスクから
独立なコマンドか、タスクに依存したキーワードかを判
定し、その判定結果を対話制御手段へ送り、さらに次に
前記対話制御手段は、前記キーワード判定手段の判定結
果を元に、該判定結果がキーワードである場合は、それ
をキーワード保持手段に送ると共に該キーワードをタス
ク管理手段に送り、 前記判定結果がコマンドである場
合は、各コマンドに対する処理を行う。
[0010] A request is issued to the task management means, and the response generation means, the recognized vocabulary supplementing means, the keyword determination means, and the keyword holding means are controlled in accordance with the response result, and the speech recognition means is controlled via the recognized vocabulary supplementing means. Further, under the control of the speech synthesizing means via the response generating means, and the dialog control means for proceeding the dialog between the system and the user, the keyword holding means, based on the request of the dialog control means, Holding, deleting, notifying of the latest keyword, the dialog control means requests the task management means for guidance content for prompting the next action in the progress of the dialogue, receives the guidance content, and sends the guidance content to the keyword holding means. Requesting the latest keyword and receiving the latest keyword, the guidance content and the latest keyword, and An instruction to generate a response sentence using them is sent to the response generation means, and the response generation means, in accordance with the instruction received from the interaction control means, includes a guidance sentence for prompting a next action in the progress of the interaction, At the same time, a response sentence including the latest keyword, which is also the recognition result of the previous stage received, is generated and output to the speech synthesis means. Then, the dialog control means sends the task management means the next recognized vocabulary in the progress of the dialog. , Receiving the recognition vocabulary, and sending the recognition vocabulary comprising task-dependent keywords to the recognition vocabulary supplementing means. The recognition vocabulary supplementing means adds “cancel”, “cancel” to the recognition vocabulary received from the dialog control means. Supplement words that represent commands that are independent of tasks, such as help, stop, repeat, don't know, or any Handover, and the keyword determination unit compares the supplemented recognized vocabulary obtained by the recognized vocabulary supplementing unit with the recognition result obtained by the voice recognition unit, and the recognition result is a command independent of the task. Or a keyword depending on the task, and sends the result of the determination to the dialogue control means. Next, the dialogue control means determines whether the keyword is a keyword based on the determination result of the keyword determination means. Sends the keyword to the keyword holding unit and sends the keyword to the task management unit, and if the determination result is a command, performs a process for each command.

【0011】[0011]

【発明の実施の形態】以下図を用いて本発明の実施例を
説明する。
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS An embodiment of the present invention will be described below with reference to the drawings.

【0012】図1は本発明による音声対話システムの一
実施例を示すブロック図である。音声対話システムは、
あるタスクを遂行することを目的として、利用者とシス
テムが音声を使ってやりとりをするようなシステムであ
る。本発明による音声対話システムは、対話制御手段
(5)の制御の元に動作する。対話制御手段(5)の動
作については、後で詳述する。
FIG. 1 is a block diagram showing one embodiment of a voice dialogue system according to the present invention. The spoken dialogue system
It is a system in which the user and the system communicate using voice for the purpose of performing a certain task. The voice dialogue system according to the present invention operates under the control of the dialogue control means (5). The operation of the dialog control means (5) will be described later in detail.

【0013】タスク管理手段(10)は、タスクに依存
した処理を全て担当する。タスクに依存した情報(タス
ク遂行の手順、場面毎の入力待ち語彙、等)を管理した
り、対話制御手段からのタスクの進行等についての問い
合わせなどに返答したりする。
The task management means (10) is in charge of all tasks dependent processing. It manages task-dependent information (task execution procedure, input waiting vocabulary for each scene, etc.) and responds to inquiries about the progress of tasks from the dialog control means.

【0014】音声認識手段(2)は、与えられた認識語
彙の範囲内で、入力される音声(1)を認識し、認識結
果を1つの単語あるいは複数の単語の列として出力す
る。音声認識手段(2)の実現方法としては様々な手法
が考えられ、本発明はその方法を限定しない。例えば、
HMM(HiddenMarkovModel)という
確率モデルを用いる手法が扱いやすい。この方法を用い
れば、任意の認識語彙をテキストとして与えるだけで、
音声認識できる構成にすることができる。詳細は、“中
川聖一:”確率モデルによる音声認識、電子情報通信学
会、1988”他の文献に詳しいので、本明細書では省
略する。
The speech recognition means (2) recognizes the input speech (1) within the range of the given recognition vocabulary, and outputs the recognition result as one word or a sequence of a plurality of words. Various methods are conceivable as a method of realizing the voice recognition means (2), and the present invention does not limit the method. For example,
A method using a probabilistic model called HMM (Hidden Markov Model) is easy to handle. With this method, just giving any recognized vocabulary as text,
A configuration capable of voice recognition can be provided. The details are described in "Seiichi Nakagawa: Speech Recognition by Probabilistic Model, IEICE, 1988" and other documents, and are omitted in this specification.

【0015】認識語彙補足手段(3)は、対話制御手段
(5)より受け取った認識語彙に、特にタスクから独立
なコマンドを表す単語等を補足して、音声認識手段に渡
す。ここで、コマンドとしては、 「違います」「違
う」「いいえ」「取消」「キャンセル」などの否定ある
いは取消を意味する単語、 「ヘルプ」などの助言要求
を意味する単語、 「ストップ」「停止」「中止」など
のシステムの処理の停止要求を意味する単語、「もう一
度」「リピート」などの直前のシステム応答の再出力要
求を意味する単語、などがある。さらに、「わかりませ
ん」「わからない」などの不明を意味する単語、「問わ
ない」「何でも良い」「任意」などの任意を意味する単
語なども認識語彙補足手段(3)で補足され得る単語と
する。図3に認識語彙の例を示しており、人名の認識語
彙を例に取り、補足前(a)と補足後(b)の認識語彙
を示している。
The recognition vocabulary supplementing means (3) supplements the recognition vocabulary received from the dialog control means (5), especially words representing commands independent of tasks, and passes them to the speech recognition means. Here, the command is a word that means negation or cancellation, such as "no,""no,""no,""cancel," or "cancel," a word that means an advisory request, such as "help,""stop,""stop." There are words that mean a request to stop the processing of the system, such as "", "stop", and words, such as "again" and "repeat", that mean a request to re-output the immediately previous system response. Furthermore, words that can be supplemented by the recognized vocabulary supplementing means (3) include words that mean unknown, such as “I don't understand” or “I don't know”, and words that mean arbitrary, such as “I don't care”, “Anything is fine”, or “Any”. And FIG. 3 shows an example of the recognition vocabulary. The recognition vocabulary of the personal name is taken as an example, and the recognition vocabulary before (a) and after (b) the supplement is shown.

【0016】キーワード判定手段(6)は、音声認識手
段(2)より得られた認識結果がタスクから独立なコマ
ンドか、タスクに依存したキーワードかを判定し、その
結果を対話制御手段へ送る。ここで、判定結果は、例え
ば図4に示すように表現される。図4に示した例では、
1つの判定結果が、2つの値の組み合わせで表現され
る。左側が判定結果の種類を表しており、「CMD」は
コマンド、「KW」はキーワードを表す。 また、「C
MD」と組み合わされている「CANCEL」、「HE
LP」はコマンドの種類を表している。さらに「KW」
と組み合わされている“資材課”や“佐藤”は実データ
値を示している。
The keyword determining means (6) determines whether the recognition result obtained from the voice recognition means (2) is a command independent of the task or a keyword dependent on the task, and sends the result to the dialog control means. Here, the determination result is expressed, for example, as shown in FIG. In the example shown in FIG.
One determination result is represented by a combination of two values. The left side shows the type of the determination result, “CMD” represents a command, and “KW” represents a keyword. Also, "C
"CANCEL", "HE" combined with "MD"
"LP" indicates the type of command. Furthermore, "KW"
“Materials section” and “Sato” combined with are showing actual data values.

【0017】キーワード保持手段(6)は、対話制御手
段の指示に従って、渡されたキーワードをスタック形式
で保持したり、スタックに積まれているキーワードを対
話制御手段に通知したりする。
The keyword holding means (6) holds the passed keywords in a stack format or notifies the dialog control means of the keywords stored in the stack, according to the instruction of the dialog control means.

【0018】応答生成手段(7)は、対話制御手段
(5)からの指示に従い、タスクを遂行するために必要
な項目の内容(名前などの実データ値)を質問するよう
な応答文を生成する。
The response generation means (7) generates a response sentence which inquires of the contents (actual data values such as names) of the items necessary for performing the task in accordance with the instruction from the interaction control means (5). I do.

【0019】音声合成手段(8)は、応答生成手段
(7)から得られる応答文を音声波形に変換してスピー
カなどの電気信号から音波へ変換するデバイスにより空
間中を伝播する音声(9)として出力する。
The voice synthesizing means (8) converts the response sentence obtained from the response generating means (7) into a voice waveform and converts the electric signal into a sound wave by a device such as a speaker. Output as

【0020】図2は対話制御手段の処理フローを示す図
である。簡単のために、本発明による音声対話システム
でデータ入力のみが行われる場合のフローを示してい
る。実際のタスクにおいては、利用者に対する結果のみ
の提示なども含まれることになる。
FIG. 2 is a diagram showing a processing flow of the dialog control means. For simplicity, the flow when only data input is performed in the voice interaction system according to the present invention is shown. In the actual task, presentation of only the result to the user is also included.

【0021】次に、図2のフローに従って、電話接続サ
ービスなどをタスクとして、所属名に‘資材課’、人名
に‘佐藤’を入力する場合を仮定して処理手順を説明す
る。
Next, the processing procedure will be described in accordance with the flow of FIG. 2 on the assumption that a telephone connection service or the like is a task and "materials section" is input as the affiliation name and "Sato" is input as the personal name.

【0022】同タスク管理手段に対話の進行上における
次のアクションを促すガイダンス内容を要求し、まずサ
ービスが開始された時点では、対話制御手段(5)は、
タスク管理手段(10)へ対話の進行上における次の
アクションを促すガイダンス内容を要求する。この要求
に対する返答は、 次のガイダンス内容が「所属の質
問」である、として得られる。次に、対話制御手段
(5)は、キーワード保持手段から最新のキーワードを
受理しようとする。システム利用開始直後は、キーワー
ド保持手段はからであり、その旨が対話制御手段へ通知
される。次に、対話制御手段(5)は、次のガイダンス
内容(「所属の質問」)と、あれば最新のキーワード
(この段階では「なし」)を応答生成手段(7)に送
り、応答文を生成するように指示する。応答生成手段
(7)では、 ガイダンス内容が「所属の質問」である
ので、例えば“所属をどうぞ”というような質問文を生
成して、音声出力手段(8)に送る。続いて、対話制御
手段(5)は、「所属」を音声認識するための認識語彙
を、認識語彙補足手段(3)へ送る。ここでの認識語彙
は、先の、次のアクションを促すガイダンス内容を要求
した際に、タスク管理手段(10)より一緒に受け取る
こととする。あるいは、この段階であらためて、タスク
管理手段に問い合わせて、受理しても良い。この後、認
識語彙補足手段(3)で補足された認識語彙は、音声認
識手段(2)に送られる。そして、利用者の発声に対す
る音声認識手段(2)よる認識結果は、キーワード判定
手段(4)を介してコマンドかキーワードかの判定が付
加され、図4に示したような形式で対話制御手段(5)
に戻される。今の仮定では、ここでの受理結果は「K
W]+「資材課」である。すなわち、対話制御手段
(5)はキーワードをキーワード保持手段(6)のスタ
ックにキーワードをPOPし、さらに、タスク管理手段
(10)に対してキーワードを通知する。この時点で、
タスク管理手段(10)は、内部でタスクの進行状態が
更新される。
The task management means is requested for guidance contents for prompting the next action in the progress of the dialog. First, when the service is started, the dialog control means (5)
It requests the task management means (10) for guidance contents for prompting the next action in the progress of the dialogue. The response to this request is obtained as the following guidance content is "affiliation question". Next, the dialog control means (5) tries to receive the latest keyword from the keyword holding means. Immediately after the start of use of the system, the keyword holding unit is empty, and the fact is notified to the dialog control unit. Next, the dialogue control means (5) sends the next guidance content ("Affiliation question") and the latest keyword ("None" at this stage) to the response generation means (7), and sends the response sentence. Instruct to generate. In the response generation means (7), since the guidance content is "question of affiliation", a question sentence such as "please belong" is generated and sent to the voice output means (8). Subsequently, the dialog control means (5) sends a recognized vocabulary for voice recognition of "belonging" to the recognized vocabulary supplementing means (3). The recognition vocabulary here is received together from the task management means (10) when the guidance content for prompting the next action is requested. Alternatively, an inquiry may be made to the task management means again at this stage to accept it. Thereafter, the recognized vocabulary supplemented by the recognized vocabulary supplementing means (3) is sent to the speech recognition means (2). Then, the recognition result of the user's utterance by the voice recognition means (2) is added with a determination as to whether it is a command or a keyword via the keyword determination means (4), and is performed in the form shown in FIG. 5)
Is returned to. According to the current assumption, the result of acceptance here is "K
W] + “Materials Section”. That is, the dialog control means (5) POPs the keyword on the stack of the keyword holding means (6) and further notifies the task management means (10) of the keyword. at this point,
The task management means (10) internally updates the task progress status.

【0023】ここで、図2のフローの最初に戻り、対話
制御手段(5)は、再びタスク管理手段(10)へ、対
話の進行上における次のアクションを促すガイダンス内
容を要求する。この要求に対する返答は、次のガイダン
ス内容が「名前の質問」であるである、として得られ
る。次に、対話制御手段(5)は、キーワード保持手段
から最新のキーワードを問い合わせ、先ほどPOPされ
たばかりの「資材課」が得られる。そして、対話管理手
段(5)は、次のガイダンス内容(「名前の質問」)と
最新のキーワード(「資材課」)を応答生成手段(7)
に送り、応答文を生成するように指示する。応答生成手
段(7)では、 「資材課」を含めて、かつ、「名前」
を問い合わせるような質問文、例えば“資材課で名前は
何ですか”とか“資材課の誰ですか”というような質問
文を生成して、音声出力手段(8)に送る。続いて、対
話制御手段(5)は、「名前」を音声認識するための認
識語彙を、認識語彙補足手段(3)へ送る。そしてま
た、音声認識手段(2)による認識結果は、キーワード
判定手段(4)を介してコマンドかキーワードかの判定
が付加され、図4に示したような形式で対話制御手段
(5)に戻される。今の仮定では、ここでの受理結果は
「KW]+「佐藤」となる。以上の繰り返しで、基本的
な対話が進行していく。
Here, returning to the beginning of the flow of FIG. 2, the dialogue control means (5) requests the task management means (10) again for guidance contents for prompting the next action in the progress of the dialogue. The answer to this request is obtained as the next guidance content is "name question". Next, the dialogue control means (5) inquires the latest keyword from the keyword holding means, and obtains the "materials section" that has just been POP. Then, the dialogue management means (5) sends the next guidance content ("question of name") and the latest keyword ("materials section") to the response generation means (7).
And instruct it to generate a response sentence. In response generation means (7), include "material section" and "name"
, Such as "What is the name in the material section?" Or "Who is the material section?", And sends it to the voice output means (8). Subsequently, the interaction control means (5) sends a recognition vocabulary for voice recognition of the "name" to the recognition vocabulary supplementing means (3). Further, the result of recognition by the voice recognition means (2) is added with a judgment as to whether it is a command or a keyword through the keyword judgment means (4), and is returned to the dialogue control means (5) in a format as shown in FIG. It is. Under the current assumption, the result of acceptance here is “KW” + “Sato”. By repeating the above, the basic dialogue proceeds.

【0024】次に、利用者の「資材課」の発声が誤認識
されて、認識結果が「施設課」になったと仮定した場合
の例を説明する。先の“資材課の誰ですか”の代わりに
“施設課の誰ですか”というシステム出力になっている
はずであり、それに対して、利用者は“違います”と答
えたとする。キーワード判定手段(4)より受け取った
結果が「KW」+「佐藤」でなく、「CMD」+「CA
NCEL」となる。この場合、対話制御手段(5)はキ
ーワード保持手段(6)に対して、キーワードを1つ
(今の場合「資材課」)POPするように指示する。さ
らに、タスク管理手段(10)に対して、POPされた
「資材課」を取り消すように通知する。ここで、対話制
御手段(5)の処理は図2のフローの最初に戻り、タス
ク管理手段(10)、キーワード保持手段(6)への問
い合わせをし、次のアクションを促すガイダンス内容が
「所属の質問」であり、最新のキーワードは「なし」で
あることを応答生成手段(7)に送り、応答生成手段
(7)で再び、“所属をどうぞ”というような質問文が
生成される。
Next, an example will be described in which it is assumed that the user's utterance of the "material section" is erroneously recognized and the recognition result is "the facility section". It is assumed that the system output is "Who is in the facility section" instead of "Who is in the material section", and the user answers "No" to the system output. The result received from the keyword determination means (4) is not "KW" + "Sato" but "CMD" + "CA
NCEL ". In this case, the dialogue control means (5) instructs the keyword holding means (6) to POP one keyword (in this case, “materials section”). Furthermore, the task management means (10) is notified to cancel the POP "material section". Here, the processing of the dialogue control means (5) returns to the beginning of the flow in FIG. 2, and inquires of the task management means (10) and the keyword holding means (6), and the guidance content for prompting the next action is "belonging Is sent to the response generating means (7), and the response generating means (7) again generates a question sentence such as "Please belong."

【0025】最後に、本発明による音声対話システムを
利用した場合の、対話例を示す。タスクを電話の接続サ
ービスとした場合の例である。
Finally, an example of a dialogue when the voice dialogue system according to the present invention is used will be described. This is an example where the task is a telephone connection service.

【0026】 システム:「相手の所属をどうぞ」 利用者 :「資材課」 システム:「資材課の誰ですか」 利用者 :「佐藤」 システム:「佐藤さんに電話をつなぎます」 次に、1番目の項目で誤認識した場合の対話例を示す。System: “Please add the other party's affiliation” User: “Materials Section” System: “Who is in the Materials Section?” User: “Sato” System: “Call Sato-san” Here is an example of a dialogue in the case of misrecognition in the second item.

【0027】 システム:「相手の所属をどうぞ」 利用者 :「資材課」 システム:「施設課の誰ですか」 利用者 :「違います」 システム:「もう一度、相手の所属をどうぞ」 利用者 :「資材課」 システム:「資材課の誰ですか」 :(以下省略) 次に、2番目以降の項目で誤認識した場合の対話例を示
す。
[0027] System: "Please go to the other party's affiliation" User: "Materials Section" System: "Who is the facility section?" User: "No," System: "Please try again to belong to the other party" User: "Materials Section" System: "Who is the Materials Section?": (Omitted hereafter) Next, an example of dialogue when misrecognition is made in the second and subsequent items is shown.

【0028】 :(途中まで省略) システム:「資材課の誰ですか」 利用者 :「佐藤」 システム:「加藤さんに電話をつなぎます」 利用者 :「違います」 システム:「もう一度、資材課の誰ですか」 利用者 :「佐藤」 システム:「佐藤さんに電話をつなぎます」 図1において、利用者とシステムとの間のメディアとし
て、音声のみしか描いていないが、文字、画像、など他
のメディアも含めた対話システムでも良い。また、ボタ
ン入力を用意し、取消などをボタン入力するようにして
も良い。遠隔地の電話機から利用している際には、取消
などにタッチトーン信号を用いても良い。
: (Some part omitted) System: "Who is in the materials section?" User: "Sato" System: "Connects the phone to Mr. Kato" User: "Is different" System: "Again, Materials section User: "Sato" System: "Connect Sato-san" In Figure 1, only audio is drawn as media between the user and the system, but characters, images, etc. A dialog system including other media may be used. Alternatively, a button input may be prepared, and a button such as a cancel button may be input. When using from a remote telephone, a touch tone signal may be used for cancellation or the like.

【0029】図1におけるタスク管理手段はタスクに特
有の処理も含んでいる。例えば、遠隔地の電話機からシ
ステムを利用する形態の場合、電話回線の制御などが考
えられる。
The task management means in FIG. 1 also includes processing specific to a task. For example, in the case of using the system from a remote telephone, control of a telephone line may be considered.

【0030】[0030]

【発明の効果】本発明によれば、複数項目のデータ入力
を音声で行う音声対話システムにおいて、音声認識を利
用する際に必須の確認に関して、確認のみのシステム出
力を省略し、認識結果を次のガイダンス文に含めて提示
することでやりとりの数を減少させることにより、 単
語音声認識程度の音声認識能力しか持たない音声対話シ
ステムでも、利用者とシステムとの間で効率良く音声対
話が進行できる効果が得られる。
According to the present invention, in a voice dialogue system for inputting data of a plurality of items by voice, a system output of only verification is omitted for a required confirmation when using voice recognition, and the recognition result is set as follows. By reducing the number of conversations by presenting it in a guidance sentence, the user and the system can proceed efficiently with a dialogue even with a spoken dialogue system that has only a speech recognition capability equivalent to word speech recognition. The effect is obtained.

【図面の簡単な説明】[Brief description of the drawings]

【図1】本発明による音声対話システムの構成の一実施
例を示すブロック図である。
FIG. 1 is a block diagram showing one embodiment of the configuration of a voice dialogue system according to the present invention.

【図2】対話制御手段の処理手順を示すフローチャート
である。
FIG. 2 is a flowchart illustrating a processing procedure of a dialog control unit.

【図3】認識語彙の一例を示す図である。FIG. 3 is a diagram showing an example of a recognized vocabulary.

【図4】キーワード判定手段から対話制御手段に渡され
るデータ構造の一例を示す図である。
FIG. 4 is a diagram showing an example of a data structure passed from a keyword determination unit to a dialogue control unit.

【図5】キーワード保持手段に保持されるデータの一例
を示す図である。
FIG. 5 is a diagram illustrating an example of data stored in a keyword storage unit.

【符号の説明】[Explanation of symbols]

1…利用者の音声、2…音声認識手段、3…認識語彙補
足手段、4…キーワード判定手段、5…対話制御手段、
6…キーワード保持手段、7…応答生成手段、8…音声
出力手段、9…システムの出力音声、10…タスク管理
手段。
DESCRIPTION OF SYMBOLS 1 ... User's voice, 2 ... Voice recognition means, 3 ... Recognition vocabulary supplement means, 4 ... Keyword judgment means, 5 ... Dialogue control means,
6 ... Keyword holding means, 7 ... Response generation means, 8 ... Speech output means, 9 ... System output speech, 10 ... Task management means.

フロントページの続き (72)発明者 畑岡 信夫 東京都国分寺市東恋ケ窪一丁目280番地 株式会社日立製作所中央研究所内Continued on the front page (72) Inventor Nobuo Hataoka 1-280 Higashi Koigakubo, Kokubunji-shi, Tokyo Inside Central Research Laboratory, Hitachi, Ltd.

Claims (18)

【特許請求の範囲】[Claims] 【請求項1】タスク管理手段の管理している対話の進行
上における次のアクションを促すガイダンス内容と、キ
ーワード保持手段が保持している前段階の認識結果であ
る最新のキーワードを用いて、 対話の進行上における
次のアクションを促すガイダンス文の中に、前記最新の
キーワードを含めた応答文を応答生成手段で生成し、出
力させると共に、認識語彙補足手段において、タスク管
理手段の管理している対話の進行上における次の場面で
の認識語彙に、タスクから独立なコマンドを表す単語を
補足し、その補足後の認識語彙の範囲内で、音声認識手
段において利用者の発話した音声を認識させ、キーワー
ド判定手段において、その認識結果が前記タスクから独
立なコマンドか、それ以外のタスクに依存したキーワー
ドかを判定し、その判定結果がコマンドである場合は、
各コマンドに対する処理を行い、さらに前記判定結果が
前記キーワードである場合は、該判定結果であるキーワ
ードを新たに最新のキーワードとしてキーワード保持手
段へ保持させると同時に、さらに該判定結果であるキー
ワードをタスク管理手段へ通知する、以上を繰り返すこ
とにより音声を進行させることを特徴とする音声対話制
御方法。
A dialogue is provided by using guidance contents for prompting a next action in the progress of a dialogue managed by a task management means and the latest keyword which is a recognition result of a previous stage held by a keyword holding means. A response sentence including the latest keyword is generated and output by the response generating unit in the guidance sentence for prompting the next action in the progress of the process, and the task is managed by the task managing unit in the recognition vocabulary supplementing unit. A word representing a command independent from the task is supplemented to the recognition vocabulary in the next scene in the progress of the dialogue, and within the range of the recognition vocabulary after the supplementation, the speech recognition means recognizes the speech uttered by the user. The keyword determination means determines whether the recognition result is a command independent of the task or a keyword dependent on other tasks. If the judgment result is a command,
The process for each command is performed, and when the result of the determination is the keyword, the keyword that is the result of the determination is newly held in the keyword holding unit as the latest keyword. A voice dialogue control method, characterized in that the voice is advanced by notifying the management means and repeating the above.
【請求項2】前記請求項1記載の音声対話制御方法にお
いて、前記コマンドは、少なくとも「違います」「違
う」「いいえ」「取消」「キャンセル」などの否定ある
いは取消を意味する単語を含み、 前記判定結果が該否
定あるいは取消を意味する単語である場合に、前記キー
ワード保持手段に最新のキーワードを1つ削除するよう
に指示し、同時に前記タスク管理手段に1つ前の要求項
目の内容が取り消されたことを通知することを特徴とす
る音声対話制御方法。
2. The method according to claim 1, wherein the command includes at least a word meaning negation or cancellation, such as "no,""no,""no,""cancel," or "cancel." If the result of the determination is a word meaning the negation or cancellation, the keyword holding means is instructed to delete one of the latest keywords, and at the same time, the content of the immediately preceding request item is instructed to the task management means. A voice dialogue control method, wherein a notification of cancellation is provided.
【請求項3】前記請求項1記載の音声対話制御方法にお
いて、前記コマンドは、少なくとも「ヘルプ」などの助
言要求を意味する単語を含み、 前記判定結果が該助言
要求を意味する単語である場合に、応答生成手段におい
て、使い方についてのメッセージ文を出力させることを
特徴とする音声対話制御方法。
3. The voice dialogue control method according to claim 1, wherein the command includes at least a word meaning an advice request such as "help", and the determination result is a word meaning the advice request. And a response generation means for outputting a message about the usage.
【請求項4】前記請求項1記載の音声対話制御方法にお
いて、前記コマンドは、少なくとも「ストップ」「停
止」「中止」「中断」などのシステムの処理の停止要求
を意味する単語を含み、前記判定結果が該停止要求を意
味する単語である場合に、対話進行を停止することを特
徴とする音声対話制御方法。
4. The voice interaction control method according to claim 1, wherein the command includes at least a word indicating a request to stop processing of the system such as “stop”, “stop”, “stop”, “stop”, and the like. A speech dialogue control method, wherein the dialogue progress is stopped when the determination result is a word meaning the stop request.
【請求項5】前記請求項1記載の音声対話制御方法にお
いて、 前記応答生成手段は1つ以上の応答文を蓄積
し、後で参照できる手段を持ち、前記コマンドは、少な
くとも「もう一度」「リピート」などの直前のシステム
応答の再出力要求を意味する単語を含み、前記判定結果
が該再出力要求を意味する単語である場合に、応答生成
手段において直前に出力した応答文の再出力させること
を特徴とする音声対話制御方法。
5. The voice dialogue control method according to claim 1, wherein said response generation means has means for storing one or more response sentences, which can be referred to later, and wherein said command is at least “again” or “repeat”. If the determination result is a word meaning the re-output request including a word meaning a re-output request of the immediately preceding system response, the response generation means may re-output the response sent immediately before. A voice interaction control method characterized by the following.
【請求項6】前記請求項1記載の音声対話制御方法にお
いて、前記コマンドは、少なくとも「わかりません」
「わからない」などの不明を意味する単語を含み、前記
判定結果が該不明を意味する単語である場合に、前記タ
スク管理手段に1つ前の要求項目の内容が不明であるこ
とを通知し、同時に同タスク管理手段に対話の進行上に
おける次のアクションを促すガイダンス内容を要求し、
その返答内容に従って対話進行することを特徴とする音
声対話制御方法。
6. The method according to claim 1, wherein the command is at least "I do not know."
Including a word meaning unknown such as "I don't know", and when the determination result is a word meaning unknown, notifies the task management means that the content of the immediately preceding request item is unknown, At the same time, request the task management means for guidance that prompts the next action in the progress of the dialogue,
A spoken dialogue control method characterized in that the dialogue proceeds according to the contents of the response.
【請求項7】前記請求項1記載の音声対話制御方法にお
いて、前記コマンドは、少なくとも「問わない」「何で
も良い」「任意」などの任意を意味する単語を含み、前
記判定結果が該任意を意味する単語である場合に、タス
ク管理手段に要求項目の内容が任意であることを通知
し、同時に同タスク管理手段に対話の進行上における次
のアクションを促すガイダンス内容を要求し、 その返
答内容に従って対話進行することを特徴とする音声対話
制御方法。
7. The voice interaction control method according to claim 1, wherein said command includes at least a word meaning an arbitrary word such as "don't care", "anything", "any", and said judgment result indicates that said arbitrary word is arbitrary. If it is a meaningful word, notify the task management means that the content of the request item is optional, and at the same time, request the task management means for guidance content that prompts the next action in the progress of the dialogue, and the response content A dialogue control method characterized in that the dialogue proceeds according to the following.
【請求項8】タスクに依存した対話の進行状況を管理
し、対話の進行上における次のアクションを促すガイダ
ンス内容のように、タスクに依存した対話の進行状況に
関連した情報提供の要求を受け、その要求に対する情報
を返答するタスク管理手段と、前記タスク管理手段に要
求を出しその返答結果に応じて、認識語彙補足手段、キ
ーワード判定手段、キーワード保持手段、および応答生
成手段を制御し、システムと利用者の間の対話を進行さ
せる対話制御手段と、前記対話制御手段より受け取った
認識語彙に新たな単語を補足して、補足した結果を音声
認識手段へ渡す認識語彙補足手段と、該認識語彙補足手
段の出力により指定された認識語彙の範囲内で、利用者
の発話した音声を認識して一つまたは複数の単語系列を
出力する音声認識手段と、前記一つまたは複数の単語系
列を判定するキーワード判定手段と、前記対話制御手段
より渡されたキーワードを保持したり、保持しているキ
ーワードの中で最新のキーワードを削除、あるいは対話
制御手段に通知したりするキーワード保持手段と、前記
対話制御手段からの指示に従い、場合によっては前記対
話制御手段からのデータを用いて、応答文を生成する応
答生成手段と、該応答生成手段から得られる応答文を音
声波形に変換して出力する音声合成手段とから構成さ
れ、特に、前記対話制御手段は、タスク管理手段に対話
の進行上における次のアクションを促すガイダンス内容
を要求して該ガイダンス内容を受け取り、また前記キー
ワード保持手段に最新のキーワードを要求して該最新の
キーワードを受け取り、前記ガイダンス内容と前記最新
のキーワード、およびそれらを用いて応答文を生成させ
る指示を応答生成手段へ通知し、前記応答生成手段は、
前記対話制御手段から受け取った指示に従い、 対話の
進行上における次のアクションを促すガイダンス文の中
に、同時に受け取った前段階の認識結果でもある最新の
キーワードを含めた応答文を生成し、音声合成手段へ出
力し、次に前記対話制御手段は、タスク管理手段に対話
の進行上における次の認識語彙を要求して該認識語彙を
受け取り、タスクに依存したキーワードからなる該認識
語彙を認識語彙補足手段へ送り、前記認識語彙補足手段
は、前記対話制御手段から受け取った認識語彙に、タス
クから独立なコマンドを表す単語を補足して、前記音声
認識手段およびキーワード判定手段に渡し、前記キーワ
ード判定手段は、前記認識語彙補足手段より得た補足後
の認識語彙と前記音声認識手段より得た認識結果とを比
較し、該認識結果が前記タスクから独立なコマンドか、
タスクに依存したキーワードかを判定し、その判定結果
を対話制御手段へ送り、さらに次に前記対話制御手段
は、前記キーワード判定手段の判定結果を元に、該判定
結果がコマンドである場合は、各コマンドに対する処理
を行い、前記判定結果がキーワードである場合は、該判
定結果であるキーワードをキーワード保持手段に送ると
同時に、該キーワードをタスク管理手段に送ることを特
徴とする音声対話システム。
8. A request for providing information related to the progress of a task-dependent dialogue, such as guidance content for managing the progress of a task-dependent dialogue and prompting the next action in the progress of the dialogue. A task management means for responding information to the request, a recognition vocabulary supplementing means, a keyword determination means, a keyword holding means, and a response generation means in response to a request issued to the task management means and the response result. Dialogue control means for promoting a dialogue between a user and a user; a recognition vocabulary supplementary means for supplementing a new word to the recognition vocabulary received from the dialogue control means and passing the supplemented result to a speech recognition means; A speech recognition method for recognizing a speech uttered by a user and outputting one or more word sequences within the range of the recognition vocabulary specified by the output of the vocabulary supplementing means. And a keyword determining unit for determining the one or more word sequences; holding a keyword passed from the dialog control unit; deleting a latest keyword from the held keywords; or a dialog control unit A response generating means for generating a response sentence using data from the dialog control means in accordance with an instruction from the dialog control means, and in some cases, obtained from the response generating means. And speech synthesis means for converting the response sentence into a speech waveform and outputting the speech. In particular, the dialogue control means requests the task management means for guidance content for prompting the next action in the progress of the dialogue. Requesting the latest keyword from the keyword holding means and receiving the latest keyword, Wherein the volume latest keywords, and notifies the response generation means an instruction to generate a response sentence by using them, the response generation means,
According to the instruction received from the interaction control means, a response sentence including the latest keyword, which is also a recognition result of the previous stage received at the same time, is generated in the guidance sentence for prompting the next action in the progress of the dialogue, and speech synthesis is performed. The dialogue control means requests the task management means for the next recognition vocabulary in the course of the dialogue, receives the recognition vocabulary, and supplements the recognition vocabulary consisting of task-dependent keywords with the recognition vocabulary supplement. Means, the recognition vocabulary supplementing means supplements the recognition vocabulary received from the dialog control means with a word representing a command independent of a task, passes the word to the speech recognition means and the keyword determination means, and passes the word to the keyword recognition means. Compares the supplemented recognition vocabulary obtained by the recognition vocabulary supplementing means with the recognition result obtained by the speech recognition means, and the recognition result is A command independent of the task,
It is determined whether the keyword depends on the task, the determination result is sent to the dialog control means, and further, the dialog control means, based on the determination result of the keyword determination means, if the determination result is a command, A voice dialogue system which performs a process for each command and, when the determination result is a keyword, sends the keyword as the determination result to the keyword holding means and simultaneously sends the keyword to the task management means.
【請求項9】前記請求項2記載の音声対話システムにお
いて、前記コマンドは、少なくとも「違います」「違
う」「いいえ」「取消」「キャンセル」などの否定ある
いは取消を意味する単語を含み、 前記判定結果が該否
定あるいは取消を意味する単語である場合に、前記対話
制御手段は前記キーワード保持手段に最新のキーワード
を1つ削除するように指示し、同時に前記タスク管理手
段に1つ前の要求項目の内容が取り消されたことを通知
することを特徴とする音声対話システム。
9. The speech dialogue system according to claim 2, wherein the command includes at least a word meaning negation or cancellation such as "no,""no,""no,""cancel," or "cancel." If the determination result is a word meaning the negation or cancellation, the dialogue control means instructs the keyword holding means to delete one newest keyword, and at the same time, issues a request to the task management means. A voice dialogue system for notifying that the content of an item has been canceled.
【請求項10】前記請求項2記載の音声対話システムに
おいて、前記コマンドは、少なくとも「ヘルプ」などの
助言要求を意味する単語を含み、 前記キーワード判定
手段の判定結果が該助言要求を意味する単語である場合
に、前記対話制御手段は応答生成手段にヘルプ応答を指
示し、該指示に従い前記応答生成手段は使い方などのメ
ッセージ文を出力することを特徴とする音声対話システ
ム。
10. The voice interaction system according to claim 2, wherein said command includes at least a word meaning an advice request such as "help", and the result of judgment by said keyword judgment means is a word meaning the advice request. Wherein the dialogue control means instructs the response generation means to provide a help response, and the response generation means outputs a message such as usage in accordance with the instruction.
【請求項11】前記請求項2記載の音声対話システムに
おいて、前記コマンドは、少なくとも「ストップ」「停
止」「中止」「中断」などのシステムの処理の停止要求
を意味する単語を含み、前記キーワード判定手段の判定
結果が該停止要求を意味する単語である場合に、前記対
話制御手段は対話進行を停止することを特徴とする音声
対話システム。
11. The voice interaction system according to claim 2, wherein the command includes at least a word indicating a request to stop processing of the system, such as "stop", "stop", "stop", "stop", and the keyword. If the result of the determination by the determination means is a word meaning the stop request, the dialogue control means stops the dialogue progress.
【請求項12】前記請求項2記載の音声対話システムに
おいて、前記応答生成手段は1つ以上の応答文を蓄積
し、後で参照できる手段を持ち、前記コマンドは、少な
くとも「もう一度」「リピート」などの直前のシステム
応答の再出力要求を意味する単語を含み、前記キーワー
ド判定手段の判定結果が該再出力要求を意味する単語で
ある場合に、前記対話制御手段は応答生成手段に直前に
出力した応答文の再出力する指示を出すことを特徴とす
る音声対話システム。
12. The speech dialogue system according to claim 2, wherein said response generating means has means for storing one or more response sentences, which can be referred to later, and wherein said command is at least "again" or "repeat". When the result of the keyword determination means is a word meaning the re-output request, the dialog control means outputs the word immediately before to the response generation means. A spoken dialogue system that issues an instruction to re-output a response sentence.
【請求項13】前記請求項2記載の音声対話システムに
おいて、前記コマンドは、少なくとも「わかりません」
「わからない」などの不明を意味する単語を含み、前記
キーワード判定手段の判定結果が該不明を意味する単語
である場合に、前記対話制御手段は前記タスク管理手段
に1つ前の要求項目の内容が不明であることを通知し、
同時に同タスク管理手段に対話の進行上における次のア
クションを促すガイダンス内容を要求し、その返答内容
に従って対話進行することを特徴とする音声対話システ
ム。
13. The voice interaction system according to claim 2, wherein said command is at least "I do not understand".
If the keyword includes a word meaning unknown, such as "I don't know," and the result of the determination by the keyword determination means is a word meaning the unknown, the dialog control means sends the content of the immediately preceding request item to the task management means. Notify that is unknown,
At the same time, a voice dialogue system which requests the task management means for guidance content for prompting the next action in the progress of the dialogue, and proceeds with the dialogue according to the response content.
【請求項14】前記請求項2記載の音声対話システムに
おいて、前記コマンドは、少なくとも「問わない」「何
でも良い」「任意」などの任意を意味する単語を含み、
前記キーワード判定手段の判定結果が該任意を意味する
単語である場合に、前記対話制御手段は、タスク管理手
段に要求項目の内容が任意であることを通知し、同時に
同タスク管理手段に対話の進行上における次のアクショ
ンを促すガイダンス内容を要求し、 その返答内容に従
って対話進行することを特徴とする音声対話システム。
14. The voice dialogue system according to claim 2, wherein the command includes at least a word meaning an arbitrary word such as “don't care”, “anything”, “any”,
When the result of the determination by the keyword determination means is a word meaning the arbitraryness, the dialogue control means notifies the task management means that the content of the request item is optional, and at the same time, notifies the task management means of the dialogue. A spoken dialogue system that requests guidance content for prompting the next action in the process, and performs dialogue according to the response content.
【請求項15】請求項8、9、10、11、12、1
3、および14記載の前記タスクの内容が電話番号の検
索サービス、あるいは電話回線の接続サービスである音
声対話システム。
15. The method of claim 8, 9, 10, 11, 12, 1.
15. A voice interaction system in which the contents of the tasks described in 3 and 14 are a telephone number search service or a telephone line connection service.
【請求項16】請求項8、9、10、11、12、1
3、および14記載の前記タスクの内容が交通機関、催
し物、あるいは施設の検索サービス、あるいは予約サー
ビスである音声対話システム。
16. The method of claim 8, 9, 10, 11, 12, or 1.
15. A spoken dialogue system in which the contents of the tasks described in 3 and 14 are a transportation service, an entertainment or facility search service, or a reservation service.
【請求項17】請求項8、9、10、11、12、1
3、および14記載の前記タスクの内容が商品の検索サ
ービス、販売サービス、あるいはサポートサービスであ
る音声対話システム。
17. The method of claim 8, 9, 10, 11, 12, 1.
15. The voice interaction system according to items 3 and 14, wherein the content of the task is a product search service, a sales service, or a support service.
【請求項18】請求項8、9、10、11、12、1
3、および14記載の前記タスクの内容が経路情報提供
サービス、あるいは経路誘導サービスである音声対話シ
ステム。
18. The method of claim 8, 9, 10, 11, 12, 1.
15. A voice interaction system in which the contents of the tasks described in 3 and 14 are a route information providing service or a route guidance service.
JP18906096A 1996-07-18 1996-07-18 Spoken dialogue control method and spoken dialogue system Expired - Fee Related JP3700266B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP18906096A JP3700266B2 (en) 1996-07-18 1996-07-18 Spoken dialogue control method and spoken dialogue system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP18906096A JP3700266B2 (en) 1996-07-18 1996-07-18 Spoken dialogue control method and spoken dialogue system

Publications (2)

Publication Number Publication Date
JPH1031497A true JPH1031497A (en) 1998-02-03
JP3700266B2 JP3700266B2 (en) 2005-09-28

Family

ID=16234630

Family Applications (1)

Application Number Title Priority Date Filing Date
JP18906096A Expired - Fee Related JP3700266B2 (en) 1996-07-18 1996-07-18 Spoken dialogue control method and spoken dialogue system

Country Status (1)

Country Link
JP (1) JP3700266B2 (en)

Cited By (96)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH11126089A (en) * 1997-10-24 1999-05-11 Nissan Motor Co Ltd Voice interaction device
JP2000020089A (en) * 1998-07-07 2000-01-21 Matsushita Electric Ind Co Ltd Speed recognition method and apparatus therefor as well as voice control system
JP2002073080A (en) * 2000-09-01 2002-03-12 Fujitsu Ten Ltd Voice interactive system
US6397188B1 (en) 1998-07-29 2002-05-28 Nec Corporation Natural language dialogue system automatically continuing conversation on behalf of a user who does not respond
JP2003091297A (en) * 2001-09-19 2003-03-28 Matsushita Electric Ind Co Ltd Voice interaction device
JP2006284677A (en) * 2005-03-31 2006-10-19 Clarion Co Ltd Voice guiding device, and control method and control program for voice guiding device
JP2007047488A (en) * 2005-08-10 2007-02-22 Nippon Telegr & Teleph Corp <Ntt> Interactive method, interactive device, interactive program, and recording medium
KR101511831B1 (en) * 2010-01-18 2015-04-14 애플 인크. Active input elicitation by intelligent automated assistant
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9535906B2 (en) 2008-07-31 2017-01-03 Apple Inc. Mobile device having human language translation capability with positional feedback
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
JP2017058545A (en) * 2015-09-17 2017-03-23 本田技研工業株式会社 Voice processing device and voice processing method
US9620104B2 (en) 2013-06-07 2017-04-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9626955B2 (en) 2008-04-05 2017-04-18 Apple Inc. Intelligent text-to-speech conversion
US9633674B2 (en) 2013-06-07 2017-04-25 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
US9633660B2 (en) 2010-02-25 2017-04-25 Apple Inc. User profiling for voice input processing
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US9646614B2 (en) 2000-03-16 2017-05-09 Apple Inc. Fast, language-independent method for user authentication by voice
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US9798393B2 (en) 2011-08-29 2017-10-24 Apple Inc. Text correction processing
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9953088B2 (en) 2012-05-14 2018-04-24 Apple Inc. Crowd sourcing information to fulfill user requests
US9966068B2 (en) 2013-06-08 2018-05-08 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US9966065B2 (en) 2014-05-30 2018-05-08 Apple Inc. Multi-command single utterance input method
US9971774B2 (en) 2012-09-19 2018-05-15 Apple Inc. Voice-based media searching
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US10078631B2 (en) 2014-05-30 2018-09-18 Apple Inc. Entropy-guided text prediction using combined word and character n-gram language models
US10079014B2 (en) 2012-06-08 2018-09-18 Apple Inc. Name recognition system
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US10089072B2 (en) 2016-06-11 2018-10-02 Apple Inc. Intelligent device arbitration and control
US10101822B2 (en) 2015-06-05 2018-10-16 Apple Inc. Language input correction
US10102359B2 (en) 2011-03-21 2018-10-16 Apple Inc. Device access using voice authentication
US10127220B2 (en) 2015-06-04 2018-11-13 Apple Inc. Language identification from short strings
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US10169329B2 (en) 2014-05-30 2019-01-01 Apple Inc. Exemplar-based natural language processing
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
US10185542B2 (en) 2013-06-09 2019-01-22 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US10186254B2 (en) 2015-06-07 2019-01-22 Apple Inc. Context-based endpoint detection
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US10255907B2 (en) 2015-06-07 2019-04-09 Apple Inc. Automatic accent detection using acoustic models
US10269345B2 (en) 2016-06-11 2019-04-23 Apple Inc. Intelligent task discovery
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US10283110B2 (en) 2009-07-02 2019-05-07 Apple Inc. Methods and apparatuses for automatic speech recognition
US10297253B2 (en) 2016-06-11 2019-05-21 Apple Inc. Application integration with a digital assistant
US10318871B2 (en) 2005-09-08 2019-06-11 Apple Inc. Method and apparatus for building an intelligent automated assistant
US10354011B2 (en) 2016-06-09 2019-07-16 Apple Inc. Intelligent automated assistant in a home environment
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10496753B2 (en) 2010-01-18 2019-12-03 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
US10521466B2 (en) 2016-06-11 2019-12-31 Apple Inc. Data driven natural language event detection and classification
US10553209B2 (en) 2010-01-18 2020-02-04 Apple Inc. Systems and methods for hands-free notification summaries
US10552013B2 (en) 2014-12-02 2020-02-04 Apple Inc. Data detection
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US10568032B2 (en) 2007-04-03 2020-02-18 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
US10607141B2 (en) 2010-01-25 2020-03-31 Newvaluexchange Ltd. Apparatuses, methods and systems for a digital conversation management platform
US10659851B2 (en) 2014-06-30 2020-05-19 Apple Inc. Real-time digital assistant knowledge updates
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US10679605B2 (en) 2010-01-18 2020-06-09 Apple Inc. Hands-free list-reading by intelligent automated assistant
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US10706373B2 (en) 2011-06-03 2020-07-07 Apple Inc. Performing actions associated with task items that represent tasks to perform
US10705794B2 (en) 2010-01-18 2020-07-07 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10733993B2 (en) 2016-06-10 2020-08-04 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US10791176B2 (en) 2017-05-12 2020-09-29 Apple Inc. Synchronization and task delegation of a digital assistant
US10789041B2 (en) 2014-09-12 2020-09-29 Apple Inc. Dynamic thresholds for always listening speech trigger
US10810274B2 (en) 2017-05-15 2020-10-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification

Cited By (132)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH11126089A (en) * 1997-10-24 1999-05-11 Nissan Motor Co Ltd Voice interaction device
JP2000020089A (en) * 1998-07-07 2000-01-21 Matsushita Electric Ind Co Ltd Speed recognition method and apparatus therefor as well as voice control system
US6397188B1 (en) 1998-07-29 2002-05-28 Nec Corporation Natural language dialogue system automatically continuing conversation on behalf of a user who does not respond
US9646614B2 (en) 2000-03-16 2017-05-09 Apple Inc. Fast, language-independent method for user authentication by voice
JP2002073080A (en) * 2000-09-01 2002-03-12 Fujitsu Ten Ltd Voice interactive system
JP2003091297A (en) * 2001-09-19 2003-03-28 Matsushita Electric Ind Co Ltd Voice interaction device
JP2006284677A (en) * 2005-03-31 2006-10-19 Clarion Co Ltd Voice guiding device, and control method and control program for voice guiding device
JP4705398B2 (en) * 2005-03-31 2011-06-22 クラリオン株式会社 Voice guidance device, control method and program for voice guidance device
JP2007047488A (en) * 2005-08-10 2007-02-22 Nippon Telegr & Teleph Corp <Ntt> Interactive method, interactive device, interactive program, and recording medium
JP4589843B2 (en) * 2005-08-10 2010-12-01 日本電信電話株式会社 Dialogue method, dialogue apparatus, dialogue program, and recording medium
US10318871B2 (en) 2005-09-08 2019-06-11 Apple Inc. Method and apparatus for building an intelligent automated assistant
US10568032B2 (en) 2007-04-03 2020-02-18 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US10381016B2 (en) 2008-01-03 2019-08-13 Apple Inc. Methods and apparatus for altering audio output signals
US9865248B2 (en) 2008-04-05 2018-01-09 Apple Inc. Intelligent text-to-speech conversion
US9626955B2 (en) 2008-04-05 2017-04-18 Apple Inc. Intelligent text-to-speech conversion
US9535906B2 (en) 2008-07-31 2017-01-03 Apple Inc. Mobile device having human language translation capability with positional feedback
US10108612B2 (en) 2008-07-31 2018-10-23 Apple Inc. Mobile device having human language translation capability with positional feedback
US10795541B2 (en) 2009-06-05 2020-10-06 Apple Inc. Intelligent organization of tasks items
US11080012B2 (en) 2009-06-05 2021-08-03 Apple Inc. Interface for a virtual digital assistant
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US10475446B2 (en) 2009-06-05 2019-11-12 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US10283110B2 (en) 2009-07-02 2019-05-07 Apple Inc. Methods and apparatuses for automatic speech recognition
US10496753B2 (en) 2010-01-18 2019-12-03 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10553209B2 (en) 2010-01-18 2020-02-04 Apple Inc. Systems and methods for hands-free notification summaries
US10741185B2 (en) 2010-01-18 2020-08-11 Apple Inc. Intelligent automated assistant
US9548050B2 (en) 2010-01-18 2017-01-17 Apple Inc. Intelligent automated assistant
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US10679605B2 (en) 2010-01-18 2020-06-09 Apple Inc. Hands-free list-reading by intelligent automated assistant
US10706841B2 (en) 2010-01-18 2020-07-07 Apple Inc. Task flow identification based on user intent
US11423886B2 (en) 2010-01-18 2022-08-23 Apple Inc. Task flow identification based on user intent
US10705794B2 (en) 2010-01-18 2020-07-07 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
KR101511831B1 (en) * 2010-01-18 2015-04-14 애플 인크. Active input elicitation by intelligent automated assistant
US10607141B2 (en) 2010-01-25 2020-03-31 Newvaluexchange Ltd. Apparatuses, methods and systems for a digital conversation management platform
US10984327B2 (en) 2010-01-25 2021-04-20 New Valuexchange Ltd. Apparatuses, methods and systems for a digital conversation management platform
US10607140B2 (en) 2010-01-25 2020-03-31 Newvaluexchange Ltd. Apparatuses, methods and systems for a digital conversation management platform
US10984326B2 (en) 2010-01-25 2021-04-20 Newvaluexchange Ltd. Apparatuses, methods and systems for a digital conversation management platform
US11410053B2 (en) 2010-01-25 2022-08-09 Newvaluexchange Ltd. Apparatuses, methods and systems for a digital conversation management platform
US10049675B2 (en) 2010-02-25 2018-08-14 Apple Inc. User profiling for voice input processing
US9633660B2 (en) 2010-02-25 2017-04-25 Apple Inc. User profiling for voice input processing
US10102359B2 (en) 2011-03-21 2018-10-16 Apple Inc. Device access using voice authentication
US10706373B2 (en) 2011-06-03 2020-07-07 Apple Inc. Performing actions associated with task items that represent tasks to perform
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US11120372B2 (en) 2011-06-03 2021-09-14 Apple Inc. Performing actions associated with task items that represent tasks to perform
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US9798393B2 (en) 2011-08-29 2017-10-24 Apple Inc. Text correction processing
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US9953088B2 (en) 2012-05-14 2018-04-24 Apple Inc. Crowd sourcing information to fulfill user requests
US10079014B2 (en) 2012-06-08 2018-09-18 Apple Inc. Name recognition system
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9971774B2 (en) 2012-09-19 2018-05-15 Apple Inc. Voice-based media searching
US9633674B2 (en) 2013-06-07 2017-04-25 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
US9966060B2 (en) 2013-06-07 2018-05-08 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9620104B2 (en) 2013-06-07 2017-04-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9966068B2 (en) 2013-06-08 2018-05-08 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US10657961B2 (en) 2013-06-08 2020-05-19 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
US10185542B2 (en) 2013-06-09 2019-01-22 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US9966065B2 (en) 2014-05-30 2018-05-08 Apple Inc. Multi-command single utterance input method
US10497365B2 (en) 2014-05-30 2019-12-03 Apple Inc. Multi-command single utterance input method
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US10078631B2 (en) 2014-05-30 2018-09-18 Apple Inc. Entropy-guided text prediction using combined word and character n-gram language models
US10169329B2 (en) 2014-05-30 2019-01-01 Apple Inc. Exemplar-based natural language processing
US11133008B2 (en) 2014-05-30 2021-09-28 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US10659851B2 (en) 2014-06-30 2020-05-19 Apple Inc. Real-time digital assistant knowledge updates
US9668024B2 (en) 2014-06-30 2017-05-30 Apple Inc. Intelligent automated assistant for TV user interactions
US10904611B2 (en) 2014-06-30 2021-01-26 Apple Inc. Intelligent automated assistant for TV user interactions
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US10431204B2 (en) 2014-09-11 2019-10-01 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US10789041B2 (en) 2014-09-12 2020-09-29 Apple Inc. Dynamic thresholds for always listening speech trigger
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US9986419B2 (en) 2014-09-30 2018-05-29 Apple Inc. Social reminders
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US10552013B2 (en) 2014-12-02 2020-02-04 Apple Inc. Data detection
US11556230B2 (en) 2014-12-02 2023-01-17 Apple Inc. Data detection
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US10311871B2 (en) 2015-03-08 2019-06-04 Apple Inc. Competing devices responding to voice triggers
US11087759B2 (en) 2015-03-08 2021-08-10 Apple Inc. Virtual assistant activation
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US10127220B2 (en) 2015-06-04 2018-11-13 Apple Inc. Language identification from short strings
US10101822B2 (en) 2015-06-05 2018-10-16 Apple Inc. Language input correction
US10255907B2 (en) 2015-06-07 2019-04-09 Apple Inc. Automatic accent detection using acoustic models
US10186254B2 (en) 2015-06-07 2019-01-22 Apple Inc. Context-based endpoint detection
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US11500672B2 (en) 2015-09-08 2022-11-15 Apple Inc. Distributed personal assistant
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
JP2017058545A (en) * 2015-09-17 2017-03-23 本田技研工業株式会社 Voice processing device and voice processing method
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US11526368B2 (en) 2015-11-06 2022-12-13 Apple Inc. Intelligent automated assistant in a messaging environment
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
US11069347B2 (en) 2016-06-08 2021-07-20 Apple Inc. Intelligent automated assistant for media exploration
US10354011B2 (en) 2016-06-09 2019-07-16 Apple Inc. Intelligent automated assistant in a home environment
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
US10733993B2 (en) 2016-06-10 2020-08-04 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
US11037565B2 (en) 2016-06-10 2021-06-15 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US11152002B2 (en) 2016-06-11 2021-10-19 Apple Inc. Application integration with a digital assistant
US10297253B2 (en) 2016-06-11 2019-05-21 Apple Inc. Application integration with a digital assistant
US10269345B2 (en) 2016-06-11 2019-04-23 Apple Inc. Intelligent task discovery
US10521466B2 (en) 2016-06-11 2019-12-31 Apple Inc. Data driven natural language event detection and classification
US10089072B2 (en) 2016-06-11 2018-10-02 Apple Inc. Intelligent device arbitration and control
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
US11405466B2 (en) 2017-05-12 2022-08-02 Apple Inc. Synchronization and task delegation of a digital assistant
US10791176B2 (en) 2017-05-12 2020-09-29 Apple Inc. Synchronization and task delegation of a digital assistant
US10810274B2 (en) 2017-05-15 2020-10-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback

Also Published As

Publication number Publication date
JP3700266B2 (en) 2005-09-28

Similar Documents

Publication Publication Date Title
JPH1031497A (en) Voice conversation control method and voice conversation system
JP3454897B2 (en) Spoken dialogue system
US8064573B2 (en) Computer generated prompting
US7747438B2 (en) Multi-slot dialog systems and methods
US11869495B2 (en) Voice to voice natural language understanding processing
US9015048B2 (en) Incremental speech recognition for dialog systems
US20210193116A1 (en) Data driven dialog management
US20120253823A1 (en) Hybrid Dialog Speech Recognition for In-Vehicle Automated Interaction and In-Vehicle Interfaces Requiring Minimal Driver Processing
WO2005064592A1 (en) Device control device, speech recognition device, agent device, on-vehicle device control device, navigation device, audio device, device control method, speech recognition method, agent processing method, on-vehicle device control method, navigation method, and audio device control method, and program
JP2001005488A (en) Voice interactive system
JP2007529831A (en) Voice interactive messaging method and apparatus
Boyce Natural spoken dialogue systems for telephony applications
US20240087562A1 (en) Interactive content output
CN114860910A (en) Intelligent dialogue method and system
JP2019138989A (en) Information processor, method for processing information, and program
US10854196B1 (en) Functional prerequisites and acknowledgments
JP4304959B2 (en) Voice dialogue control method, voice dialogue control apparatus, and voice dialogue control program
KR20230007502A (en) Hotword-free preemption of automated assistant response presentations
JP2003036094A (en) Device for speech dialogue and method for processing speech dialogue
Wang et al. Cross Cultural Comparison of Users’ Barge-in with the In-Vehicle Speech System
JP7058305B2 (en) Information processing device, audio output method, audio output program
JP2003228393A (en) Device and method for voice interaction, voice interaction program and recording medium therefor
CN114268694A (en) Service request response method, device, equipment, system and medium
CN113885825A (en) Method and device for intelligently creating application form
CN113314112A (en) Dialogue method, dialogue device, electronic equipment and storage medium

Legal Events

Date Code Title Description
A131 Notification of reasons for refusal

Free format text: JAPANESE INTERMEDIATE CODE: A131

Effective date: 20040316

A131 Notification of reasons for refusal

Free format text: JAPANESE INTERMEDIATE CODE: A131

Effective date: 20050405

A521 Request for written amendment filed

Free format text: JAPANESE INTERMEDIATE CODE: A523

Effective date: 20050601

TRDD Decision of grant or rejection written
A01 Written decision to grant a patent or to grant a registration (utility model)

Free format text: JAPANESE INTERMEDIATE CODE: A01

Effective date: 20050621

A61 First payment of annual fees (during grant procedure)

Free format text: JAPANESE INTERMEDIATE CODE: A61

Effective date: 20050704

LAPS Cancellation because of no payment of annual fees