CN109063035A - A kind of man-machine more wheel dialogue methods towards trip field - Google Patents

A kind of man-machine more wheel dialogue methods towards trip field Download PDF

Info

Publication number
CN109063035A
CN109063035A CN201810779805.9A CN201810779805A CN109063035A CN 109063035 A CN109063035 A CN 109063035A CN 201810779805 A CN201810779805 A CN 201810779805A CN 109063035 A CN109063035 A CN 109063035A
Authority
CN
China
Prior art keywords
question sentence
current question
current
slot position
model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201810779805.9A
Other languages
Chinese (zh)
Other versions
CN109063035B (en
Inventor
赵铁军
郑德权
林先辉
曹海龙
朱聪慧
徐冰
杨沐昀
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin Institute of Technology
Original Assignee
Harbin Institute of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin Institute of Technology filed Critical Harbin Institute of Technology
Priority to CN201810779805.9A priority Critical patent/CN109063035B/en
Publication of CN109063035A publication Critical patent/CN109063035A/en
Application granted granted Critical
Publication of CN109063035B publication Critical patent/CN109063035B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • G06F40/35Discourse or dialogue representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/211Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
    • G06F18/24133Distances to prototypes
    • G06F18/24137Distances to cluster centroïds
    • G06F18/2414Smoothing the distance, e.g. radial basis function networks [RBFN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • G06F40/295Named entity recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N7/00Computing arrangements based on specific mathematical models
    • G06N7/01Probabilistic graphical models, e.g. probabilistic networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Probability & Statistics with Applications (AREA)
  • Algebra (AREA)
  • Computational Mathematics (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Pure & Applied Mathematics (AREA)
  • Machine Translation (AREA)

Abstract

A kind of man-machine more wheel dialogue methods towards trip field, it is used for information technology field.The present invention solves the problems, such as that current more wheel conversational systems have difficulties to the intent information of user's question sentence and slot position information extraction.The present invention carries out standardization processing to short text question sentence, and using the intent information of the short text question sentence after DAN, CNN or BLSTM model extraction standardization processing, and the model based on BLSTM achieves the ideal effect that Micro-F1 value is 93.47%;Using the slot position information of the short text question sentence after the BLSTM-CRF model extraction standardization processing for introducing intent features word, and model achieves the ideal effect that F1 value is 89.47%;The slot position information of history and the slot position information of current question sentence are determined into current dialog state information as input, and the intent information of the current question sentence of combination determines the reply strategy of next step;User is replied to according to the determining corresponding template of reply policy selection.Present invention could apply to information technology field use.

Description

A kind of man-machine more wheel dialogue methods towards trip field
Technical field
The invention belongs to interactive fields, and in particular to a kind of man-machine more wheel dialogue sides towards trip field Method.
Background technique
Interactive system is the man-machine bi-directional exchanges of information system that machine is considered as to a cognitive subject, is that realization is man-machine A kind of interactive mode.In conversational system, user is with specific purpose, it is desirable to obtain the letter for meeting specific restrictive condition Breath or service, such as: booking tickets, make a reservation, finding commodity etc..User demand is taken turns and is stated usually needs point more, by user right The demand of oneself is improved during words.
At present both at home and abroad for field of going on a journey more wheel conversational systems research, mainly include by its pipeline (be divided into from Right three language understanding, dialogue management, spatial term parts) and using based on the building of neural network model end to end Task conversational system.Wherein, it is the semantic expressiveness of structuring that natural language understanding, which is the text conversion that will be inputted, judges question sentence Intention and slot position information;Dialogue management determines current dialogue state, and determines the reply strategy of next step;Natural language is raw It is replied at according to the obtained corresponding template of reply policy selection;Although at present both at home and abroad for more wheels pair in field of going on a journey The research of telephone system makes some progress, but it must be more complete that the premise of research method application, which is user's question sentence, One question sentence.
And in actual conversational system, user's question sentence is usually to be presented in the form of short text, and short text again can not The meeting avoided causes current more wheel conversational systems to the intent information and slot of user's question sentence there are some omissions and reference phenomenon There are certain difficulties for position information extraction.
Summary of the invention
The purpose of the present invention is the intent informations and slot position information for the current more wheel conversational systems of solution to user's question sentence The problem of extraction has difficulties.
The technical solution adopted by the present invention to solve the above technical problem is:
A kind of man-machine more wheel dialogue methods towards trip field, the specific steps of this method are as follows:
Step 1: carrying out standardization processing to the current question sentence of user, then pronoun is explicitly indicated to existing in current question sentence Or the case where lacking sentence minor structure, according to the slot position information involved in being interacted before user, to the instruction in current question sentence The sentence minor structure of pronoun and missing is substituted or is filled;The current question sentence that obtains that treated;
Step 2: obtaining step 1 treated the intent information of current question sentence using DAN, CNN or BLSTM model: will Treated that current question sentence inputs DAN, CNN or BLSTM model for step 1;The output of DAN, CNN or BLSTM model is passed through Softmax operates to obtain the intention probability of current question sentence, the intent information by the intention of maximum probability as current question sentence;
Step 3: the BLSTM-CRF model of step 1 treated current question sentence input introduces intent features vocabulary is obtained To the slot position information of current question sentence;
Step 4: determining current dialogue shape using the slot position information of history and the slot position information of current question sentence as input State information, and the intent information of the current question sentence of combination determines the reply strategy of next step;
Step 5: replying to user according to the corresponding template of reply policy selection that step 4 determines.
The beneficial effects of the present invention are: the present invention provides a kind of more wheel dialogue methods towards trip field, the present invention Solve the standardization processing, intent information and slot position in more wheel conversational systems towards trip field to short text user's question sentence Information extraction problem, in terms of being intended to information extraction, the model based on BLSTM that the present invention uses achieve Micro-F1 value for 93.47% ideal effect, in terms of slot position information extraction, the BLSTM-CRF for the introducing intent features vocabulary that the present invention uses Model achieves the ideal effect that F1 value is 89.47%, and the present invention considers the influence of short text in more wheel dialogues, overcomes existing There is the limitation of technology.
Detailed description of the invention
Fig. 1 is a kind of flow chart of man-machine more wheel dialogue methods towards trip field of the present invention;
Fig. 2 is DAN model schematic of the present invention;
Fig. 3 is CNN model schematic of the present invention;
Fig. 4 is BLSTM model schematic of the present invention;
Fig. 5 is the BLSTM-CRF model schematic of the present invention for introducing intent features vocabulary;
Fig. 6 is the screenshot that conversational system described in the embodiment of the present invention demonstrates interface.
Specific embodiment
Further description of the technical solution of the present invention with reference to the accompanying drawing, and however, it is not limited to this, all to this Inventive technique scheme is modified or replaced equivalently, and without departing from the spirit and scope of the technical solution of the present invention, should all be covered Within the protection scope of the present invention.
Specific embodiment 1: embodiment is described with reference to Fig. 1.One kind described in present embodiment is towards trip field Man-machine take turns dialogue methods, the specific steps of this method more are as follows:
Step 1: carrying out standardization processing to the current question sentence of user, then pronoun is explicitly indicated to existing in current question sentence Or the case where lacking sentence minor structure, according to the slot position information involved in being interacted before user, to the instruction in current question sentence The sentence minor structure of pronoun and missing is substituted or is filled;The current question sentence that obtains that treated;
Step 2: obtaining step 1 treated the intent information of current question sentence using DAN, CNN or BLSTM model: will Treated that current question sentence inputs DAN, CNN or BLSTM model for step 1;The output of DAN, CNN or BLSTM model is passed through Softmax operates to obtain the intention probability of current question sentence, the intent information by the intention of maximum probability as current question sentence;
Step 3: the BLSTM-CRF model of step 1 treated current question sentence input introduces intent features vocabulary is obtained To the slot position information of current question sentence;
Step 4: determining current dialogue shape using the slot position information of history and the slot position information of current question sentence as input State information, and the intent information of the current question sentence of combination determines the reply strategy of next step;
Step 5: replying to user according to the corresponding template of reply policy selection that step 4 determines.
Specific embodiment 2: embodiment is described with reference to Fig. 2.Present embodiment is to one kind described in embodiment one Man-machine more wheel dialogue methods towards trip field are further limited;Step 1 processing is obtained using the DAN model The detailed process of the intent information of current question sentence afterwards are as follows:
Before being tested using deep learning, the present invention use first from Google increase income word2vec training word to Amount, Chinese web page text data of the data used from wikipedia in April, 2018, using CBOW model, term vector dimension Degree is 300 dimensions.Subsequent deep learning model be all be subject to the training completion term vector;
The question sentence data R in artificial constructed trip field1Item crawls the question sentence number in trip field using customized query According to R2Item obtains the question sentence data R of chat in SMP2017 evaluation and test task class and vertical class3Item;By whole R1+R2+R3Question sentence carries out It is randomly ordered, the question sentence of a part is therefrom randomly selected after randomly ordered as training set, the question sentence of rest part is as test Collection;
The DAN model includes input layer (embedding), average term vector layer (Average Embedding) and entirely Articulamentum;
The training process of DAN model is as follows:
The parameter setting of DAN model: term vector dimension is 300, and full articulamentum size is that 128, dropout is set as 0.5, Mini-batch is dimensioned to 64, the number of iterations B1, patience is set as 5, i.e., trained after 5 tests are without promotion Stop, using Adam learning rate update method, learning rate is set as 0.001;It is non-linear in full articulamentum and softmax layers of progress Before converting (activation primitive), standardized operation is carried out to data using batch-normalization method.
By the input layer of step 1 treated current question sentence inputs trained DAN model, input layer is by current question sentence In each word be mapped as corresponding term vector, it is assumed that the dimension of each term vector be q, then obtain dimension be n*q term vector Matrix;Wherein: n is the length of current question sentence, if the curtailment n of current question sentence, is filled by zero-padding method;? Input layer, using above-mentioned trained term vector in advance, obtained term vector matrix can be static state, that is, train for experiment It immobilizes in journey;
The term vector matrix that input layer obtains is averaged by average term vector layer according to dimension;And will it is average after word to Moment matrix is input to full articulamentum;The output of full articulamentum operates to obtain current question sentence to belong to by softmax is intended to the general of i Rate;Wherein, the formula of softmax operation is as follows:
Wherein: y0Represent current question sentence, p (y0=i) it is that current question sentence belongs to the probability for being intended to i, i=1,2 ..., m, m generation The total number for figure classification of expressing the meaning, eiIt is intended to the e index of i.
Deep learning experiment of the invention is carried out in the environment of tensorflow1.2.
Specific embodiment 3: embodiment is described with reference to Fig. 3 present embodiment is to one kind described in embodiment one Man-machine more wheel dialogue methods towards trip field are further limited;Step 1 processing is obtained using the CNN model The detailed process of the intent information of current question sentence afterwards are as follows:
The question sentence data R in artificial constructed trip field1Item crawls the question sentence number in trip field using customized query According to R2Item obtains the question sentence data R of chat in SMP2017 evaluation and test task class and vertical class3Item;By whole R1+R2+R3Question sentence carries out It is randomly ordered, the question sentence of a part is therefrom randomly selected after randomly ordered as training set, the question sentence of rest part is as test Collection;CNN model includes input layer, convolutional layer, pond layer and full articulamentum;
The training process of CNN model is as follows:
The parameter setting of CNN model: term vector dimension is 300, and convolution kernel size is 256, dropout 0.5, mini- Batch is dimensioned to 64, the number of iterations B2, patience is set as 5, i.e., and it is trained after 5 tests are without promotion to stop, Using Adam learning rate update method, learning rate is set as 0.001;
By the input layer of step 1 treated current question sentence inputs trained CNN model, input layer is by current question sentence In each word be mapped as corresponding term vector, it is assumed that the dimension of each term vector be q, then obtain dimension be n*q term vector Matrix;Wherein: n is the length of current question sentence, if the curtailment n of current question sentence, passes through zero padding (zero-padding) Method is filled;In input layer, using above-mentioned trained term vector in advance, obtained term vector matrix can be quiet for experiment State, i.e., it immobilizes in the training process.
Use the convolution of different convolution kernel size h*k (size of the window of word of the h for institute's convolution, k are term vector dimension) Layer carries out convolution operation to the term vector matrix of input layer output, respectively obtains the corresponding spy of convolutional layer of each size convolution kernel Sign, experiment use multiple convolution and carry out convolution operation, we can extract the key message of similar n-gram in sentence, obtain Richer feature representation.Feature (feature map) size obtained due to the convolution kernel of multiple and different sizes is also different Sample, therefore we are input to pond layer to by each feature, by maximizing pondization operation, make the dimension phase of each feature Together;
The identical feature of the dimension that pond layer is obtained is passed through as the input of full articulamentum, the output of full articulamentum Softmax operates to obtain the probability that current question sentence belongs to intention i;Wherein, the formula of softmax operation is as follows:
y0Represent current question sentence, p (y0=i) it is that current question sentence belongs to the probability for being intended to i, i=1,2 ..., m, m, which is represented, to be intended to The total number of classification, eiIt is intended to the e index of i.
Specific embodiment 4: embodiment is described with reference to Fig. 4.Present embodiment is to one kind described in embodiment one Man-machine more wheel dialogue methods towards trip field are further limited, and are obtained at step 1 using the BLSTM model The detailed process of the intent information of current question sentence after reason are as follows:
The question sentence data R in artificial constructed trip field1Item crawls the question sentence number in trip field using customized query According to R2Item obtains the question sentence data R of chat in SMP2017 evaluation and test task class and vertical class3Item;By whole R1+R2+R3Question sentence carries out It is randomly ordered, the question sentence of a part is therefrom randomly selected after randomly ordered as training set, the question sentence of rest part is as test Collection;BLSTM model includes input layer, LSTM layers of forward direction, backward LSTM layers and full articulamentum;
The training process of BLSTM model is as follows:
The parameter setting of BLSTM model: term vector dimension is 300, and LSTM layers of forward direction and backward LSTM layers of hidden layer are big Small is 128, dropout 0.5, and mini-batch is dimensioned to 64, the number of iterations B3, patience is set as 5, that is, passes through It crosses 5 tests to stop without training after promotion, using Adam learning rate update method, learning rate is set as 0.001;
By the input layer of step 1 treated current question sentence inputs trained BLSTM model, input layer will be asked currently Sentence in each word be mapped as corresponding term vector, it is assumed that the dimension of each term vector be q, then obtain dimension be n*q word to Moment matrix;Wherein: n is the length of current question sentence, if the curtailment n of current question sentence, passes through zero padding (zero- Padding) method is filled;
The term vector matrix that input layer is obtained operates before carrying out to LSTM and backward LSTM, and will be preceding to LSTM and backward The hidden layer that LSTM is operated is spliced according to dimension, obtains the hiding layer state h at each moment1,h2,…,hT, wherein h1、h2And hTRespectively represent the hiding layer state at the 1st moment, the 2nd moment and T moment;
Using obtained hiding layer state as input, the output of full articulamentum operates to obtain full articulamentum by softmax Current question sentence belongs to the probability for being intended to i;Wherein, the formula of softmax operation is as follows:
y0Represent current question sentence, p (y0=i) it is that current question sentence belongs to the probability for being intended to i, i=1,2 ..., m, m, which is represented, to be intended to The total number of classification, eiIt is intended to the e index of i.
Specific embodiment 5: embodiment is described with reference to Fig. 5.Present embodiment is to embodiment one, two, three or four A kind of man-machine more wheel dialogue methods towards trip field are further limited, by step 1 in the step 3 Treated, and current question sentence input introduces the BLSTM-CRF model of intent features vocabulary, obtains the slot position information of current question sentence Detailed process are as follows:
The word vector that this section experiment uses trains word vector, the data used by the word2vec that Google increases income to use Chinese web page text data from wikipedia in April, 2018, is different from term vector training process, in training process In text is not segmented in advance, use SKIP-GRAM model, word vector dimension is set as 50 dimensions.
The question sentence data R in artificial constructed trip field1Item crawls the question sentence number in trip field using customized query According to R2Item obtains the question sentence data R of chat in SMP2017 evaluation and test task class and vertical class3Item;By whole R1+R2+R3Question sentence carries out It is randomly ordered, the question sentence of a part is therefrom randomly selected after randomly ordered as training set, the question sentence of rest part is as test Collection;BLSTM-CRF model includes input layer, LSTM layers of forward direction, backward LSTM layers and CRF output layer;
The training process of BLSTM-CRF model is as follows:
The parameter setting of BLSTM-CRF model, word vector dimension are 50 dimensions, and LSTM layer of forward direction and backward LSTM's layers is hiding Layer dimension is 128, it is intended that characteristic dimension 10, dropout 0.5, mini-batch are dimensioned to 64, and the number of iterations is B4, patience is set as 5, i.e., and it is trained after 5 tests are without promotion to stop, using Adam learning rate update method, study Rate is set as 0.001;
By the input layer of step 1 treated current question sentence inputs trained BLSTM-CRF model, input layer output The word vector of current question sentence is expressed as X=(X1,X2,…,XN), X1For the word vector of the first character of current question sentence, N is current The total number of word in question sentence;
Construct intent features vocabulary, it is intended that feature vocabulary includes corresponding Feature Words in each intention;Such as inquire train meaning Train website, inquiry hotel under figure are intended to trade name etc. of going with wine.According to intent features word pair included in intent features vocabulary Treated that current question sentence is labeled for step 1, the arrangement set Y=(y after being marked1,y2,…,yN′), y1,y2,…, yN′The sequence obtained after respectively marking, N ' is the sequence total number in set Y;If there are intent features in current question sentence Word then marks rule using IOBES and is labeled to the intent features word in current question sentence;If a word is that a name is real The beginning of body is then labeled as B-label, if it is name entity inside but be not that first character is then labeled as I-label, Ending if it is name entity is so labeled as E-label, is then labeled as S- if it is the information of single name entity Label is labeled as O if other situations;
Arrangement set Y=(y after mark1,y2,…,yN′) by the input layer of trained BLSTM-CRF model (embedding) vector that operation obtains each sequence indicates, respectively by the vector of each sequence indicate with the word of current question sentence to Amount indicates to be spliced, using spliced result as LSTM layers of the forward direction of BLSTM-CRF model and backward LSTM layers of input;
To LSTM layers and backward LSTM layers of output before splicing, and using random initializtion to preceding to LSTM layers and backward LSTM layers of parameter matrix is initialized, and the hiding layer state H at each moment is obtained1,H2,…,HT, wherein H1、H2And HTPoint The hiding layer state at the 1st moment, the 2nd moment and T moment is not represented;
The splicing result of LSTM layers and backward LSTM layers output of forward direction is Q, and the size of splicing result Q is 2*N*h, and h is hidden The dimension of layer is hidden, splicing result Q obtains state matrix P by CRF output layer;What state matrix P was calculated is hidden state to sight The emission probability of survey state;
The output layer of model is not to directly adopt softmax layers, but pass through CRF layers and final label is calculated. If directly passing through softmax in output obtains optimal label, this method does not account for the label of context, model it is defeated Outgoing label is unable to fully use adjacent label information.Global optimization characteristic based on CRF, be utilized adjacent output label it Between related information predict current label, achieved the effect that global optimization;
By state matrix P, each sequences y in the arrangement set Y after calculating markkProbability score s (X, yk), k= 1,2,…,N′;
State matrix P operates each sequences y in the arrangement set Y after being marked by softmaxkProbability P (yk| X), wherein;
Wherein: e represents e index.
Each sequences y in sequence of calculation set YkLog probability log (p (yk| X)):
By the maximum sequences y of log probabilitykSlot position information as current question sentence, in which: log probability refers to is with 10 Bottom takes probability P (yk| X) logarithm.
By state matrix P, each sequences y in the arrangement set Y after calculating markkProbability score s (X, yk) tool Body process are as follows:
A indicates transfer matrix, and calculating is the transition probability size hidden between layer state,It indicates from label yk It is transferred to mark yk+1Probability;List entries X is represented to observation state ykEmission probability;y0And yNIndicate current sentence Beginning (start) and terminate (eos) label, therefore the size of A is (k+2) * (k+2) (k is hidden state number), all On possible flag sequence.
By constructing intent features vocabulary, current question sentence is labeled according to intent features vocabulary, slot can be improved The accuracy rate of position information extraction.
Specific embodiment 6: present embodiment is to a kind of man-machine more wheels towards trip field described in embodiment one Dialogue method is further limited, the detailed process of step 4 described in present embodiment are as follows:
Dialogue state tracking is using the slot position information of history and the slot position information of current question sentence as input, in history sentence There are the slot positions of vacancy in son, then current dialogue state is other than the slot position information of Historical heritage sentence, also by current question sentence The correspondence slot position information of appearance, which is filled into, to be come;The slot position information and history slot position information that occur in current question sentence are rushed It is prominent, then it needs to carry out confirmation operation to conflict slot position;With the current dialog state information of determination, determined in conjunction with step 2 current The intent information of question sentence, to determine the reply strategy of next step.
Determine that strategy is replied in the dialogue of next step according to current dialogue states: dialogue state, which tracks, believes the slot position of history The slot position information conduct input of breath, current question sentence, finally exports current status information.In this stage, system maintenance slot The state of position, slot position state table are as shown in table 1:
1 slot position state table of table
For the slot position of the vacancy in historical information, if there is the slot position information in current question sentence, this is by the slot position Information is dissolved into current state.In the event of the slot position of conflict, being as previously mentioned " departure place " is " Beijing ", such as current Refer to that " departure place " is " Foochow " in question sentence, then with regard to needing to carry out confirmation operation to conflict slot position.
Then, dialog strategy optimization module determines current system mode and next step according to current dialogue states Reply strategy.The system mode of setting is as shown in table 2:
2 system state table of table
After dialogue management module generates final movement, dialogue responder module is transferred to specifically to execute.Talk with responder module according to The final movement generated according to dialogue management generates corresponding language as agent speech transmission to webpage front-end and is presented to use Family, this completes the complete dialogues of a wheel.
Talking with used in responder module is the method based on template.Such as the sentence of final query result is generated, it is first Some revert statement templates are manually first formulated, then the corresponding value of slot position is substituted into template sentence.Such as order the sentence of air ticket Subtemplate is " time sets out, and the train from from_loc to to_loc is as follows: ", if active user wishes that the departure time is " tomorrow ", departure place are " Harbin ", and destination is " Beijing ", then the final revert statement generated after replacement is " tomorrow It sets out, as follows from Harbin to Pekinese's train: ".In order to improve user experience, avoiding the sentence generated every time all is identical sentence Son, I has formulated multiple template to each clause, randomly chooses one from template when generating and replying and is replied.
Specific embodiment 7: present embodiment is to a kind of man-machine more wheels towards trip field described in embodiment five Dialogue method is further limited, the detailed process of step 4 described in present embodiment are as follows:
Dialogue state tracking is using the slot position information of history and the slot position information of current question sentence as input, in history sentence There are the slot positions of vacancy in son, then current dialogue state is other than the slot position information of Historical heritage sentence, also by current question sentence The correspondence slot position information of appearance, which is filled into, to be come;The slot position information and history slot position information that occur in current question sentence are rushed It is prominent, then it needs to carry out confirmation operation to conflict slot position;With the current dialog state information of determination, determined in conjunction with step 2 current The intent information of question sentence, to determine the reply strategy of next step.
Specific embodiment 8: present embodiment is to a kind of man-machine more wheels towards trip field described in embodiment one Dialogue method is further limited, and carries out standardization processing to the current question sentence of user in step 1 described in present embodiment Process are as follows: delete expression present in current question sentence and Chinese incorrect codes, correct wrong word existing for current question sentence, and set phase The synonym table answered replaces with correct word to the near synonym for including in current question sentence.
Embodiment
The method mentioned according to the present invention, we have built the Task conversational system towards trip field, such as Fig. 6 It is shown;
Entire more wheel conversational systems are distributed according to three levels of front end, Intermediate Control Layer, background system.It mainly bears front end Duty receives the input sentence of user, is sent into dialogue Understanding Module, while system is generated corresponding reply and shows user, realizes user With more wheels interaction of machine.Middle layer is responsible for connection front end and backstage, according to the input of front end and semaphore control backstage System, while the operation result for receiving backstage feeds back to front-end interface.The system on backstage is mainly appointing towards trip field Business type conversational system.Front-end interface is that the form of webpage mainly uses the technologies such as html, ajax and javascript to realize, is led here Introduce the realization of the Task conversational system towards trip field on backstage.
Firstly, conversation sentence often will appear demonstrative pronoun and sentence lacks part minor structure in daily dialogue Situation, and in the conversational system using text as form, in fact it could happen that situations such as messy code, wrong word, therefore the present invention attempts base Standardization processing is carried out to question sentence in single-wheel and more wheels, in the case where more wheels, the slot position of task based access control type conversational system setting Mechanism makes phase to current sentence in conjunction with global information (slot position information above) and current information (this slot position information) The standardization processing answered.
Understand followed by dialogue.After user's input problem, we identify the intention of question sentence and the slot position letter of carrying Breath.Intention assessment is based on LSTM model training and obtains, and slot position is identified by trained using BLSTM and conditional random field models It arrives, the input for talking with understanding is the input text of user, and output is the current intention demand of user, setting out of including in text The key slot position information such as ground, destination, time.The attitude (positive or negative) of user is also judged simultaneously, it is assumed that current round System the slot position information extracted is confirmed, if dialogue understand without extraction negative attitude, default use Family confirms that slot information is correct.
Conversational system towards trip field includes the slot positions such as departure place, destination, time, hotel's name, each slot altogether With slot value be sky, slot value is to be confirmed and slot value has confirmed that three kinds of situations.Slot position information above and current question sentence are combined first Slot position information, determine current dialogue state.
Next the reply strategy of next step conversational system is determined.Slot position acquisition of information is completed or history slot position information and is worked as Before the slot position information extracted when clashing, need to confirm whether slot position information collects correctly;It is not collected in some slot position values To when, it is necessary to inquired to user;Such as when slot position corresponding under intention is all own acknowledgement state, then selection terminates This conversation tasks.
Finally, selecting suitable sentence to show from the data of json format according to the reply strategy that dialogue management exports To user.
The research contents of this paper is more wheel conversational systems towards trip field, and intention assessment and slot position identification are dialogues The key that query understands takes following several method to solve the problems, such as to lack in the corresponding corpus in trip field herein It carries out the collection of corpus: the form of manual compiling being taken to obtain the higher a collection of specific subject corpus of quality, artificial constructed number It is 3000 according to collection size;Know from Baidu and crawled the training set that size is 18048 using customized query, right The mass data obtained through crawler carries out later period screening and is left more than 10651 datas after arranging.In addition, SMP2017 evaluation and test is appointed It include 3069 corpus of chat class and vertical class (including totally 30 vertical fields such as flight, train) in business, I am by flight, train Relevant corpus is put into accordingly vertical class, and remaining hang down class and chat class corpus arrangement are put into " other " class, such as 3 institute of table Show.
Table 3
In conjunction with above-mentioned three kinds of approach, arranges obtain 13852 query corpus altogether.It, will to the data that manual sorting is collected into Data carry out it is randomly ordered after, randomly select wherein 5% data as test set, it is remaining as training set.And it will The training set and test set of the dialogue evaluation and test data of smp2017 are put into corresponding training set and test set, it is intended that identification number It is shown in Table 4 according to the statistical conditions of collection, wherein training set shares 12503 problems, and development set has 1349 problems.
Each theme corpus situation after table 4 manually marks
For two classification problems, more commonly used index includes accuracy rate precision, recall rate recall and F1 Value.It due to 30 vertical classes in SMP and chats in training set corpus, sample is 62 under " flight " classification, under " train " classification Sample is 70, and sample is 2167 under " other " classification, and the sample of " flight " and " train " class is very few, therefore this project not needle Model is trained to the training corpus of SMP and is tested on testing material, but by the training set of SMP and artificially collects mark Corpus combine and be trained, and compare experimental analysis to the test set after integration.
Table 5 is the intention assessment experimental result of DAN model.The recall of " flight " classification compares pair of the other three classification It should be worth relatively low, illustrate that classifier tends to will to belong under the question sentence misclassification to other classifications of " flight " classification.
5 DAN intention assessment experimental result of table
Table 6 is the intention assessment experimental result of CNN model.Intention assessment model based on CNN all table in each classification Now preferably.
6 CNN intention assessment experimental result of table
Table 7 is the intention assessment experimental result of BLSTM model, the intention assessment modelling effect based on BLSTM compared with other three A model ideal is a bit.
7 BLSTM intention assessment experimental result of table
Table 8 is the intention assessment Comparative result of DAN model, CNN model and BLSTM model on trip FIELD Data, Middle first three items precision, recall, F1 value is calculated in a manner of weighted, that is, is calculating overall assessment index When use lower sample number of all categories as weight.It can be seen that CNN model compares biography with LSTM model in intention assessment problem System method effect is good, and the micro-F1 value of LSTM model compares other three classifier height with macro-F1 value.Therefore this project Have finally chosen intention assessment model of the model based on LSTM as conversational system.
The comparison of 8 intention assessment experimental result of table
The main research of this paper be towards trip field more wheel conversational systems, therefore herein to user query into Artificial mark is gone.
This subject study scalability and diversity of slot position, to can be used regular expression or rule carries out piece Obtained slot position is lifted, is extracted using the mode based on regular expression, such as " train with seats type, train type, train vehicle Secondary code, commercial air flights code, Hotel Star " etc. then uses sequence labelling mould to be difficult to be extracted with regular expression Type is identified.Specific mark system is as shown in table 9, such as " departure place " and " destination ", for flight ticket booking and fire It is required information for vehicle ticket booking task;" time " element can distinguish user and specifically it is expected departure times, and " hotel " is right User is answered to expect the associated hotel moved in.
9 slot position explanation of table
The slot position composition quantity of each classification included in training set, development set is as shown in table 10.Wherein, departure place Quantity with two ingredients in destination be it is most, the two ingredients are the key messages in trip field, in train, flight, wine Three, shop etc. refers in being intended to.And the minimum number that hotel occurs, this is because hotel's element only has in an intention It is involved.
10 slot position quantity of table
In addition, mainly passing through internet channel (including 12306 official websites, major airline official website, ctrip.com station etc.) herein It is collected intent features vocabulary, is illustrated as shown in table 11:
11 intent features vocabulary of table
There are mainly two types of the labelling schemes for naming Entity recognition, and one is IOB formats, if a word is that a name is real The beginning of body is then labeled as B-label, if it is name entity inside but be not that first character is then labeled as I-label, Otherwise being labeled as O.Another scheme is IOBES, it is a kind of variant of IOB scheme, is marked than the former more E-label Remember the ending of a name entity.It determines to use IOBES labelling schemes herein, it encodes the single information for naming entity, and (S is simultaneously And the ending (E) of name entity is clearly marked.Using this scheme, a word mark can be had later at I-label High confidence level the selectional restriction of subsequent word in I-label or E-label, and IOB scheme be merely capable of determining it is subsequent Word not the inside of another entity tag (i.e. the latter word can only be novel entities beginning B-label perhaps O or continue A upper entity I-label, but it is unlikely to be the I-label of a novel entities).
The common evaluation index of sequence labelling includes accuracy rate accuracy, rate of precision precision, recall rate recall With F1 value (F1-score).
Wherein, it is predicted as positive example, practical is also positive example, we are known as true positive (TP), second situation, in advance Surveying is positive example, and be actually negative example, we are known as false positive (FP), the third situation, predicts the example that is negative, is actually positive Example, referred to as false negative (FN), a kind of last situation predict the example that is negative, practical be also negative example, referred to as true negative(TN)。
Recall indicates that the sample for being actually positive example, model are classified into the specific gravity of positive example, and recall is higher, Illustrate that model successfully recalls more positive examples;Precision expression is judged as in the sample of positive example that actually it is to model The specific gravity of positive example, precision is higher, illustrates that model is stronger to the separating capacity of positive example sample.F1-score is the comprehensive of the two It closes.F1-score is higher, illustrates that disaggregated model is more steady, effect is more ideal.
The CRF tool that the present invention uses is CRF++0.58 version.It is named Entity recognition using CRF++, mainly includes Following steps: template file is determined;Handle training data and test data;Model training and parameter adjustment.
The feature templates of this paper CRF are as follows:
#Unigram
U00:% × [- 2,0]
U01:% × [- 1,0]
U02:% × [0,0]
U03:% × [1,0]
U04:% × [2,0]
U05:% × [- 1,0]/% × [0,0]
U06:% × [0,0]/% × [1,0]
U07:% × [- 2,0]/% × [- 1,0]/% × [0,0]
U08:% × [- 1,0]/% × [0,0]/% × [1,0]
U09:% × [0,0]/% × [1,0]/% × [2,0]
#Bigram
B
The feature templates of CRF include two kinds of forms:
The first is Unigram template: first character is U, this is for describing unigram feature Template.The %x [#, #] of every a line generates the state probability function in a CRFs: f (s, o), and wherein s is the label of t moment (output), o is the example function in context such as the CRF++ supporting paper of t moment:
Func1=if (output=B and feature=" U02: that ") return 1else return 0
It is the state probability function generated by U02:%x [0,0] in the first row of input file.By the of input file A line " substitution " is into function, function return 1, meanwhile, if certain a line of input file is also " that " in the 1st column, and it Output (the 2nd column) be also equally B, then this function also returns to 1 in this line.
Second is Bigram template: first character is B, and every a line %x [#, #] generates in a CRFs Side (Edge) function: f (s', s, o), wherein s' is the label at t -1 moment.That is, Bigram type and Unigram be substantially Machine is same, only it is contemplated that the label at t -1 moment.If only writing a B, default generates f (s', s), it means that preceding One output token and current token would be combined into bigram features.
In CRF++, the parameter of CRF model mainly includes following several:
12 CRF parameter list of table
Model is adjusted mainly for-f parameter and-c parameter herein.
This project is shown in Table 13 using the BLSTM-CRF model for introducing intent features vocabulary, specific experimental result.
The BLSTM-CRF experimental result of the introducing intent features vocabulary of table 13
As can be seen that being 10 dimensions and the BLSTM-CRF model for introducing intent features vocabulary being used to imitate being intended to characteristic dimension It is highly desirable on fruit.
Table 14 is slot position identification experimental result comparison.From the point of view of the experimental result of each element, for " departure place " and The recognition effect of " destination " two elements is preferable, the reason is that being directed to departure place and destination, the context of question sentence has bright True preposition instruction, such as " from Beijing to Harbin ", " from Beijing to Foochow " etc.;It is poor for the recognition effect of hotel's name, it is former Because may be that training corpus is limited, the study for hotel's name fails to learn completely, and wants in test data for " hotel's name " There is the noun not occurred in training data much in element, and it is bad that this shows model in " hotel's name " element, and BLSTM-CRF model in the case where introducing intent features vocabulary, to the recognition effect in " hotel name " compared with other models promoted compared with Greatly, this also illustrates the models to have the ability of certain identification unregistered word.
From the point of view of the comparison of the experimental result of model, the recognition effect of BLSTM is worst compared with other models, and CRF is as statistical machine Classic sequence dimensioning algorithm in device study, has had reached comparatively ideal effect in small data set.BLSTM-CRF models coupling BLSTM is to the understandability of contextual information and global optimization's ability of CRF, and effect is compared with CRF model and BLSTM model It is good;And BLSTM-CRF (introduced feature vocabulary) model is obviously more outstanding than other models in the recognition capability to unregistered word, because This effect is most ideal.Combination learning model based on BLSTM-CRF is taken it as a whole, and recognition capability is not so good as BLSTM-CRF mould Type.
14 slot position of table identifies experimental result comparison

Claims (8)

1. a kind of man-machine more wheel dialogue methods towards trip field, which is characterized in that the specific steps of this method are as follows:
Step 1: carry out standardization processing to the current question sentence of user, then pronoun is explicitly indicated or lacks to existing in current question sentence The case where weary sentence minor structure, according to the slot position information involved in being interacted before user, to the demonstrative pronoun in current question sentence It is substituted or is filled with the sentence minor structure of missing;The current question sentence that obtains that treated;
Step 2: obtaining step 1 treated the intent information of current question sentence using DAN, CNN or BLSTM model: by step One treated that current question sentence inputs DAN, CNN or BLSTM model;Softmax is passed through in the output of DAN, CNN or BLSTM model Operation obtains the intention probability of current question sentence, the intent information by the intention of maximum probability as current question sentence;
Step 3: the BLSTM-CRF model of step 1 treated current question sentence input introduces intent features vocabulary is worked as The slot position information of preceding question sentence;
Step 4: determining current dialogue state letter using the slot position information of history and the slot position information of current question sentence as input Breath, and the intent information of the current question sentence of combination determines the reply strategy of next step;
Step 5: replying to user according to the corresponding template of reply policy selection that step 4 determines.
2. a kind of man-machine more wheel dialogue methods towards trip field according to claim 1, which is characterized in that use institute State the detailed process that DAN model obtains step 1 treated the intent information of current question sentence are as follows:
The DAN model includes input layer, average term vector layer and full articulamentum;The training process of DAN model is as follows:
The parameter setting of DAN model: term vector dimension is 300, and full articulamentum size is that 128, dropout is set as 0.5, Mini-batch is dimensioned to 64, the number of iterations B1, patience is set as 5, i.e., by 5 test accuracy rates without mentioning When rising, deconditioning;Using Adam learning rate update method, learning rate is set as 0.001;
By the input layer of step 1 treated current question sentence inputs trained DAN model, input layer will be in current question sentence Each word is mapped as corresponding term vector, it is assumed that the dimension of each term vector is q, then obtains the term vector matrix that dimension is n*q; Wherein: n is the length of current question sentence, if the curtailment n of current question sentence, is filled by zero-padding method;
The term vector matrix that input layer obtains is averaged by average term vector layer according to dimension;And the term vector square after being averaged Battle array is input to full articulamentum;The output of full articulamentum passes through softmax and operates to obtain the probability that current question sentence belongs to intention i;Its In, the formula of softmax operation is as follows:
Wherein: y0Represent current question sentence, p (y0=i) it is that current question sentence belongs to the probability for being intended to i, i=1,2 ..., m, m represents meaning The total number of figure classification, eiIt is intended to the e index of i.
3. a kind of man-machine more wheel dialogue methods towards trip field according to claim 1, which is characterized in that use institute State the detailed process that CNN model obtains step 1 treated the intent information of current question sentence are as follows:
The CNN model includes input layer, convolutional layer, pond layer and full articulamentum;The training process of CNN model is as follows:
The parameter setting of CNN model: term vector dimension is 300, and convolution kernel size is 256, dropout 0.5, mini-batch 64 are dimensioned to, the number of iterations B2, patience is set as 5, i.e., by 5 test accuracy rates without being promoted when, stop Training;Using Adam learning rate update method, learning rate is set as 0.001;
By the input layer of step 1 treated current question sentence inputs trained CNN model, input layer will be in current question sentence Each word is mapped as corresponding term vector, it is assumed that the dimension of each term vector is q, then obtains the term vector matrix that dimension is n*q; Wherein: n is the length of current question sentence, if the curtailment n of current question sentence, is filled by zero-padding method;
Convolution operation is carried out to the term vector matrix that input layer exports using the convolutional layer of different convolution kernel sizes, is respectively obtained every The corresponding feature of convolutional layer of a size convolution kernel, is input to pond layer for each feature, by maximizing pondization operation, makes every The dimension of a feature is identical;
The identical feature of the dimension that pond layer is obtained passes through softmax as the input of full articulamentum, the output of full articulamentum Operation obtains current question sentence and belongs to the probability for being intended to i;Wherein, the formula of softmax operation is as follows:
y0Represent current question sentence, p (y0=i) it is that current question sentence belongs to the probability for being intended to i, i=1,2 ..., m, m, which is represented, is intended to classification Total number, eiIt is intended to the e index of i.
4. a kind of man-machine more wheel dialogue methods towards trip field according to claim 1, which is characterized in that use institute State the detailed process that BLSTM model obtains step 1 treated the intent information of current question sentence are as follows:
The BLSTM model includes input layer, LSTM layers of forward direction, backward LSTM layers and full articulamentum;BLSTM model was trained Journey is as follows:
The parameter setting of BLSTM model: term vector dimension is 300, and the hidden layer size that LSTM layers and backward LSTM layers of forward direction is 128, dropout 0.5, mini-batch are dimensioned to 64, the number of iterations B3, patience is set as 5, i.e., by 5 Secondary test accuracy rate without promoted when, deconditioning;Using Adam learning rate update method, learning rate is set as 0.001;
By the input layer of step 1 treated current question sentence inputs trained BLSTM model, input layer will be in current question sentence Each word be mapped as corresponding term vector, it is assumed that the dimension of each term vector be q, then obtain dimension be n*q term vector square Battle array;Wherein: n is the length of current question sentence, if the curtailment n of current question sentence, is filled by zero-padding method;
The term vector matrix that input layer is obtained operates before carrying out to LSTM and backward LSTM, and by preceding to LSTM and backward LSTM It operates obtained hidden layer to be spliced according to dimension, obtains the hiding layer state h at each moment1,h2,…,hT, wherein h1、h2 And hTRespectively represent the hiding layer state at the 1st moment, the 2nd moment and T moment;
Using obtained hiding layer state as input, the output of full articulamentum operates to obtain current full articulamentum by softmax Question sentence belongs to the probability for being intended to i;Wherein, the formula of softmax operation is as follows:
y0Represent current question sentence, p (y0=i) it is that current question sentence belongs to the probability for being intended to i, i=1,2 ..., m, m, which is represented, is intended to classification Total number, eiIt is intended to the e index of i.
5. a kind of man-machine more wheel dialogue methods towards trip field according to claim 1,2,3 or 4, feature exist In obtaining by the BLSTM-CRF model of step 1 treated current question sentence input introduces intent features vocabulary in the step 3 To the detailed process of the slot position information of current question sentence are as follows:
The BLSTM-CRF model includes input layer, LSTM layers of forward direction, backward LSTM layers and CRF output layer;BLSTM-CRF mould The training process of type is as follows:
The parameter setting of BLSTM-CRF model, word vector dimension are 50 dimensions, and LSTM layers of forward direction and backward LSTM layers of hidden layer are tieed up Degree is 128, it is intended that characteristic dimension 10, dropout 0.5, mini-batch are dimensioned to 64, the number of iterations B4, Patience is set as 5, i.e., by 5 test accuracy rates without promoted when, deconditioning;Using Adam learning rate update side Method, learning rate are set as 0.001;
By the input layer of step 1 treated current question sentence inputs trained BLSTM-CRF model, input layer output is current The word vector of question sentence is expressed as X=(X1,X2,…,XN), X1For the word vector of the first character of current question sentence, N is current question sentence The total number of middle word;
Construct intent features vocabulary, it is intended that feature vocabulary includes corresponding Feature Words in each intention;According to intent features vocabulary In included intent features word to step 1, treated that current question sentence is labeled, the arrangement set Y=after being marked (y1,y2,…,yN′), y1,y2,…,yN′The sequence obtained after respectively marking, N ' is the sequence total number in set Y;
Arrangement set Y=(y after mark1,y2,…,yN′) obtained by the input layer operation of trained BLSTM-CRF model The vector of each sequence indicates, indicates to splice with the expression of the word vector of current question sentence by the vector of each sequence respectively, will spell Result after connecing is as LSTM layers of the forward direction of trained BLSTM-CRF model and backward LSTM layers of input;
To LSTM layers and backward LSTM layers of output before splicing, and using random initializtion to preceding to LSTM layers and LSTM layers backward Parameter matrix initialized, obtain the hiding layer state H at each moment1,H2,…,HT, wherein H1、H2And HTIt respectively represents The hiding layer state at the 1st moment, the 2nd moment and T moment;
The splicing result of LSTM layers and backward LSTM layers output of forward direction is Q, and the size of splicing result Q is 2*N*h, and h is hidden layer Dimension, splicing result Q obtains state matrix P by CRF output layer;
By state matrix P, each sequences y in the arrangement set Y after calculating markkProbability score s (X, yk), k=1, 2,…,N′;
State matrix P operates each sequences y in the arrangement set Y after being marked by softmaxkProbability P (yk| X), Wherein;
Each sequences y in sequence of calculation set YkLog probability log (p (yk| X)):
By the maximum sequences y of log probabilitykSlot position information as current question sentence.
6. a kind of man-machine more wheel dialogue methods towards trip field according to claim 1, which is characterized in that the step Rapid four detailed process are as follows:
Dialogue state tracking is using the slot position information of history and the slot position information of current question sentence as input, in history sentence There are the slot positions of vacancy, then current dialogue state also current question sentence occurs other than the slot position information of Historical heritage sentence Correspondence slot position information be filled into come;Conflict for the slot position information occurred in current question sentence with history slot position information, then It needs to carry out confirmation operation to conflict slot position;With the current dialog state information of determination, the current question sentence determined in conjunction with step 2 Intent information, to determine the reply strategy of next step.
7. a kind of man-machine more wheel dialogue methods towards trip field according to claim 5, which is characterized in that the step Rapid four detailed process are as follows:
Dialogue state tracking is using the slot position information of history and the slot position information of current question sentence as input, in history sentence There are the slot positions of vacancy, then current dialogue state also current question sentence occurs other than the slot position information of Historical heritage sentence Correspondence slot position information be filled into come;Conflict for the slot position information occurred in current question sentence with history slot position information, then It needs to carry out confirmation operation to conflict slot position;With the current dialog state information of determination, the current question sentence determined in conjunction with step 2 Intent information, to determine the reply strategy of next step.
8. a kind of man-machine more wheel dialogue methods towards trip field according to claim 1, which is characterized in that step 1 In the process of standardization processing is carried out to the current question sentence of user are as follows: delete expression present in current question sentence and Chinese incorrect codes, Wrong word existing for current question sentence is corrected, and sets corresponding synonym table, the near synonym for including in current question sentence are replaced with Correct word.
CN201810779805.9A 2018-07-16 2018-07-16 Man-machine multi-turn dialogue method for trip field Active CN109063035B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810779805.9A CN109063035B (en) 2018-07-16 2018-07-16 Man-machine multi-turn dialogue method for trip field

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810779805.9A CN109063035B (en) 2018-07-16 2018-07-16 Man-machine multi-turn dialogue method for trip field

Publications (2)

Publication Number Publication Date
CN109063035A true CN109063035A (en) 2018-12-21
CN109063035B CN109063035B (en) 2021-11-09

Family

ID=64816886

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810779805.9A Active CN109063035B (en) 2018-07-16 2018-07-16 Man-machine multi-turn dialogue method for trip field

Country Status (1)

Country Link
CN (1) CN109063035B (en)

Cited By (45)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109686360A (en) * 2019-01-08 2019-04-26 哈尔滨理工大学 A kind of voice is made a reservation robot
CN109783819A (en) * 2019-01-18 2019-05-21 广东小天才科技有限公司 A kind of generation method and system of regular expression
CN109885668A (en) * 2019-01-25 2019-06-14 中译语通科技股份有限公司 A kind of expansible field interactive system status tracking method and apparatus
CN109979457A (en) * 2019-05-29 2019-07-05 南京硅基智能科技有限公司 A method of thousand people, thousand face applied to Intelligent dialogue robot
CN110019688A (en) * 2019-01-23 2019-07-16 艾肯特公司 The method that robot is trained
CN110059193A (en) * 2019-06-21 2019-07-26 南京擎盾信息科技有限公司 Legal advice system based on law semanteme part and document big data statistical analysis
CN110110057A (en) * 2019-04-22 2019-08-09 南京硅基智能科技有限公司 A kind of dynamic words art group system and its application in Intelligent dialogue robot
CN110110846A (en) * 2019-04-24 2019-08-09 重庆邮电大学 Auxiliary driver's vehicle exchange method based on convolutional neural networks
CN110111787A (en) * 2019-04-30 2019-08-09 华为技术有限公司 A kind of semanteme analytic method and server
CN110209561A (en) * 2019-05-09 2019-09-06 北京百度网讯科技有限公司 Evaluating method and evaluating apparatus for dialogue platform
CN110413756A (en) * 2019-07-29 2019-11-05 北京小米智能科技有限公司 The method, device and equipment of natural language processing
CN110442676A (en) * 2019-07-02 2019-11-12 北京邮电大学 Patent retrieval method and device based on more wheel dialogues
CN110502608A (en) * 2019-07-05 2019-11-26 平安科技(深圳)有限公司 The interactive method and human-computer dialogue device of knowledge based map
CN110532374A (en) * 2019-09-04 2019-12-03 深圳市慧择时代科技有限公司 The processing method and processing device of insurance information
CN110569344A (en) * 2019-08-22 2019-12-13 阿里巴巴集团控股有限公司 Method and device for determining standard question sentence corresponding to dialog text
CN110717027A (en) * 2019-10-18 2020-01-21 易小博(武汉)科技有限公司 Multi-round intelligent question-answering method, system, controller and medium
CN110765252A (en) * 2019-10-18 2020-02-07 北京邮电大学 Knowledge-driven task-oriented dialogue management method and system easy to configure
CN110825949A (en) * 2019-09-19 2020-02-21 平安科技(深圳)有限公司 Information retrieval method based on convolutional neural network and related equipment thereof
CN110928997A (en) * 2019-12-04 2020-03-27 北京文思海辉金信软件有限公司 Intention recognition method and device, electronic equipment and readable storage medium
CN111177341A (en) * 2019-12-11 2020-05-19 江苏艾佳家居用品有限公司 End-to-end ID + SF model-based user conversation demand extraction method and system
CN111259668A (en) * 2020-05-07 2020-06-09 腾讯科技(深圳)有限公司 Reading task processing method, model training device and computer equipment
CN111353035A (en) * 2020-03-11 2020-06-30 镁佳(北京)科技有限公司 Man-machine conversation method and device, readable storage medium and electronic equipment
CN111353029A (en) * 2020-02-22 2020-06-30 杭州电子科技大学 Semantic matching-based multi-turn spoken language understanding method
CN111382238A (en) * 2018-12-27 2020-07-07 上海智臻智能网络科技股份有限公司 Interactive method and device of multi-element interactive process
CN111414195A (en) * 2019-02-03 2020-07-14 北京邮电大学 Universal configurable, multi-model-compatible and domain-transportable multi-round dialogue system
WO2020143381A1 (en) * 2019-01-10 2020-07-16 华为技术有限公司 Dialogue processing method and device
CN111524533A (en) * 2020-03-02 2020-08-11 北京声智科技有限公司 Voice operation method and device, electronic equipment and storage medium
CN111723574A (en) * 2020-07-09 2020-09-29 腾讯科技(深圳)有限公司 Information processing method and device and computer readable storage medium
CN111813899A (en) * 2020-08-31 2020-10-23 腾讯科技(深圳)有限公司 Intention identification method and device based on multiple rounds of conversations
CN111813900A (en) * 2019-04-10 2020-10-23 北京猎户星空科技有限公司 Multi-turn conversation processing method and device, electronic equipment and storage medium
CN112000787A (en) * 2020-08-17 2020-11-27 上海小鹏汽车科技有限公司 Voice interaction method, server and voice interaction system
WO2020244065A1 (en) * 2019-06-04 2020-12-10 平安科技(深圳)有限公司 Character vector definition method, apparatus and device based on artificial intelligence, and storage medium
CN112100349A (en) * 2020-09-03 2020-12-18 深圳数联天下智能科技有限公司 Multi-turn dialogue method and device, electronic equipment and storage medium
CN112131885A (en) * 2019-06-24 2020-12-25 北京猎户星空科技有限公司 Semantic recognition method and device, electronic equipment and storage medium
CN112364147A (en) * 2020-12-01 2021-02-12 四川长虹电器股份有限公司 Cross-domain multi-turn dialogue method based on knowledge graph and implementation system
CN112380332A (en) * 2020-11-17 2021-02-19 深圳追一科技有限公司 Interactive knowledge feedback method, device and computer storage medium
CN112445888A (en) * 2019-09-05 2021-03-05 腾讯科技(深圳)有限公司 Information extraction method and related equipment
CN112632254A (en) * 2020-12-28 2021-04-09 平安科技(深圳)有限公司 Conversation state determining method, terminal device and storage medium
CN112800737A (en) * 2019-10-29 2021-05-14 京东数字科技控股有限公司 Natural language text generation method and device and dialogue system
CN112988960A (en) * 2021-02-09 2021-06-18 中国科学院自动化研究所 Dialog state tracking method, device, equipment and storage medium
WO2021135603A1 (en) * 2019-12-31 2021-07-08 华为技术有限公司 Intention recognition method, server and storage medium
CN113254610A (en) * 2021-05-14 2021-08-13 廖伟智 Multi-round conversation generation method for patent consultation
CN113282736A (en) * 2021-07-08 2021-08-20 北京百度网讯科技有限公司 Dialogue understanding and model training method, device, equipment and storage medium
CN113326360A (en) * 2021-04-25 2021-08-31 哈尔滨工业大学 Natural language understanding method in small sample scene
CN113569918A (en) * 2021-07-05 2021-10-29 北京淇瑀信息科技有限公司 Classification temperature adjusting method, classification temperature adjusting device, electronic equipment and medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130158980A1 (en) * 2011-12-15 2013-06-20 Microsoft Corporation Suggesting intent frame(s) for user request(s)
CN106156003A (en) * 2016-06-30 2016-11-23 北京大学 A kind of question sentence understanding method in question answering system
CN106874441A (en) * 2017-02-07 2017-06-20 腾讯科技(上海)有限公司 Intelligent answer method and apparatus
CN107369443A (en) * 2017-06-29 2017-11-21 北京百度网讯科技有限公司 Dialogue management method and device based on artificial intelligence
CN107451115A (en) * 2017-07-11 2017-12-08 中国科学院自动化研究所 The construction method and system of Chinese Prosodic Hierarchy forecast model end to end

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130158980A1 (en) * 2011-12-15 2013-06-20 Microsoft Corporation Suggesting intent frame(s) for user request(s)
CN106156003A (en) * 2016-06-30 2016-11-23 北京大学 A kind of question sentence understanding method in question answering system
CN106874441A (en) * 2017-02-07 2017-06-20 腾讯科技(上海)有限公司 Intelligent answer method and apparatus
CN107369443A (en) * 2017-06-29 2017-11-21 北京百度网讯科技有限公司 Dialogue management method and device based on artificial intelligence
CN107451115A (en) * 2017-07-11 2017-12-08 中国科学院自动化研究所 The construction method and system of Chinese Prosodic Hierarchy forecast model end to end

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
王东升等: "面向限定领域问答***的自然语言理解方法综述", 《计算机科学》 *

Cited By (69)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111382238A (en) * 2018-12-27 2020-07-07 上海智臻智能网络科技股份有限公司 Interactive method and device of multi-element interactive process
CN109686360A (en) * 2019-01-08 2019-04-26 哈尔滨理工大学 A kind of voice is made a reservation robot
US11977815B2 (en) 2019-01-10 2024-05-07 Huawei Technologies Co., Ltd. Dialogue processing method and device
WO2020143381A1 (en) * 2019-01-10 2020-07-16 华为技术有限公司 Dialogue processing method and device
CN111427992A (en) * 2019-01-10 2020-07-17 华为技术有限公司 Conversation processing method and device
CN109783819A (en) * 2019-01-18 2019-05-21 广东小天才科技有限公司 A kind of generation method and system of regular expression
CN109783819B (en) * 2019-01-18 2023-10-20 广东小天才科技有限公司 Regular expression generation method and system
CN110019688A (en) * 2019-01-23 2019-07-16 艾肯特公司 The method that robot is trained
CN109885668A (en) * 2019-01-25 2019-06-14 中译语通科技股份有限公司 A kind of expansible field interactive system status tracking method and apparatus
WO2020151017A1 (en) * 2019-01-25 2020-07-30 中译语通科技股份有限公司 Scalable field human-machine dialogue system state tracking method and device
CN111414195B (en) * 2019-02-03 2022-03-25 北京邮电大学 Universal configurable, multi-model-compatible and domain-transportable multi-round dialogue system
CN111414195A (en) * 2019-02-03 2020-07-14 北京邮电大学 Universal configurable, multi-model-compatible and domain-transportable multi-round dialogue system
CN111813900B (en) * 2019-04-10 2023-12-08 北京猎户星空科技有限公司 Multi-round dialogue processing method and device, electronic equipment and storage medium
CN111813900A (en) * 2019-04-10 2020-10-23 北京猎户星空科技有限公司 Multi-turn conversation processing method and device, electronic equipment and storage medium
CN110110057A (en) * 2019-04-22 2019-08-09 南京硅基智能科技有限公司 A kind of dynamic words art group system and its application in Intelligent dialogue robot
CN110110846A (en) * 2019-04-24 2019-08-09 重庆邮电大学 Auxiliary driver's vehicle exchange method based on convolutional neural networks
CN110111787A (en) * 2019-04-30 2019-08-09 华为技术有限公司 A kind of semanteme analytic method and server
CN110111787B (en) * 2019-04-30 2021-07-09 华为技术有限公司 Semantic parsing method and server
US11900924B2 (en) 2019-04-30 2024-02-13 Huawei Technologies Co., Ltd. Semantic parsing method and server
CN110209561A (en) * 2019-05-09 2019-09-06 北京百度网讯科技有限公司 Evaluating method and evaluating apparatus for dialogue platform
CN110209561B (en) * 2019-05-09 2024-02-09 北京百度网讯科技有限公司 Evaluation method and evaluation device for dialogue platform
CN109979457A (en) * 2019-05-29 2019-07-05 南京硅基智能科技有限公司 A method of thousand people, thousand face applied to Intelligent dialogue robot
WO2020244065A1 (en) * 2019-06-04 2020-12-10 平安科技(深圳)有限公司 Character vector definition method, apparatus and device based on artificial intelligence, and storage medium
CN110059193A (en) * 2019-06-21 2019-07-26 南京擎盾信息科技有限公司 Legal advice system based on law semanteme part and document big data statistical analysis
CN112131885A (en) * 2019-06-24 2020-12-25 北京猎户星空科技有限公司 Semantic recognition method and device, electronic equipment and storage medium
CN110442676A (en) * 2019-07-02 2019-11-12 北京邮电大学 Patent retrieval method and device based on more wheel dialogues
WO2021003819A1 (en) * 2019-07-05 2021-01-14 平安科技(深圳)有限公司 Man-machine dialog method and man-machine dialog apparatus based on knowledge graph
CN110502608B (en) * 2019-07-05 2023-10-10 平安科技(深圳)有限公司 Man-machine conversation method and man-machine conversation device based on knowledge graph
CN110502608A (en) * 2019-07-05 2019-11-26 平安科技(深圳)有限公司 The interactive method and human-computer dialogue device of knowledge based map
EP3772012A1 (en) * 2019-07-29 2021-02-03 Beijing Xiaomi Intelligent Technology Co., Ltd. Method and device for natural language processing and storage medium
CN110413756B (en) * 2019-07-29 2022-02-15 北京小米智能科技有限公司 Method, device and equipment for processing natural language
KR102330061B1 (en) 2019-07-29 2021-11-26 베이징 시아오미 인텔리전트 테크놀로지 컴퍼니 리미티드 Natural language processing method, apparatus, device and computer readable storage medium
CN110413756A (en) * 2019-07-29 2019-11-05 北京小米智能科技有限公司 The method, device and equipment of natural language processing
KR20210016262A (en) * 2019-07-29 2021-02-15 베이징 시아오미 인텔리전트 테크놀로지 컴퍼니 리미티드 Methods, devices and devices for processing natural language
US11501078B2 (en) 2019-07-29 2022-11-15 Beijing Xiaomi Intelligent Technology Co., Ltd. Method and device for performing reinforcement learning on natural language processing model and storage medium
WO2021017173A1 (en) * 2019-07-29 2021-02-04 北京小米智能科技有限公司 Natural language processing method, apparatus and device
CN110569344A (en) * 2019-08-22 2019-12-13 阿里巴巴集团控股有限公司 Method and device for determining standard question sentence corresponding to dialog text
CN110532374A (en) * 2019-09-04 2019-12-03 深圳市慧择时代科技有限公司 The processing method and processing device of insurance information
CN110532374B (en) * 2019-09-04 2022-05-10 深圳市慧择时代科技有限公司 Insurance information processing method and device
CN112445888A (en) * 2019-09-05 2021-03-05 腾讯科技(深圳)有限公司 Information extraction method and related equipment
CN110825949A (en) * 2019-09-19 2020-02-21 平安科技(深圳)有限公司 Information retrieval method based on convolutional neural network and related equipment thereof
CN110765252A (en) * 2019-10-18 2020-02-07 北京邮电大学 Knowledge-driven task-oriented dialogue management method and system easy to configure
CN110717027A (en) * 2019-10-18 2020-01-21 易小博(武汉)科技有限公司 Multi-round intelligent question-answering method, system, controller and medium
CN112800737A (en) * 2019-10-29 2021-05-14 京东数字科技控股有限公司 Natural language text generation method and device and dialogue system
CN110928997A (en) * 2019-12-04 2020-03-27 北京文思海辉金信软件有限公司 Intention recognition method and device, electronic equipment and readable storage medium
CN111177341A (en) * 2019-12-11 2020-05-19 江苏艾佳家居用品有限公司 End-to-end ID + SF model-based user conversation demand extraction method and system
CN111177341B (en) * 2019-12-11 2022-07-08 江苏艾佳家居用品有限公司 End-to-end ID + SF model-based user conversation demand extraction method and system
WO2021135603A1 (en) * 2019-12-31 2021-07-08 华为技术有限公司 Intention recognition method, server and storage medium
CN111353029A (en) * 2020-02-22 2020-06-30 杭州电子科技大学 Semantic matching-based multi-turn spoken language understanding method
CN111353029B (en) * 2020-02-22 2020-09-22 杭州电子科技大学 Semantic matching-based multi-turn spoken language understanding method
CN111524533B (en) * 2020-03-02 2023-12-08 北京声智科技有限公司 Voice operation method, device, electronic equipment and storage medium
CN111524533A (en) * 2020-03-02 2020-08-11 北京声智科技有限公司 Voice operation method and device, electronic equipment and storage medium
CN111353035A (en) * 2020-03-11 2020-06-30 镁佳(北京)科技有限公司 Man-machine conversation method and device, readable storage medium and electronic equipment
CN111259668A (en) * 2020-05-07 2020-06-09 腾讯科技(深圳)有限公司 Reading task processing method, model training device and computer equipment
CN111723574A (en) * 2020-07-09 2020-09-29 腾讯科技(深圳)有限公司 Information processing method and device and computer readable storage medium
CN112000787A (en) * 2020-08-17 2020-11-27 上海小鹏汽车科技有限公司 Voice interaction method, server and voice interaction system
CN112000787B (en) * 2020-08-17 2021-05-14 上海小鹏汽车科技有限公司 Voice interaction method, server and voice interaction system
CN111813899A (en) * 2020-08-31 2020-10-23 腾讯科技(深圳)有限公司 Intention identification method and device based on multiple rounds of conversations
CN112100349B (en) * 2020-09-03 2024-03-19 深圳数联天下智能科技有限公司 Multi-round dialogue method and device, electronic equipment and storage medium
CN112100349A (en) * 2020-09-03 2020-12-18 深圳数联天下智能科技有限公司 Multi-turn dialogue method and device, electronic equipment and storage medium
CN112380332A (en) * 2020-11-17 2021-02-19 深圳追一科技有限公司 Interactive knowledge feedback method, device and computer storage medium
CN112364147A (en) * 2020-12-01 2021-02-12 四川长虹电器股份有限公司 Cross-domain multi-turn dialogue method based on knowledge graph and implementation system
CN112632254A (en) * 2020-12-28 2021-04-09 平安科技(深圳)有限公司 Conversation state determining method, terminal device and storage medium
CN112632254B (en) * 2020-12-28 2024-05-28 平安科技(深圳)有限公司 Dialogue state determining method, terminal device and storage medium
CN112988960A (en) * 2021-02-09 2021-06-18 中国科学院自动化研究所 Dialog state tracking method, device, equipment and storage medium
CN113326360A (en) * 2021-04-25 2021-08-31 哈尔滨工业大学 Natural language understanding method in small sample scene
CN113254610A (en) * 2021-05-14 2021-08-13 廖伟智 Multi-round conversation generation method for patent consultation
CN113569918A (en) * 2021-07-05 2021-10-29 北京淇瑀信息科技有限公司 Classification temperature adjusting method, classification temperature adjusting device, electronic equipment and medium
CN113282736A (en) * 2021-07-08 2021-08-20 北京百度网讯科技有限公司 Dialogue understanding and model training method, device, equipment and storage medium

Also Published As

Publication number Publication date
CN109063035B (en) 2021-11-09

Similar Documents

Publication Publication Date Title
CN109063035A (en) A kind of man-machine more wheel dialogue methods towards trip field
CN110838288B (en) Voice interaction method and system and dialogue equipment
US11693894B2 (en) Conversation oriented machine-user interaction
CN107239444B (en) A kind of term vector training method and system merging part of speech and location information
CN105095182B (en) A kind of return information recommendation method and device
CN110110335B (en) Named entity identification method based on stack model
CN107203511A (en) A kind of network text name entity recognition method based on neutral net probability disambiguation
CN111081220B (en) Vehicle-mounted voice interaction method, full-duplex dialogue system, server and storage medium
CN108846063A (en) Determine the method, apparatus, equipment and computer-readable medium of problem answers
CN103593412B (en) A kind of answer method and system based on tree structure problem
CN114168749A (en) Question generation system based on knowledge graph and question word drive
CN109447266A (en) A kind of agricultural science and technology service intelligent sorting method based on big data
WO2022048194A1 (en) Method, apparatus and device for optimizing event subject identification model, and readable storage medium
CN109325780A (en) A kind of exchange method of the intelligent customer service system in E-Governance Oriented field
CN116562265B (en) Information intelligent analysis method, system and storage medium
CN110297889A (en) A kind of enterprise's Sentiment orientation analysis method based on Fusion Features
CN113901208A (en) Method for analyzing emotion tendentiousness of intermediate-crossing language comments blended with theme characteristics
CN116757652A (en) Online recruitment recommendation system and method based on large language model
CN116737922A (en) Tourist online comment fine granularity emotion analysis method and system
Joukhadar et al. Arabic dialogue act recognition for textual chatbot systems
Hongli [Retracted] Design and Application of English Grammar Error Correction System Based on Deep Learning
CN110689359A (en) Method and device for dynamically updating model
CN117632098A (en) Intelligent building design system based on AIGC
CN114298011B (en) Neural network, training method, aspect emotion analysis method, device and storage medium
CN111708870A (en) Deep neural network-based question answering method and device and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant