CN113326387A - Intelligent conference information retrieval method - Google Patents
Intelligent conference information retrieval method Download PDFInfo
- Publication number
- CN113326387A CN113326387A CN202110603641.6A CN202110603641A CN113326387A CN 113326387 A CN113326387 A CN 113326387A CN 202110603641 A CN202110603641 A CN 202110603641A CN 113326387 A CN113326387 A CN 113326387A
- Authority
- CN
- China
- Prior art keywords
- information
- conference
- text
- video
- audio
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 27
- 238000001514 detection method Methods 0.000 claims description 6
- 238000003058 natural language processing Methods 0.000 claims description 6
- 230000011218 segmentation Effects 0.000 claims description 3
- 238000006243 chemical reaction Methods 0.000 claims 1
- 238000000605 extraction Methods 0.000 claims 1
- 238000010586 diagram Methods 0.000 description 3
- 230000002452 interceptive effect Effects 0.000 description 2
- 230000007547 defect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/48—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/483—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/43—Querying
- G06F16/438—Presentation of query results
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Library & Information Science (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
The invention discloses an intelligent retrieval method for meeting information, which relates to the technical field of meeting records and comprises the following steps: recording conference information, recording and extracting audio streams of conference video contents in a multimedia mode in the whole process in real time, sending the audio streams to a voice recognition module to convert voice into character information, storing the character information, marking according to the conference progress time, inputting text information or voice information to inquire, matching and inquiring with the conference information stored previously, and returning corresponding audio or video information. According to the invention, the conference information is recorded in a multimedia manner, the retrieval information and the conference information are matched and inquired through multi-level processing, when the conference record is matched, the information of the time axis where the corresponding record is located is displayed, and simultaneously, the audio information which enables a user to intuitively hear the speaking of the current conference is played, so that the later analysis and understanding of the conference are more convenient, and the conference record retrieval experience is greatly improved.
Description
Technical Field
The invention relates to the technical field of conference recording, in particular to an intelligent conference information retrieval method.
Background
As technology advances, many products that automatically record conference content are being launched. From the earliest recorders to automated speech-to-text equipment. These recording methods record a lot of contents, which often last for several hours. Resulting in time and effort for reviewing or retrieving the meeting record. Although some advanced products label conference participants according to human biological characteristics such as voiceprints, fingerprints and the like, and then quickly locate conference recording contents through the labels, even labeling by using geographic information and administrative levels, the advanced products have the disadvantages of being not humanized, such as: the conference records can not be inquired and retrieved according to the content, the inquiry record mode is single, and the conference records can only be watched and listened back manually and can not be positioned quickly.
Disclosure of Invention
The invention aims to provide an intelligent conference information retrieval method to overcome the defects in the prior art.
In order to achieve the above purpose, the invention provides the following technical scheme: an intelligent conference information retrieval method comprises the following steps:
recording conference information in a multimedia mode in real time in the whole process, wherein the whole process comprises archiving in the forms of whole video, audio, text and the like of a conference;
step two, extracting the audio stream of the conference video content, copying the audio stream from a media file or a Container (Container) of a stream file by using demultiplexing (demux) to extract the audio stream from the video stream, and sending the audio stream to a voice recognition module to convert voice into text information and store the text information, wherein the original video file is unchanged;
marking the video, audio and text of the conference record according to the time of the conference, taking a speaking detection technology or a silence detection technology as a starting and ending judgment basis, further combining with a context judgment technology of NLP (natural language processing) including but not limited to SBD (sequence boundary prediction) and WS (word segmentation) with smaller granularity to process the speaking content as a unit or a word as a unit, and adding marks on the processed conference record content according to sentences and words and storing the processed conference record content;
step four, the user searches the conference record, inputs text information or voice information for inquiry, converts the voice into a text through the voice-to-text module if the voice information is received, matches and inquires the text with the conference information stored previously, returns corresponding audio or video information and attaches the text information converted by the voice;
and step five, when the user views the returned result, the recorded content of the context can be quickly retrieved, namely, the user can simultaneously view the conference information before and after the retrieved time period, the recorded content is displayed to the user in text, audio or video information through highlighting, and the user can intuitively position, select and modify the corresponding content.
Preferably, in the first step, if the conference is a network video conference, the conference information is directly obtained through the network, and if the conference is a non-network conference, the conference is recorded through multimedia devices such as audio recording and video recording, and is extracted and converted.
Preferably, the text information converted by the voice in the second step can be used for displaying and recording the real-time conference subtitles while being stored.
Preferably, the time interval marked in step three is marked by a sentence or a pause in the audio containing the content of the utterance.
Preferably, the marked video Segments, audio Segments and text Segments in the third step are stored in a one-to-one correspondence with a time sequence table, wherein the video Segments are recorded in a list vsrl (video Segments Recording list) in a time sequence, the audio Segments are recorded in a list ssrl (speech Segments Recording list) in a time sequence, and the text segment information is recorded in a list tsrl (text Segments Recording list) in a time sequence.
Preferably, the matching process in the fourth step includes the following steps:
step a, first-level character matching, wherein text information generated by user searching is used for matching text information stored in a TSRL, if the text information can be matched, audio information of a corresponding time period is returned, and if corresponding video information exists, the video information of the corresponding time period is directly returned.
B, second-level character matching, if the first level can not be matched, reducing the text information to smaller granularity through SBD for matching again, if the text information can be matched, returning the corresponding audio or video information,
and c, second-stage processing, namely decomposing the information into smaller granularity for re-matching through WS if the second-stage processing cannot be matched, returning corresponding audio or video information if the information can be matched, and otherwise, ensuring that the query information cannot be matched.
In the technical scheme, the invention provides the following technical effects and advantages:
the invention records the meeting information in a multimedia mode, marks and stores the video, audio and text recorded by the meeting according to the meeting running time, and users search and match through the text information and perform multi-stage processing, matching and inquiring the retrieval information and the conference information, when the conference record is matched, displaying the information of the time axis where the corresponding record is located, the user can select the text information through the interactive equipment, the corresponding audio is highlighted, and simultaneously, the audio information which enables the user to intuitively hear the speaking of the current conference is played, the user can randomly select any paragraph in the text module, the corresponding audio or video will be positioned and played synchronously, otherwise, the user will quickly retrieve the audio or video content, corresponding text information can also be displayed immediately, so that the later analysis and understanding of the conference are more convenient, and the experience of conference record retrieval is greatly improved.
Drawings
In order to more clearly illustrate the embodiments of the present application or technical solutions in the prior art, the drawings needed to be used in the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments described in the present invention, and other drawings can be obtained by those skilled in the art according to the drawings.
FIG. 1 is a flow chart of the present invention.
FIG. 2 is a flow chart of a query matching process of the present invention.
FIG. 3 is an exemplary diagram of an interaction interface when the present invention returns a result.
Fig. 4 is a diagram of another example of a pickup interface for the case where only audio and text information are returned as a result according to the present invention.
FIG. 5 is an exemplary diagram of an interface for a user to select a query message status in the state of FIG. 4 according to the present invention.
Description of reference numerals:
A. a video information display module; B. a video information clip display module of a time axis; C. a text information display module; D. an audio information display module; E. and a time position display module.
Detailed Description
In order to make the technical solutions of the present invention better understood, those skilled in the art will now describe the present invention in further detail with reference to the accompanying drawings.
The invention provides an intelligent conference information retrieval method, which comprises the following steps:
recording conference information, recording the whole process in a multimedia mode in real time, filing the whole conference in forms of whole video, audio, text and the like, directly acquiring the conference information through a network if the conference is a network video conference, and recording the conference through multimedia equipment such as sound recording and video recording if the conference is a non-network conference, and extracting and converting;
step two, extracting the audio stream of the conference video content, copying the audio stream from a media file or a Container (Container) of a stream file by using demultiplexing (demux) to extract the audio stream from the video stream, and sending the audio stream to a voice recognition module to convert voice into text information and store the text information while the original video file is kept unchanged, wherein the audio stream can be used for displaying and recording a real-time conference subtitle;
marking the video, audio and text of the conference record according to the time of the conference, taking a talk detection technology or a silence detection technology as a starting and ending judgment basis, taking a sentence or a pause containing the talk content in the audio as a mark at a time interval, further combining with a context judgment technology of NLP (natural language processing) including but not limited to SBD (sequence boundary prediction) and WS (word segmentation) with smaller granularity to process the talk content according to a sentence unit or a word unit, and respectively adding marks according to a sentence and a word and storing the processed conference record content;
the marked video Segments, audio Segments and text fields are respectively stored in a one-to-one correspondence way by setting a time sequence table, wherein the video Segments are recorded in a list VSRL (video Segments Recording List) according to the time sequence, the audio Segments are recorded in a list SSRL (speech Segments Recording List) according to the time sequence, the text segment information is recorded in a list TSRL (text Segments Recording List) according to the time sequence, and the structures of the VSRL, the SSRL and the TSRL are respectively shown in a table 1, a table 2 and a table 3:
TABLE 1 VSRL example
Sequence No. | Time Offset | Duration | SegmentsURL |
0 | 00:00:00.000 | 1000 | VS001.mp4 |
1 | 00:00:01.000 | 1000 | VS002.mp4 |
2 | 00:00:02.000 | 1500 | VS003.mp4 |
… | … | … | … |
Wherein,
sequence No. represents a mark serial number, and the key value of the mark similarity relation table is unique and is corresponding to the SSRL and the TSRL;
time Offset represents the Offset from the entire video, from the beginning to the current;
duration represents the time length of the current segment in milliseconds ms;
segmentsrurl indicates the URL information of the video file storing the current segment; the streaming media player can directly play the corresponding video by using the URL; in actual use, the address should be further encrypted, and the data security is improved through encryption.
TABLE 2 SSRL examples
Sequence No. | Time Offset | Duration | SegmentsURL |
0 | 00:00:00.000 | 1000 | SS001.wav |
1 | 00:00:01.000 | 1000 | SS002.wav |
2 | 00:00:02.000 | 1500 | SS003.wav |
… | … | … | … |
Wherein,
sequence No. indicates a tag number, which is the same as VSRL;
time Offset represents the Offset from the entire video, from the beginning to the current;
duration represents the time length of the current segment in milliseconds ms;
segmentsrurl indicates the audio file URL information that stores the current segment; the streaming media player can directly play the corresponding audio by using the URL; in actual use, the address should be further encrypted, and the data security is improved through encryption.
Wherein
Sequence No.VSRL=Sequence No.SSRL=Sequence No.TSRL
TABLE 3 TSRL example
Wherein,
sequence No. indicates a tag number, which is the same as VSRL;
the Original Language Code represents the Language of an Original text, and is represented by an ISO-639-1 standard, wherein en is English, zh is Chinese and the like;
code Page, representing character set of literal Code, 1209 UTF-8 Unicode;
characters, representing a file URL where text is stored;
step four, the user searches the conference record, inputs text information or voice information for inquiry, converts the voice into a text through the voice-to-text module if the voice information is received, matches and inquires the text with the conference information stored previously, returns corresponding audio or video information and attaches the text information converted by the voice;
the text matching process comprises the following steps:
step a, first-level character matching, wherein text information generated by user searching is used for matching text information stored in a TSRL, if the text information can be matched, audio information of a corresponding time period is returned, and if corresponding video information exists, the video information of the corresponding time period is directly returned.
B, second-level character matching, if the first level can not be matched, reducing the text information to smaller granularity through SBD for matching again, if the text information can be matched, returning the corresponding audio or video information,
step c, second level processing, if the second level can not be matched, the information is decomposed into smaller granularity through WS and matched again, if the information can be matched, the corresponding audio or video information is returned, otherwise, the inquiry information can not be matched
And step five, when the user views the returned result, the recorded content of the context can be quickly retrieved, namely, the user can simultaneously view the conference information before and after the retrieved time period, the recorded content is displayed to the user in text, audio or video information through highlighting, and the user can intuitively position, select and modify the corresponding content.
To sum up, the invention records in a multimedia manner, includes filing of the whole video, audio, text and other forms of the conference, sends an audio stream to a voice recognition module to convert voice into character information, marks the video, audio and text recorded in the conference according to the time of the conference, and stores the video, audio and text in one-to-one correspondence according to the time mark, a user inquires by inputting text information or voice information, if the voice information is received, converts the voice into text through a voice-to-text module, matches and inquires the conference information through multi-stage processing, when the conference record is matched, displays the information of the time axis where the corresponding record is located, including the information of upper and lower paragraphs, the user can select the text information through an interactive device such as a mouse or a touch screen, the text information is highlighted, and the corresponding audio is also highlighted, meanwhile, audio information of the current conference speech is played by the user visually, if video information corresponding to the recorded video information exists, a corresponding video clip is played, the user can randomly select any paragraph in the text module, the corresponding audio or video can be synchronously positioned and played, otherwise, the user quickly retrieves the audio or video content, and the corresponding text information can be displayed immediately, so that the later analysis and understanding of the conference are more convenient, and the experience of conference record retrieval is greatly improved.
While certain exemplary embodiments of the present invention have been described above by way of illustration only, it will be apparent to those of ordinary skill in the art that the described embodiments may be modified in various different ways without departing from the spirit and scope of the invention. Accordingly, the drawings and description are illustrative in nature and should not be construed as limiting the scope of the invention.
Claims (6)
1. An intelligent conference information retrieval method is characterized by comprising the following steps:
recording conference information in a multimedia mode in real time in the whole process, wherein the whole process comprises archiving in the forms of whole video, audio, text and the like of a conference;
step two, extracting the audio stream of the conference video content, copying the audio stream from a media file or a Container (Container) of a stream file by using demultiplexing (demux) to extract the audio stream from the video stream, and sending the audio stream to a voice recognition module to convert voice into text information and store the text information, wherein the original video file is unchanged;
marking the video, audio and text of the conference record according to the time of the conference, taking a speaking detection technology or a silence detection technology as a starting and ending judgment basis, further combining with a context judgment technology of NLP (natural language processing) including but not limited to SBD (sequence boundary prediction) and WS (word segmentation) with smaller granularity to process the speaking content as a unit or a word as a unit, and adding marks on the processed conference record content according to sentences and words and storing the processed conference record content;
step four, the user searches the conference record, inputs text information or voice information for inquiry, converts the voice into a text through the voice-to-text module if the voice information is received, matches and inquires the text with the conference information stored previously, returns corresponding audio or video information and attaches the text information converted by the voice;
and step five, when the user views the returned result, the recorded content of the context can be quickly retrieved, namely, the user can simultaneously view the conference information before and after the retrieved time period, the recorded content is displayed to the user in text, audio or video information through highlighting, and the user can intuitively position, select and modify the corresponding content.
2. The intelligent conference information retrieval method according to claim 1, wherein: in the first step, if the conference is a network video conference, the conference information is directly acquired through the network, and if the conference is a non-network conference, the conference is recorded through multimedia equipment such as audio recording and video recording, and extraction and conversion are performed.
3. The intelligent conference information retrieval method according to claim 1, wherein: and the text information converted by the voice in the second step can be used for displaying and recording the real-time conference subtitles while being stored.
4. The intelligent conference information retrieval method according to claim 1, wherein: the time interval marked in step three is marked by a sentence or a pause in the audio containing the content of the utterance.
5. The intelligent conference information retrieval method according to claim 1, wherein: the marked video Segments, audio Segments and text Segments in the third step are respectively set with a time sequence table for one-to-one storage, wherein the video Segments are recorded in a list VSRL (video Segments Recording List) according to the time sequence, the audio Segments are recorded in a list SSRL (speech Segments Recording List) according to the time sequence, and the text segment information is recorded in a list TSRL (text Segments Recording List) according to the time sequence.
6. The intelligent retrieval method for meeting information of claim 5, wherein the matching process in the fourth step comprises the following steps:
step a, first-level character matching, wherein text information generated by user searching is used for matching text information stored in a TSRL, if the text information can be matched, audio information of a corresponding time period is returned, and if corresponding video information exists, the video information of the corresponding time period is directly returned;
b, second-level character matching, wherein if the first level character matching cannot be achieved, the text information is reduced to smaller granularity through SBD and matched again, and if the text information can be matched, the corresponding audio or video information is returned;
and c, second-stage processing, namely decomposing the information into smaller granularity for re-matching through WS if the second-stage processing cannot be matched, returning corresponding audio or video information if the information can be matched, and otherwise, ensuring that the query information cannot be matched.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110603641.6A CN113326387B (en) | 2021-05-31 | 2021-05-31 | Intelligent conference information retrieval method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110603641.6A CN113326387B (en) | 2021-05-31 | 2021-05-31 | Intelligent conference information retrieval method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113326387A true CN113326387A (en) | 2021-08-31 |
CN113326387B CN113326387B (en) | 2022-12-13 |
Family
ID=77422786
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110603641.6A Active CN113326387B (en) | 2021-05-31 | 2021-05-31 | Intelligent conference information retrieval method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113326387B (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114385859A (en) * | 2021-12-29 | 2022-04-22 | 北京理工大学 | Multi-modal retrieval method for video content |
CN114661943A (en) * | 2022-05-21 | 2022-06-24 | 中科云策(深圳)科技成果转化信息技术有限公司 | Conference information storage management system |
CN115828907A (en) * | 2023-02-16 | 2023-03-21 | 南昌航天广信科技有限责任公司 | Intelligent conference management method, system, readable storage medium and computer equipment |
WO2023093092A1 (en) * | 2021-11-26 | 2023-06-01 | 华为技术有限公司 | Minuting method, and terminal device and minuting system |
CN116708055A (en) * | 2023-06-06 | 2023-09-05 | 深圳市艾姆诗电商股份有限公司 | Intelligent multimedia audiovisual image processing method, system and storage medium |
WO2024093442A1 (en) * | 2022-10-31 | 2024-05-10 | 北京字跳网络技术有限公司 | Method and apparatus for checking audiovisual content, and device and storage medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105045828A (en) * | 2015-06-26 | 2015-11-11 | 徐信 | Retrieval system and method for accurate positioning of audio/video speech information |
CN108345679A (en) * | 2018-02-26 | 2018-07-31 | 科大讯飞股份有限公司 | A kind of audio and video search method, device, equipment and readable storage medium storing program for executing |
CN111814028A (en) * | 2020-09-14 | 2020-10-23 | 腾讯科技(深圳)有限公司 | Information searching method and device |
CN112765460A (en) * | 2021-01-08 | 2021-05-07 | 北京字跳网络技术有限公司 | Conference information query method, device, storage medium, terminal device and server |
CN112839195A (en) * | 2020-12-30 | 2021-05-25 | 深圳市皓丽智能科技有限公司 | Method and device for consulting meeting record, computer equipment and storage medium |
-
2021
- 2021-05-31 CN CN202110603641.6A patent/CN113326387B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105045828A (en) * | 2015-06-26 | 2015-11-11 | 徐信 | Retrieval system and method for accurate positioning of audio/video speech information |
CN108345679A (en) * | 2018-02-26 | 2018-07-31 | 科大讯飞股份有限公司 | A kind of audio and video search method, device, equipment and readable storage medium storing program for executing |
CN111814028A (en) * | 2020-09-14 | 2020-10-23 | 腾讯科技(深圳)有限公司 | Information searching method and device |
CN112839195A (en) * | 2020-12-30 | 2021-05-25 | 深圳市皓丽智能科技有限公司 | Method and device for consulting meeting record, computer equipment and storage medium |
CN112765460A (en) * | 2021-01-08 | 2021-05-07 | 北京字跳网络技术有限公司 | Conference information query method, device, storage medium, terminal device and server |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2023093092A1 (en) * | 2021-11-26 | 2023-06-01 | 华为技术有限公司 | Minuting method, and terminal device and minuting system |
CN114385859A (en) * | 2021-12-29 | 2022-04-22 | 北京理工大学 | Multi-modal retrieval method for video content |
CN114661943A (en) * | 2022-05-21 | 2022-06-24 | 中科云策(深圳)科技成果转化信息技术有限公司 | Conference information storage management system |
WO2024093442A1 (en) * | 2022-10-31 | 2024-05-10 | 北京字跳网络技术有限公司 | Method and apparatus for checking audiovisual content, and device and storage medium |
CN115828907A (en) * | 2023-02-16 | 2023-03-21 | 南昌航天广信科技有限责任公司 | Intelligent conference management method, system, readable storage medium and computer equipment |
CN116708055A (en) * | 2023-06-06 | 2023-09-05 | 深圳市艾姆诗电商股份有限公司 | Intelligent multimedia audiovisual image processing method, system and storage medium |
CN116708055B (en) * | 2023-06-06 | 2024-02-20 | 深圳市艾姆诗电商股份有限公司 | Intelligent multimedia audiovisual image processing method, system and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN113326387B (en) | 2022-12-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN113326387B (en) | Intelligent conference information retrieval method | |
CN108305632B (en) | Method and system for forming voice abstract of conference | |
US9576581B2 (en) | Metatagging of captions | |
US10225625B2 (en) | Caption extraction and analysis | |
CN102075695B (en) | New generation intelligent cataloging system and method facing large amount of broadcast television programs | |
JP4466564B2 (en) | Document creation / viewing device, document creation / viewing robot, and document creation / viewing program | |
CN109246472A (en) | Video broadcasting method, device, terminal device and storage medium | |
CN107968959B (en) | Knowledge point segmentation method for teaching video | |
US20100299131A1 (en) | Transcript alignment | |
CN105245917A (en) | System and method for generating multimedia voice caption | |
JP2007519987A (en) | Integrated analysis system and method for internal and external audiovisual data | |
CN101202864A (en) | Player for movie contents | |
CN110781328A (en) | Video generation method, system, device and storage medium based on voice recognition | |
Bougrine et al. | Toward a Web-based speech corpus for Algerian dialectal Arabic varieties | |
JP2012181358A (en) | Text display time determination device, text display system, method, and program | |
JP2018033048A (en) | Metadata generation system | |
US20230281248A1 (en) | Structured Video Documents | |
CN102136001B (en) | Multi-media information fuzzy search method | |
US6813624B1 (en) | Method and apparatus for archival and retrieval of multiple data streams | |
KR20210138311A (en) | Apparatus for generating parallel corpus data between text language and sign language and method therefor | |
CN106550268B (en) | Video processing method and video processing device | |
JP3437617B2 (en) | Time-series data recording / reproducing device | |
CN109376145B (en) | Method and device for establishing movie and television dialogue database and storage medium | |
Shahraray et al. | Pictorial transcripts: Multimedia processing applied to digital library creation | |
KR20010037652A (en) | Audio indexing system and method, and audio retrieval system and method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |