CN114092674A - Multimedia data analysis method and system - Google Patents

Multimedia data analysis method and system Download PDF

Info

Publication number
CN114092674A
CN114092674A CN202210076247.6A CN202210076247A CN114092674A CN 114092674 A CN114092674 A CN 114092674A CN 202210076247 A CN202210076247 A CN 202210076247A CN 114092674 A CN114092674 A CN 114092674A
Authority
CN
China
Prior art keywords
information
content
scene
marking
user
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202210076247.6A
Other languages
Chinese (zh)
Other versions
CN114092674B (en
Inventor
易星
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Peiruiweihang Interconnection Technology Co ltd
Original Assignee
Beijing Peiruiweihang Interconnection Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Peiruiweihang Interconnection Technology Co ltd filed Critical Beijing Peiruiweihang Interconnection Technology Co ltd
Priority to CN202210076247.6A priority Critical patent/CN114092674B/en
Publication of CN114092674A publication Critical patent/CN114092674A/en
Application granted granted Critical
Publication of CN114092674B publication Critical patent/CN114092674B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9535Search customisation based on user profiles and personalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Computer Hardware Design (AREA)
  • Computer Graphics (AREA)
  • Data Mining & Analysis (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The invention discloses a multimedia data analysis method and a multimedia data analysis system, wherein high-freedom VR tourism of a user is realized through the arrangement of a VR output module, an eye movement marking module, a media analysis module and a content selection module, scene objects in a VR scene are marked through the eye movement marking module, the user can synchronously screen similar or identical contents through a cloud end to carry out recommendation display and share tourism ideas, a virtual tourism interaction platform based on the cloud end and VR is built, people can conveniently and clearly visit according to self information, and the experience of VR tourism is well improved.

Description

Multimedia data analysis method and system
Technical Field
The invention relates to the field related to data processing, in particular to a multimedia data analysis method and a multimedia data analysis system.
Background
Multimedia data generally refers to data carrier types such as files, audio, pictures and videos, the multimedia data can generally transmit information content to personnel intuitively and quickly, and with the rapid development of technologies such as internet technology and VR, the display modes of the multimedia data become richer gradually and the application of the multimedia data is wider.
In the prior art, VR technology has been developed and used for online tourism in scenic spots and online tourism in exhibitions, and has been recognized and pursued by people because of the presence experience that the traditional multimedia method does not have.
However, in the prior art, VR tour modes are limited to a simple scene passive output mode, and the use experience is single, so that online tour of a large scene is tedious, a user cannot conveniently select and acquire interesting contents to actively select tour, and fatigue is easily caused after long-time use.
Disclosure of Invention
The present invention is directed to a multimedia data analysis method and system, which solve the problems set forth in the background art.
In order to achieve the purpose, the invention provides the following technical scheme:
multimedia data analysis system is applicable to scenes such as VR travel and VR exhibition, includes:
the VR output module is used for acquiring and outputting multimedia content in real time, wherein the multimedia content is a local VR scene of a display scene;
the eye movement marking module is used for acquiring and generating eyeball movement information of a user, analyzing and generating watching direction information according to the eyeball movement information, receiving and responding to object marking information, and marking corresponding scene content according to the multimedia content and the watching direction information;
the media analysis module is used for extracting the marked scene content in the multimedia content, performing content identification analysis on the scene content, and generating a marking tag according to an analysis result, wherein the marking tag is used for representing the characteristic information of the scene content, and the marking tag is uploaded through a cloud server and is synchronized into a cloud tag;
and the content selecting and listing module is used for comparing and screening the cloud tags of the display scene according to the marking tags and marking the display scene according to the position information corresponding to the cloud tags in the comparison and screening result.
As a further scheme of the invention: the eye movement marking module comprises:
the eye tracking unit is used for tracking the movement of the eyeballs of the user through sensing and image acquisition equipment to generate a group of eye movement information of the user;
the focusing calculation simulating unit is used for carrying out motion analysis on a group of eyeball motion information to generate watching direction information, the motion analysis is used for carrying out focusing calculation on the watching direction of the user according to the eyeball motion information, and the watching direction information is used for representing the watching direction of the eyeballs of the user;
the scene focusing unit is used for carrying out focusing prompt on the corresponding scene content in the multimedia content according to the gazing direction information, the focusing prompt is used for displaying an identification result of the system on the gazing content of the user to the user, and the scene content corresponds to the unique position information;
and the object marking unit is used for receiving and responding to object marking information and marking the scene content of the focusing prompt.
As a further scheme of the invention: the content selection module comprises:
the tag acquisition unit is used for accessing the cloud tags of the display scene through the cloud server;
the tag screening unit is used for comparing and screening the cloud tags according to the content of the marked tags to generate screening results, and the content coincidence rate of the cloud tags and the marked tags in the screening results is greater than or equal to a preset screening standard;
and the tag output unit is used for acquiring the position information corresponding to the cloud tag in the screening result, and marking and updating the display scene according to the position information.
As a further scheme of the invention: the VR output module includes:
the scene acquisition unit is used for acquiring scene model data and a ground data distribution map of the display scene, wherein the scene model data is arranged corresponding to the data distribution map and is used for representing a space model and image information of the display scene at a certain position;
the motion simulation unit is used for generating an observation motion point location in the data distribution map, receiving motion control information from a user terminal, and controlling and updating the position information of the observation motion point location relative to the data distribution map according to the motion control information, wherein the observation motion point location is used for representing the simulated observation position information of a user, and the observation motion point location comprises direction information;
and the scene output unit is used for acquiring the corresponding scene model data according to the observed motion point, rendering and outputting the scene model data.
As a further scheme of the invention: the eye movement marking module further comprises:
and the content marking unit is used for receiving and responding message marking information and marking the scene content of the focusing prompt, wherein the message marking information is used for representing the subjective marking content of the user.
As a further scheme of the invention: the media analysis module comprises a subjective tag unit, and the content selection module comprises a subjective screening unit;
the subjective tag unit is used for extracting the content of the message marking information to generate a subjective tag, and the subjective tag is uploaded and synchronized into a subjective cloud tag through the cloud server;
and the subjective screening unit is used for screening the subjective cloud tags through the subjective tags and marking the display scene.
The embodiment of the invention aims to provide a multimedia data analysis method, which is characterized by comprising the following steps:
acquiring and outputting multimedia content in real time, wherein the multimedia content is a local VR scene of a display scene;
acquiring and generating eyeball motion information of a user, analyzing and generating watching orientation information according to the eyeball motion information, receiving and responding object marking information, and marking corresponding scene content according to the multimedia content and the watching orientation information;
extracting the marked scene content in the multimedia content, performing content identification analysis on the scene content, and generating a marking tag according to an analysis result, wherein the marking tag is used for representing the characteristic information of the scene content, and the marking tag is synchronously updated through a cloud server;
and comparing and screening the cloud tags of the display scene according to the marked tags, and marking the display scene according to the azimuth information corresponding to the cloud tags in the comparison and screening result.
As a further scheme of the invention: the steps of collecting and generating the eyeball motion information of the user, analyzing and generating the gazing direction information according to the eyeball motion information, receiving and responding to object marking information, and marking the corresponding scene content according to the multimedia content and the gazing direction information specifically comprise:
the method comprises the steps that the eyeball of a user is tracked through sensing and image acquisition equipment, and a group of eyeball movement information of the user is generated;
performing motion analysis on the group of eyeball motion information to generate gaze direction information, wherein the motion analysis is used for performing focusing calculation on the gaze direction of the user according to the eyeball motion information, and the gaze direction information is used for representing the gaze direction of the eyeballs of the user;
focusing prompt is carried out on the corresponding scene content in the multimedia content according to the gazing direction information, and the focusing prompt is used for displaying the recognition result of the system on the gazing content of the user to the user;
and receiving and responding to object marking information, and marking the scene content of the focusing prompt.
As a further scheme of the invention: the step of comparing and screening the cloud tags of the display scene according to the labeled tags and labeling the display scene according to the orientation information corresponding to the cloud tags in the comparison and screening result specifically includes:
accessing a cloud tag of the display scene through the cloud server;
comparing and screening the cloud labels through the content of the marked labels to generate a screening result, wherein the content coincidence rate of the cloud labels and the marked labels in the screening result is greater than or equal to a preset screening standard;
and acquiring position information corresponding to the cloud label in the screening result, and marking and updating the display scene according to the position information.
Compared with the prior art, the invention has the beneficial effects that: through VR output module, eye movement mark module, the setting of media analysis module and content option module, user's high degree of freedom VR tourism has been realized, simultaneously mark the scene article in the VR scene through eye movement mark module, the user can recommend the sharing that shows and the tourism is thought through the synchronous screening similar or the same content in high in the clouds, a virtual tour interaction platform based on high in the clouds and VR has been built, can make people more convenient and clear and definite purposive tourism according to self information, good promotion VR tourism's experience.
Drawings
Fig. 1 is a block diagram showing a configuration of a multimedia data analysis system.
Fig. 2 is a block diagram of an eye movement labeling module in the multimedia data analysis system.
FIG. 3 is a block diagram of a content selection module in the multimedia data analysis system.
FIG. 4 is a block diagram of a VR output module in the multimedia data analysis system.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
The following detailed description of specific embodiments of the present invention is provided in connection with specific embodiments.
As shown in fig. 1, the multimedia data analysis system provided for an embodiment of the present invention is suitable for VR travel and VR exhibition, and includes:
and the VR output module 100 is configured to acquire and output multimedia content in real time, where the multimedia content is a local VR scene of the display scene.
The eye movement marking module 300 is configured to collect and generate eye movement information of a user, analyze and generate gaze direction information according to the eye movement information, receive and respond to object marking information, and mark corresponding scene content according to the multimedia content and the gaze direction information.
The media analysis module 500 is configured to extract the scene content marked in the multimedia content, perform content identification analysis on the scene content, and generate a mark tag according to an analysis result, where the mark tag is used to represent feature information of the scene content, and the mark tag is uploaded through a cloud server and synchronized into a cloud tag.
The content selecting and listing module 700 is configured to compare and screen the cloud tags of the display scene according to the tag tags, and tag the display scene according to the position information corresponding to the cloud tags in the comparison and screening result.
In this embodiment, in use, firstly, a display scene is selected by a user, for example, a user a wants to watch a certain park through a VR device, then VR data content of the display scene of the park, that is, the whole park is downloaded and obtained, and then output through the VR device (the data content of a part of the display scene output through the VR device at a certain moment is defined as multimedia content here) (the above is realized by the VR output module 100), when the user visits the display scene of the park through the VR device, the sensing device tracks the eyeball movement of the user a, so as to determine a scene object in the park watched by the user a through simulation calculation (the above is realized by the eye movement marking module 300), for example, a plant m that the user a never sees is watching in the park, the user a has a great interest in the plant m, and wants to find and watch other same plantlets m, at this time, the user a marks the plant m watched by the user a through the control unit, then the system performs feature analysis on the marked plant m to generate a mark label (for example, feature 1, feature 2 or name) (which is implemented by the media analysis module 500), and then the system accesses all the park labels according to the label to find out the same or similar labels, and marks the same or similar labels in the display scene of the park, and outputs the labels through the VR device of the user to assist the user in viewing the park visit (which is implemented by the content listing module 700) (where the sensor device, the VR device, and the like are all one of the functional components of each module).
As shown in fig. 2, as another preferred embodiment of the present invention, the eye movement marking module 300 includes:
an eye tracking unit 301, configured to perform motion tracking on the eyeballs of the user through a sensing and image obtaining device, and generate a set of eye motion information of the user.
And a focusing calculation simulating unit 302, configured to perform motion analysis on a set of the eyeball motion information to generate gaze direction information, where the motion analysis is used to perform focusing calculation on a gaze direction of a user according to the eyeball motion information, and the gaze direction information is used to represent a gaze direction of an eyeball of the user.
And the scene focusing unit 303 is configured to perform focusing prompt on the corresponding scene content in the multimedia content according to the gazing direction information, where the focusing prompt is used to show an identification result of the system on the user gazing content to the user, and the scene content corresponds to the unique position information.
An object marking unit 304, configured to receive and mark the scene content of the focus hint in response to object marking information.
In this embodiment, the eye movement labeling module 300 is divided into more detailed functions and some terms are described, the eye movement tracking of the user can be realized by arranging a camera and a sensor in the VR device, and the focusing prompt here can be understood as a frame selection labeling of an object, etc., which is displayed in an output image of the VR device of the user, and the generation of the frame selection labeling changes along with the continuous adjustment change of the observation direction, etc., of the user.
As shown in fig. 3, as another preferred embodiment of the present invention, the content selection module 700 includes:
a tag obtaining unit 701, configured to access the cloud tag of the display scene through the cloud server.
A tag screening unit 702, configured to compare and screen the cloud tag according to the content of the tag, and generate a screening result, where a content coincidence rate of the cloud tag and the tag in the screening result is greater than or equal to a preset screening standard.
And a tag output unit 703 configured to obtain location information corresponding to the cloud tag in the screening result, and mark and update the display scene according to the location information.
In this embodiment, the content sorting module 700 is explained in more detail and the noun definition, where the cloud tag is determined by a preset sorting criterion during the sorting process, and a user may adjust the sorting criterion according to his own requirement and a sorting scheme for sorting results, for example, a matching sorting criterion that the tag overlap ratio reaches eighty percent, and then the matching tag is sorted and only displayed in a certain number in a circular range at the position of the display scene where the user is located, and the tag is hidden and displayed outside the range, which may be customized at will.
As shown in fig. 4, as another preferred embodiment of the present invention, the VR output module 100 includes:
the scene obtaining unit 101 is configured to obtain scene model data and a data distribution map of the display scene, where the scene model data is set corresponding to the data distribution map and is used to represent a spatial model and image information of the display scene at a certain position.
The motion simulation unit 102 is configured to generate an observed motion point location in the data distribution map, receive motion control information from a user, and control and update position information of the observed motion point location relative to the data distribution map according to the motion control information, where the observed motion point location is used to represent simulated observed position information of the user, and the observed motion point location includes direction information.
And the scene output unit 103 is configured to obtain the corresponding scene model data according to the observed motion point, render and output the scene model data.
In this embodiment, the VR scene output module 100 is further described, when a user uses VR equipment to visit a display scene, the content that the user can obtain and view at a certain time is limited, that is, the obtained content is a sector area of the human field of view, so that it is necessary to simulate the actual scene of the user visiting the park by establishing a view point moving along with the user control in the display scene, and update the position and orientation of the view point in the display scene by the motion simulation unit 102, so as to obtain and output scene model data to the VR equipment of the user.
As another preferred embodiment of the present invention, the eye movement marking module 300 further includes:
and the content marking unit is used for receiving and responding message marking information and marking the scene content of the focusing prompt, wherein the message marking information is used for representing the subjective marking content of the user.
Further, the media analysis module 500 includes a subjective tag unit, and the content selection module 700 includes a subjective screening unit;
the subjective tag unit is used for extracting the content of the left message marking information to generate a subjective tag, and the subjective tag is uploaded and synchronized into a subjective cloud tag through the cloud server.
And the subjective screening unit is used for screening the subjective cloud tags through the subjective tags and marking the display scene.
In this embodiment, message marking information and a subjective label are newly introduced, and the function of the message marking information and the subjective label can be equivalent to an evaluation message function, for example, after the user a marks the plant m, the message marking information is further set for sharing the feeling of the person who subsequently sees the mark, and meanwhile, the plant can be further described as a basis for generating the subjective label for further screening other similar landscape contents in the park, so as to expand the screening range.
As shown in fig. 4, the present invention also provides a multimedia data analysis method, including the following steps:
and acquiring and outputting multimedia content in real time, wherein the multimedia content is a local VR scene of a display scene.
The method comprises the steps of collecting and generating eyeball motion information of a user, analyzing and generating watching orientation information according to the eyeball motion information, receiving and responding object marking information, and marking corresponding scene content according to the multimedia content and the watching orientation information.
Extracting the marked scene content in the multimedia content, performing content identification analysis on the scene content, and generating a marking label according to an analysis result, wherein the marking label is used for representing the characteristic information of the scene content, and the marking label is synchronously updated through a cloud server.
And comparing and screening the cloud tags of the display scene according to the marked tags, and marking the display scene according to the azimuth information corresponding to the cloud tags in the comparison and screening result.
As another preferred embodiment of the present invention, the step of collecting and generating eye movement information of a user, analyzing and generating gaze direction information according to the eye movement information, receiving and responding to object tagging information, and tagging corresponding scene content according to the multimedia content and the gaze direction information specifically includes:
and tracking the movement of the eyeballs of the user through sensing and image acquisition equipment to generate a group of eye movement information of the user.
And performing motion analysis on the eyeball motion information to generate gaze direction information, wherein the motion analysis is used for performing focusing and fitting calculation on the gaze direction of the user according to the eyeball motion information, and the gaze direction information is used for representing the gaze direction of the eyeballs of the user.
And carrying out focusing prompt on the corresponding scene content in the multimedia content according to the gazing direction information, wherein the focusing prompt is used for displaying the recognition result of the system on the gazing content of the user to the user.
And receiving and responding to object marking information, and marking the scene content of the focusing prompt.
As another preferred embodiment of the present invention, the step of comparing and screening the cloud tags of the display scene according to the labeled tags, and labeling the display scene according to the orientation information corresponding to the cloud tags in the comparison and screening result specifically includes:
and accessing the cloud label of the display scene through the cloud server.
And comparing and screening the cloud labels according to the content of the marked labels to generate a screening result, wherein the content coincidence rate of the cloud labels and the marked labels in the screening result is greater than or equal to a preset screening standard.
And acquiring position information corresponding to the cloud label in the screening result, and marking and updating the display scene according to the position information.
It will be understood by those skilled in the art that all or part of the processes of the methods of the embodiments described above can be implemented by a computer program, which can be stored in a non-volatile computer-readable storage medium, and can include the processes of the embodiments of the methods described above when the program is executed. Any reference to memory, storage, database, or other medium used in the embodiments provided herein may include non-volatile and/or volatile memory, among others. Non-volatile memory can include read-only memory (ROM), Programmable ROM (PROM), Electrically Programmable ROM (EPROM), Electrically Erasable Programmable ROM (EEPROM), or flash memory. Volatile memory can include Random Access Memory (RAM) or external cache memory. By way of illustration and not limitation, RAM is available in a variety of forms such as Static RAM (SRAM), Dynamic RAM (DRAM), Synchronous DRAM (SDRAM), Double Data Rate SDRAM (DDRSDRAM), Enhanced SDRAM (ESDRAM), Synchronous Link DRAM (SLDRAM), Rambus Direct RAM (RDRAM), direct bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM).
Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosure herein. This application is intended to cover any variations, uses, or adaptations of the disclosure following, in general, the principles of the disclosure and including such departures from the present disclosure as come within known or customary practice within the art to which the disclosure pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the disclosure being indicated by the following claims.
It will be understood that the present disclosure is not limited to the precise arrangements described above and shown in the drawings and that various modifications and changes may be made without departing from the scope thereof. The scope of the present disclosure is limited only by the appended claims.

Claims (9)

1. Multimedia data analysis system is applicable to VR travel and VR exhibition scene, its characterized in that includes:
the VR output module is used for acquiring and outputting multimedia content in real time, wherein the multimedia content is a local VR scene of a display scene;
the eye movement marking module is used for acquiring and generating eyeball movement information of a user, analyzing and generating watching direction information according to the eyeball movement information, receiving and responding to object marking information, and marking corresponding scene content according to the multimedia content and the watching direction information;
the media analysis module is used for extracting the marked scene content in the multimedia content, performing content identification analysis on the scene content, and generating a marking tag according to an analysis result, wherein the marking tag is used for representing the characteristic information of the scene content, and the marking tag is uploaded through a cloud server and is synchronized into a cloud tag;
and the content selecting and listing module is used for comparing and screening the cloud tags of the display scene according to the marking tags and marking the display scene according to the position information corresponding to the cloud tags in the comparison and screening result.
2. The multimedia data analysis system of claim 1, wherein the eye movement labeling module comprises:
the eye tracking unit is used for tracking the movement of the eyeballs of the user through sensing and image acquisition equipment to generate a group of eye movement information of the user;
the focusing calculation simulating unit is used for carrying out motion analysis on a group of eyeball motion information to generate watching direction information, the motion analysis is used for carrying out focusing calculation on the watching direction of the user according to the eyeball motion information, and the watching direction information is used for representing the watching direction of the eyeballs of the user;
the scene focusing unit is used for carrying out focusing prompt on the corresponding scene content in the multimedia content according to the gazing direction information, the focusing prompt is used for displaying an identification result of the system on the gazing content of the user to the user, and the scene content corresponds to the unique position information;
and the object marking unit is used for receiving and responding to object marking information and marking the scene content of the focusing prompt.
3. The multimedia data analysis system of claim 2, wherein the content selection module comprises:
the tag acquisition unit is used for accessing the cloud tags of the display scene through the cloud server;
the tag screening unit is used for comparing and screening the cloud tags according to the content of the marked tags to generate screening results, and the content coincidence rate of the cloud tags and the marked tags in the screening results is greater than or equal to a preset screening standard;
and the tag output unit is used for acquiring the position information corresponding to the cloud tag in the screening result, and marking and updating the display scene according to the position information.
4. The multimedia data analysis system of claim 1, wherein the VR output module comprises:
the scene acquisition unit is used for acquiring scene model data and a ground data distribution map of the display scene, wherein the scene model data is arranged corresponding to the data distribution map and is used for representing a space model and image information of the display scene at a certain position;
the motion simulation unit is used for generating an observation motion point location in the data distribution map, receiving motion control information from a user terminal, and controlling and updating the position information of the observation motion point location relative to the data distribution map according to the motion control information, wherein the observation motion point location is used for representing the simulated observation position information of a user, and the observation motion point location comprises direction information;
and the scene output unit is used for acquiring the corresponding scene model data according to the observed motion point, rendering and outputting the scene model data.
5. The multimedia data analysis system of claim 2, wherein the eye movement labeling module further comprises:
and the content marking unit is used for receiving and responding message marking information and marking the scene content of the focusing prompt, wherein the message marking information is used for representing the subjective marking content of the user.
6. The system of claim 5, wherein the media analysis module comprises a subjective tagging element and the content listing module comprises a subjective filtering element;
the subjective tag unit is used for extracting the content of the message marking information to generate a subjective tag, and the subjective tag is uploaded and synchronized into a subjective cloud tag through the cloud server;
and the subjective screening unit is used for screening the subjective cloud tags through the subjective tags and marking the display scene.
7. The multimedia data analysis method is characterized by comprising the following steps:
acquiring and outputting multimedia content in real time, wherein the multimedia content is a local VR scene of a display scene;
acquiring and generating eyeball motion information of a user, analyzing and generating watching orientation information according to the eyeball motion information, receiving and responding object marking information, and marking corresponding scene content according to the multimedia content and the watching orientation information;
extracting the marked scene content in the multimedia content, performing content identification analysis on the scene content, and generating a marking tag according to an analysis result, wherein the marking tag is used for representing the characteristic information of the scene content, and the marking tag is synchronously updated through a cloud server;
and comparing and screening the cloud tags of the display scene according to the marked tags, and marking the display scene according to the azimuth information corresponding to the cloud tags in the comparison and screening result.
8. The method according to claim 7, wherein the step of collecting and generating eye movement information of the user, analyzing and generating gaze direction information according to the eye movement information, receiving and responding to object tagging information, and tagging corresponding scene content according to the multimedia content and the gaze direction information specifically comprises:
the method comprises the steps that the eyeball of a user is tracked through sensing and image acquisition equipment, and a group of eyeball movement information of the user is generated;
performing motion analysis on the group of eyeball motion information to generate gaze direction information, wherein the motion analysis is used for performing focusing calculation on the gaze direction of the user according to the eyeball motion information, and the gaze direction information is used for representing the gaze direction of the eyeballs of the user;
focusing prompt is carried out on the corresponding scene content in the multimedia content according to the gazing direction information, and the focusing prompt is used for displaying the recognition result of the system on the gazing content of the user to the user;
and receiving and responding to object marking information, and marking the scene content of the focusing prompt.
9. The multimedia data analysis method according to claim 8, wherein the step of comparing and screening the cloud tags of the display scene according to the tag tags and tagging the display scene according to the orientation information corresponding to the cloud tags in the comparison and screening result specifically comprises:
accessing a cloud tag of the display scene through the cloud server;
comparing and screening the cloud labels through the content of the marked labels to generate a screening result, wherein the content coincidence rate of the cloud labels and the marked labels in the screening result is greater than or equal to a preset screening standard;
and acquiring position information corresponding to the cloud label in the screening result, and marking and updating the display scene according to the position information.
CN202210076247.6A 2022-01-24 2022-01-24 Multimedia data analysis method and system Active CN114092674B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210076247.6A CN114092674B (en) 2022-01-24 2022-01-24 Multimedia data analysis method and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210076247.6A CN114092674B (en) 2022-01-24 2022-01-24 Multimedia data analysis method and system

Publications (2)

Publication Number Publication Date
CN114092674A true CN114092674A (en) 2022-02-25
CN114092674B CN114092674B (en) 2022-04-22

Family

ID=80309163

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210076247.6A Active CN114092674B (en) 2022-01-24 2022-01-24 Multimedia data analysis method and system

Country Status (1)

Country Link
CN (1) CN114092674B (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106843468A (en) * 2016-12-27 2017-06-13 努比亚技术有限公司 A kind of man-machine interaction method in terminal and VR scenes
CN107861629A (en) * 2017-12-20 2018-03-30 杭州埃欧哲建设工程咨询有限公司 A kind of practice teaching method based on VR
CN108958460A (en) * 2017-05-19 2018-12-07 深圳市掌网科技股份有限公司 Building sand table methods of exhibiting and system based on virtual reality
US20190130647A1 (en) * 2017-09-27 2019-05-02 Goertek Technology Co.,Ltd. Display control method and system, and virtual reality device
CN110285818A (en) * 2019-06-28 2019-09-27 武汉大学 A kind of Relative Navigation of eye movement interaction augmented reality
CN111105294A (en) * 2019-12-20 2020-05-05 武汉市奥拓智能科技有限公司 VR navigation method, system, client, server and storage medium thereof
CN112666714A (en) * 2015-08-07 2021-04-16 托比股份公司 Gaze direction mapping
CN113194410A (en) * 2021-04-28 2021-07-30 云景文旅科技有限公司 5G and virtual augmented reality fused tourism information processing method and system

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112666714A (en) * 2015-08-07 2021-04-16 托比股份公司 Gaze direction mapping
CN106843468A (en) * 2016-12-27 2017-06-13 努比亚技术有限公司 A kind of man-machine interaction method in terminal and VR scenes
CN108958460A (en) * 2017-05-19 2018-12-07 深圳市掌网科技股份有限公司 Building sand table methods of exhibiting and system based on virtual reality
US20190130647A1 (en) * 2017-09-27 2019-05-02 Goertek Technology Co.,Ltd. Display control method and system, and virtual reality device
CN107861629A (en) * 2017-12-20 2018-03-30 杭州埃欧哲建设工程咨询有限公司 A kind of practice teaching method based on VR
CN110285818A (en) * 2019-06-28 2019-09-27 武汉大学 A kind of Relative Navigation of eye movement interaction augmented reality
CN111105294A (en) * 2019-12-20 2020-05-05 武汉市奥拓智能科技有限公司 VR navigation method, system, client, server and storage medium thereof
CN113194410A (en) * 2021-04-28 2021-07-30 云景文旅科技有限公司 5G and virtual augmented reality fused tourism information processing method and system

Also Published As

Publication number Publication date
CN114092674B (en) 2022-04-22

Similar Documents

Publication Publication Date Title
DE102009049849B4 (en) Method for determining the pose of a camera, method for recognizing an object in a real environment and method for creating a data model
US10984602B1 (en) Facial expression tracking during augmented and virtual reality sessions
US20130022947A1 (en) Method and system for generating behavioral studies of an individual
CN114332374A (en) Virtual display method, equipment and storage medium
CN112040273B (en) Video synthesis method and device
WO2017177259A1 (en) System and method for processing photographic images
US20200257121A1 (en) Information processing method, information processing terminal, and computer-readable non-transitory storage medium storing program
CN113435236A (en) Home old man posture detection method, system, storage medium, equipment and application
CN113112612A (en) Positioning method and system for dynamic superposition of real person and mixed reality
CN112884556A (en) Shop display method, system, equipment and medium based on mixed reality
CN113641836A (en) Display method and related equipment thereof
CN117333645A (en) Annular holographic interaction system and equipment thereof
CN115933930A (en) Method, terminal and device for analyzing attention of learning object in education meta universe
CN115130493A (en) Face deformation recommendation method, device, equipment and medium based on image recognition
CN114387679A (en) System and method for realizing sight line estimation and attention analysis based on recursive convolutional neural network
CN114092674B (en) Multimedia data analysis method and system
CN109191229A (en) Augmented reality ornament recommended method and device
KR20160136833A (en) medical education system using video contents
CN117333644A (en) Virtual reality display picture generation method, device, equipment and medium
US20230103116A1 (en) Content utilization platform system and method of producing augmented reality (ar)-based image output
CN110866168A (en) Information recommendation method and device, terminal and server
Shasha et al. Object Recognition of Environmental Information in the Internet of Things Based on Augmented Reality
CN116484091B (en) Card information program interaction method and device
CN118364340B (en) Student course short board positioning method, system and storage medium based on deep learning
CN116091147A (en) Product recommendation method, device and equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant