CN107783652A - Implementation method, system and the device of virtual reality - Google Patents

Implementation method, system and the device of virtual reality Download PDF

Info

Publication number
CN107783652A
CN107783652A CN201710964798.5A CN201710964798A CN107783652A CN 107783652 A CN107783652 A CN 107783652A CN 201710964798 A CN201710964798 A CN 201710964798A CN 107783652 A CN107783652 A CN 107783652A
Authority
CN
China
Prior art keywords
virtual reality
destination object
depth
location information
action
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201710964798.5A
Other languages
Chinese (zh)
Other versions
CN107783652B (en
Inventor
刘杰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
GCI Science and Technology Co Ltd
Original Assignee
GCI Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by GCI Science and Technology Co Ltd filed Critical GCI Science and Technology Co Ltd
Priority to CN201710964798.5A priority Critical patent/CN107783652B/en
Publication of CN107783652A publication Critical patent/CN107783652A/en
Application granted granted Critical
Publication of CN107783652B publication Critical patent/CN107783652B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Multimedia (AREA)
  • User Interface Of Digital Computer (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The present invention relates to a kind of implementation method of virtual reality, system and device, computer-readable storage medium and equipment, wherein method includes:The surface each point of destination object in several adjacent timeslices is obtained relative to the depth location information of preset reference point;Position according to where the depth location information determines the predetermined patterns of the destination object;The action message of the destination object is determined according to the variable quantity of the position in different time piece;The operational factor for realizing system of virtual reality is adjusted according to the action message, and the virtual reality video flowing after output adjustment.Such scheme can realize virtual reality technology in the case of without extra body-sensing peripheral hardware, reduce the complexity of virtual reality operation.

Description

Implementation method, system and the device of virtual reality
Technical field
The present invention relates to technical field of virtual reality, more particularly to a kind of implementation method of virtual reality, system and dress Put.
Background technology
Virtual reality (Virtual Reality, abbreviation VR) technology is a kind of meter that can be created with the experiencing virtual world Calculation machine analogue system, it can generate a virtual information environment in Multi information space and be interacted with actual environment, and make User is on the spot in person, immerses wherein.
However, realization of the conventional art for all kinds of virtual realities supports it is by the peripheral hardware such as somatosensory handle, body-sensing headgear Realize, some related sensors, such as gyroscope, gravity sensor, pressure inductor etc. are carried on these body-sensing peripheral hardwares. After the user of virtual reality must wear these peripheral hardwares, the virtual reality video shown in virtual reality glasses can be coordinated to enter The problem of operation complexity is high be present in the corresponding action of row, such a implementation method.
The content of the invention
Based on this, it is necessary to for above-mentioned technical problem, there is provided a kind of implementation method for operating easy virtual reality, be System and device.
A kind of implementation method of virtual reality, comprises the following steps:
The surface each point for obtaining destination object in several adjacent timeslices is believed relative to the depth location of preset reference point Breath;
Position according to where the depth location information determines the predetermined patterns of the destination object;
The action message of the destination object is determined according to the variable quantity of the position in different time piece;
The operational factor for realizing system of virtual reality is adjusted according to the action message, and after output adjustment Virtual reality video flowing.
Wherein, the surface each point of destination object in several adjacent timeslices is obtained relative to the depth position of preset reference point The step of confidence ceases is performed by depth camera.
The data acquisition instructions of control terminal transmission are received by depth camera, according to the data acquisition instructions, are obtained The surface each point of destination object allows users to certainly relative to the depth location information of depth camera in several adjacent timeslices By the operating process of ground control virtual reality;Or if depth camera detects that destination object enters the visual field model of depth camera Enclose, then the surface each point for obtaining the destination object in several adjacent timeslices is believed relative to the depth location of depth camera Breath, by the automatic detection of depth camera, improve the convenience of virtual reality operation.
The surface each point of destination object in several adjacent timeslices is being obtained relative to the depth location of preset reference point After information, include determining the geometric center of destination object, and according to the geometric center, adjust the angle of the depth camera Degree, make the geometric center for including the destination object within sweep of the eye of the depth camera, so as to improve virtual reality The free degree so that the closer reality of virtual reality experience.
The implementation method of the virtual reality, the predetermined patterns of the destination object are determined according to the depth location information The step of position at place, including according to the depth location information, the destination object is generated in 3D depth coordinates system The 3D rendering of predetermined patterns, according to the 3D rendering, the position where the predetermined patterns of the destination object is obtained, so that Positioning to the position where destination object predetermined patterns is more accurate.
Wherein, above-mentioned action message includes the type of action and action extent index of destination object, and destination object is moved It is more accurate to make information identification.
The implementation method of the virtual reality, the target pair is determined according to the variable quantity of the position in different time piece The step of action message of elephant, including the variable quantity reference value of the variable quantity and corresponding site deliberate action type is carried out pair Than if the variable quantity is more than or equal to the variable quantity reference value, it is determined that the type of action of the destination object is default Type of action, according to the variable quantity of the position in different time piece, the action extent index of the destination object is determined, so as to Improve the degree of accuracy of identification destination object action message.
A kind of virtual reality realizes system, including:
Information acquisition module, for obtaining the surface each point of destination object in several adjacent timeslices relative to default ginseng The depth location information of examination point;
Data processing module, where determining the predetermined patterns of the destination object according to the depth location information Position;
Action recognition module, for determining the dynamic of the destination object according to the variable quantity of the position in different time piece Make information;
Parameter adjustment module, for being adjusted according to the action message to the operational factor for realizing system of virtual reality It is whole, and the virtual reality video flowing after output adjustment.
A kind of realization device of virtual reality, including:
Depth camera and intelligent gateway;
The depth camera obtains the surface each point of destination object in several adjacent timeslices relative to depth camera Depth location information, and the depth location information is sent to intelligent gateway;
Position of the intelligent gateway according to where the depth location information determines the predetermined patterns of the destination object, The action message of the destination object is determined according to the variable quantity of the position in different time piece, according to the action message pair The operational factor for realizing system of virtual reality is adjusted, and the virtual reality video flowing after output adjustment.
Wherein, the realization device of virtual reality also includes virtual reality glasses, and the virtual reality glasses are used for Rreceive output Adjustment after virtual reality video flowing, and shown, be easy to check the virtual reality video flowing after adjustment.
The realization device of virtual reality, in addition to control terminal, the control terminal are used to send number to the depth camera According to acquisition instructions, depth camera obtains destination object in several adjacent timeslices after the data acquisition instructions are received Surface each point relative to the depth location information of depth camera, improve the controlling of virtual reality operation.
A kind of computer-readable storage medium, is stored thereon with computer program, it is characterised in that the computer program is located Reason device realizes the implementation method of virtual reality when performing.
A kind of computer equipment, including memory, processor and storage can be run on a memory and on a processor Computer program, it is characterised in that the implementation method of virtual reality is realized during the computing device described program.
Implementation method, system and the device of above-mentioned virtual reality, computer-readable storage medium and equipment, according to the target of acquisition The depth location information of object, the predetermined patterns position of destination object is determined, and according to the position in different time piece Variable quantity determine the action message of destination object, virtual reality parameter is adjusted according to the action message, in the reality of virtual reality Without extra body-sensing peripheral hardware during existing, the complexity of virtual reality operation is reduced.
Brief description of the drawings
Fig. 1 is the implementation method flow chart of the virtual reality of one embodiment;
Fig. 2 is that the intelligent gateway of one embodiment identifies the schematic diagram of user action;
Fig. 3 is the structural representation for realizing system of the virtual reality of one embodiment;
Fig. 4 is the schematic diagram of the realization device of the virtual reality of one embodiment.
Embodiment
Technical scheme is described in detail with reference to specific embodiment and accompanying drawing, so that it is more It is clear.
As shown in figure 1, the present invention provides a kind of implementation method of virtual reality, there can be following steps:
S10, the surface each point of destination object in several adjacent timeslices is obtained relative to the depth position of preset reference point Confidence ceases;
This step can be performed by depth camera, and depth camera can be infrared Ancillary textures depth camera, flight time (Time Of Flight, abbreviation TOF) camera or other depth cameras.The number that control terminal is sent can be received by depth camera According to acquisition instructions, when can also detect that destination object enters the field range of depth camera by depth camera, make depth camera The surface each point of destination object in several adjacent timeslices is obtained relative to the depth location information of depth camera.Wherein, control Terminal processed can be the control of intelligent terminal such as mobile phone, tablet personal computer.
In actual applications, destination object can be the user using virtual reality device.For the ease of description, below Illustrated so that destination object is user as an example.
In one embodiment, user can make TOF camera by controlling terminal to send data acquisition instructions to TOF camera Start to gather in its visual field in several adjacent timeslices user relative to the depth location information of TOF camera.Pass through peripheral hardware control The mode of terminal triggering virtual reality processed, can make user more freely control the operating process of virtual reality.
In another embodiment, the depth camera uses infrared Ancillary textures binocular depth camera, works as depth camera Detect user take virtual reality glasses and enter depth camera within sweep of the eye when, 3D depth cameras are automatically according to the time The body of piece continuous collecting user relative to depth camera depth location information.By depth camera automatic detection and start to adopt The mode of collection, improve the convenience of virtual reality operation process.
The surface each point of destination object in several adjacent timeslices is being obtained relative to the depth location of preset reference point After information, the geometric center of destination object is may further determine that, and the angle of 3D depth cameras is adjusted according to the geometric center, makes depth Degree camera includes the geometric center within sweep of the eye.
In one embodiment, depth camera gathers in its visual field in several adjacent timeslices user relative to depth phase The depth location information of machine, and the geometric center of user is determined, according to the movement of this geometric center, adjust acquisition angles, control Depth camera includes this geometric center within sweep of the eye, so as to improve the free degree of virtual reality operation process so that empty Intend experience of reality closer to reality.
The depth information is exported, entered after the depth location information of the destination object is obtained by depth camera Enter step S20.
S20, the position according to where the depth location information determines the predetermined patterns of the destination object;
This step can be performed by intelligent gateway, and the intelligent gateway can establish wireless connection with above-mentioned depth camera, connection Mode may include Ethernet, USB or WIFI etc., so as to realize wireless telecommunications, improve the Discussing Convenience of virtual reality.Intelligence can be passed through Gateway receives the depth location information, and according to the depth location information, the pre- of destination object is generated in 3D depth coordinates system If the 3D rendering at position, and according to the 3D rendering, obtain the position where the predetermined patterns of the destination object.Wherein target pair The predetermined patterns of elephant can be the positions such as the trunk, arm, lower limb of user.
In one embodiment, the depth position by the user's body that intelligent gateway collects relative to depth camera Confidence ceases, and in 3D depth coordinates system, is schemed according to the 3D of the trunk of depth location information generation user, arm and lower extremity Picture, and according to the 3D rendering, obtain the position where user's trunk, arm and lower extremity.By in 3D depth coordinates system, Rebuild the 3D rendering of destination object predetermined patterns so that the positioning to the position where destination object predetermined patterns is more accurate Really.
S30, the action message of the destination object is determined according to the variable quantity of the position in different time piece;
This step can be performed by intelligent gateway, and intelligent gateway is according to the positions of different time piece destination object predetermined patterns Variable quantity, determine the action message of the destination object.Wherein, action message can be the type of action and action degree of destination object Parameter, the deliberate action Change of types amount reference value of the location variation of predetermined patterns and same area can be contrasted, when When the variable quantity is more than or equal to variable quantity reference value, the type of action that determines predetermined patterns is deliberate action type, and root The action extent index of the type of action of the destination object is determined according to the variable quantity of the position, passes through deliberate action type Variable quantity reference value can improve the degree of accuracy of identification destination object action message, and strengthen virtual reality realizes details.Wherein, move It may include that boxing, pin are kicked, the type such as embrace as type, action extent index can be the distance of action, the time of execution.
As shown in Fig. 2 the schematic diagram of user action is identified for intelligent gateway.
Specifically, user carries out reality-virtualizing game in the depth camera visual field, and it is each to gather user surface by depth camera The depth location information relative to camera is put, and this depth location information is returned to intelligent gateway, by intelligent gateway to the depth Degree positional information handled, determine user body parts, as trunk, arm, lower limb position in the position of different time piece Variable quantity, location variation and default running parameter are contrasted, so that it is determined that the action message of user, can be once to box Action, a kick, once embrace action.
In one embodiment, intelligent gateway is by the position of user's lower extremity in 3D depth coordinates system in different time piece The kicking variable quantity reference value of variable quantity and lower extremity is contrasted, and is determined that location variation is more than variable quantity reference value, is sentenced The action message for determining user is a kick, and the time that analysis obtains completing this kick is 3S, distance of kicking For 50cm.
In one of the embodiments, depth camera gathers the right hand position of user in different time piece relative to camera Depth location information, including the right hand depth location information of timeslice 1, the right hand depth location information of timeslice 2, depth camera pass through The mode of wireless connection is established with intelligent gateway and connected, and the data of acquisition are returned into intelligent gateway, and intelligent gateway is not according to With change in location of user's right hand position in 3D coordinate systems in timeslice, judge that user makes right hand punch action, and identify Go out the action extent index of this right hand boxing, including the completion punch action time is 2S, boxing distance is 20cm.
S40, the operational factor for realizing system of virtual reality is adjusted according to the action message, and output adjustment Virtual reality video flowing afterwards.
This step can be performed by intelligent gateway, and intelligent gateway is according to the action message of the destination object of acquisition to virtual existing Real realizes that the operational factor of system is adjusted, and the virtual reality video flowing after adjustment is exported.In a reality Apply in example, it has been determined that the action message of user is a kick, and distance is 50cm, and direction is due east direction, and side herein To having a vase from user 50cm, so as to which intelligent gateway is according to the kick of this user, adjust the realization system of virtual reality The operational factor of system, i.e., the scene that user's kicking smashes vase is realized in virtual reality, and will be virtual existing after adjustment Real video flowing is exported.
After virtual reality video flowing after output adjustment, step S10 can be returned to.
The implementation method of above-mentioned virtual reality, according to the depth location information of the destination object of acquisition, determine destination object Predetermined patterns position, and determine according to variable quantity of the position in different time piece the action message of destination object, Virtual reality parameter is adjusted according to the action message, user need not be worn outside extra body-sensing in the implementation process of virtual reality If reducing the complexity of virtual reality operation, also cause virtual reality experience closer to real-life real scene;And And this method the type of action and action extent index that can be made with interpretation user, has in the case where not wearing body-sensing peripheral hardware Help abundant virtual reality realizes details.
As shown in figure 3, the present invention a kind of virtual reality is also provided realize system, it may include:
Information acquisition module 10, for obtaining the surface each point of destination object in several adjacent timeslices relative to default The depth location information of reference point;
This functions of modules can be realized that depth camera can be infrared Ancillary textures depth camera, flight by depth camera Time (Time Of Flight, abbreviation TOF) camera or other depth cameras..Control terminal can be received by depth camera to send out The data acquisition instructions sent, when can also detect that destination object enters the field range of depth camera by depth camera, make depth Spend camera and obtain the surface each point of destination object in several adjacent timeslices relative to the depth location information of depth camera.Its In, control terminal can be the control of intelligent terminal such as mobile phone, tablet personal computer.
In actual applications, destination object can be the user using virtual reality device.For the ease of description, below Illustrated so that destination object is user as an example.
In one embodiment, user can make TOF camera by controlling terminal to send data acquisition instructions to TOF camera Start to gather in its visual field in several adjacent timeslices user relative to the depth location information of TOF camera.Pass through peripheral hardware control The mode of terminal triggering virtual reality processed, can make user more freely control the operating process of virtual reality.
In another embodiment, the depth camera uses infrared Ancillary textures binocular depth camera, works as depth camera Detect user take virtual reality glasses and enter depth camera within sweep of the eye when, 3D depth cameras are automatically according to the time The body of piece continuous collecting user relative to depth camera depth location information.By depth camera automatic detection and start to adopt The mode of collection, improve the convenience of virtual reality operation process.
The surface each point of destination object in several adjacent timeslices is being obtained relative to the depth location of preset reference point , can also be by determining the geometric center of destination object after information, and the angle of 3D depth cameras is adjusted according to the geometric center, Make depth camera includes the geometric center within sweep of the eye.
In one embodiment, depth camera gathers in its visual field in several adjacent timeslices user relative to depth phase The depth location information of machine, and the geometric center of user is determined, according to the movement of this geometric center, adjust acquisition angles, control Depth camera includes this geometric center within sweep of the eye, so as to improve the free degree of virtual reality operation process so that empty Intend experience of reality closer to reality.
Information acquisition module 10 is carried out the depth information defeated after the depth location information of the destination object is obtained Go out, into data processing module 20.
Data processing module 20, for determining the predetermined patterns place of the destination object according to the depth location information Position;
This functions of modules can realize that the intelligent gateway can be with above-mentioned depth phase by the data processing unit on intelligent gateway Machine establishes wireless connection, and connected mode may include Ethernet, USB or WIFI etc., so as to realize wireless telecommunications, improves virtual reality Discussing Convenience.Depth location information can be received by data processing unit on intelligent gateway, according to the depth location information, The 3D rendering of the predetermined patterns of destination object is generated in 3D depth coordinates system, and according to the 3D rendering, obtains the destination object Predetermined patterns where position.Wherein the predetermined patterns of destination object can be the positions such as the trunk, arm, lower limb of user.
In one embodiment, the user's body collected by the processing unit on intelligent gateway is relative to depth The depth location information of camera, in 3D depth coordinates system, according to the depth location information generate the trunk of user, arm and under The 3D rendering at limb position, and according to the 3D rendering, obtain the position where user's trunk, arm and lower extremity.By in 3D In depth coordinate system, the 3D rendering of destination object predetermined patterns is rebuild so as to the position where destination object predetermined patterns The positioning put is more accurate.
Action recognition module 30, for determining the destination object according to the variable quantity of the position in different time piece Action message;
This functions of modules can realize that action recognition unit is according to different time by the action recognition unit on intelligent gateway The location variation of piece destination object predetermined patterns, determine the action message of the destination object.Wherein, action message can be target The type of action and action extent index of object, can be by the location variation of predetermined patterns and the deliberate action type of same area Variable quantity reference value is contrasted, and when the variable quantity is more than or equal to variable quantity reference value, determines the action of predetermined patterns Type is deliberate action type, and determines according to the variable quantity of the position action degree of the type of action of the destination object Parameter, the degree of accuracy of identification destination object action message can be improved by the variable quantity reference value of deliberate action type, and enhancing is empty That intends reality realizes details.Wherein, type of action may include that boxing, pin such as kick, embraced at the type, and action extent index can be The distance of work, the time of execution.
Specifically, user carries out reality-virtualizing game in the depth camera visual field, and it is each to gather user surface by depth camera The depth location information relative to camera is put, and this depth location information is returned to intelligent gateway, by intelligent gateway to the depth Degree positional information handled, determine user body parts, as trunk, arm, lower limb position in the position of different time piece Variable quantity, location variation and default running parameter are contrasted, so that it is determined that the action message of user, can be once to box Action, a kick, once embrace action.
In one embodiment, the action recognition unit on intelligent gateway will be used in different time piece in 3D depth coordinates system The location variation of family lower extremity and the kicking variable quantity reference value of lower extremity are contrasted, and determine that location variation is more than Variable quantity reference value, the action message for judging user are a kick, and analysis obtain completing this kick when Between be 3S, kicking distance be 50cm.
Parameter adjustment module 40, for being carried out according to the action message to the operational factor for realizing system of virtual reality Adjustment, and the virtual reality video flowing after output adjustment.
This functions of modules can be performed by the parameter adjustment unit on intelligent gateway, and parameter adjustment unit is according to the mesh of acquisition The action message of mark object is adjusted to the operational factor for realizing system of virtual reality, and by the virtual reality after adjustment Video flowing is exported.In one embodiment, it has been determined that the action message of user is a kick, and distance is 50cm, Direction is due east direction, and in this direction, has a vase from user 50cm, so as to parameter adjustment unit kicking according to this user Leg acts, and adjusts the operational factor for realizing system of virtual reality, i.e., realizes that user's kicking smashes vase in virtual reality Scene, and the virtual reality video flowing after adjustment is exported.
Said system, without extra body-sensing peripheral hardware, reduces virtual reality operation in the implementation process of virtual reality Complexity, also cause virtual reality experience closer to real-life real scene;And the system is not being worn outside body-sensing In the case of if, the type of action and action extent index that can be made with interpretation user, contribute to the realization of abundant virtual reality Details.
The implementation method for realizing system and the virtual reality of the present invention of the virtual reality of the present invention corresponds, above-mentioned The technical characteristic and its advantage that the embodiment of the implementation method of virtual reality illustrates are applied to the realization system of virtual reality In the embodiment of system, hereby give notice that.
As shown in figure 4, the present invention also provides a kind of realization device of virtual reality, it may include:
Depth camera 101 and intelligent gateway 201;
The depth camera 101 obtains the surface each point of destination object in several adjacent timeslices relative to depth camera Depth location information, and the depth location information is sent to intelligent gateway 201;
Position of the intelligent gateway 201 according to where the depth location information determines the predetermined patterns of the destination object Put, the action message of the destination object is determined according to the variable quantity of the position in different time piece, believed according to the action Cease and the operational factor for realizing system of virtual reality is adjusted, and the virtual reality video flowing after output adjustment.
Be the field range of depth camera 101 on the left of Fig. 4, depth camera 101 can within sweep of the eye it is adjacent several when Between in piece the surface each point of destination object be acquired relative to the depth location information of camera.
Depth camera 101 can be infrared Ancillary textures depth camera, flight time (Time Of Flight, abbreviation TOF) camera or other depth cameras.The data acquisition instructions that control terminal is sent can be received by depth camera 101, can also be led to When crossing depth camera 101 and detecting that destination object enters the field range of camera, camera is set to obtain in adjacent several timeslices The surface each point of destination object relative to camera depth location information.Wherein, it can be mobile phone, tablet personal computer etc. to control terminal Control of intelligent terminal.
In actual applications, destination object can be the user using virtual reality device.For the ease of description, below Illustrated so that destination object is user as an example.
In one embodiment, user can make TOF camera by controlling terminal to send data acquisition instructions to TOF camera Start to gather in its visual field in several adjacent timeslices user relative to the depth location information of TOF camera.Pass through peripheral hardware control The mode of terminal triggering virtual reality processed, can make user more freely control the operating process of virtual reality.
In another embodiment, the depth camera uses infrared Ancillary textures binocular depth camera, works as depth camera Detect user take virtual reality glasses and enter depth camera within sweep of the eye when, 3D depth cameras are automatically according to the time The body of piece continuous collecting user relative to depth camera depth location information.By depth camera automatic detection and start to adopt The mode of collection, improve the convenience of virtual reality operation process.
The surface each point of destination object in several adjacent timeslices is being obtained relative to the depth location of preset reference point After information, the geometric center of destination object is may further determine that, and according to the angle of the geometric center percentage regulation camera 101, make Depth camera 101 includes the geometric center within sweep of the eye.
In one embodiment, depth camera 101 gathers in its visual field in several adjacent timeslices user relative to depth The depth location information of camera 101 is spent, and determines the geometric center of user, according to the movement of this geometric center, adjusts acquisition angle Degree, controlling depth camera 101 includes this geometric center within sweep of the eye, so as to improve the freedom of virtual reality operation process Degree so that the closer reality of virtual reality experience.
Depth camera 101 is returned the depth information after the depth location information of the destination object is obtained To the intelligent gateway 201 of connection.
It is the intelligent gateway 201 being connected with depth camera 101 on the right side of Fig. 4, the intelligent gateway 201 can be with depth camera 101 Wireless connection is established, connected mode may include Ethernet, USB or WIFI etc., so as to realize wireless telecommunications, improve virtual reality Discussing Convenience.The depth location information can be received by the data processing unit 202 on intelligent gateway 201, according to the depth position Confidence is ceased, and the 3D rendering of the predetermined patterns of destination object is generated in 3D depth coordinates system, and according to the 3D rendering, described in acquisition Position where the predetermined patterns of destination object.Wherein the predetermined patterns of destination object can be trunk, arm, lower limb of user etc. Position.
In one embodiment, the user's body collected by the data processing unit 202 on intelligent gateway 201 Relative to the depth location information of depth camera, in 3D depth coordinates system, the body of user is generated according to the depth location information The 3D rendering of dry, arm and lower extremity, and according to the 3D rendering, obtain the position where user's trunk, arm and lower extremity Put.By in 3D depth coordinates system, rebuilding the 3D rendering of destination object predetermined patterns so that preset portion to destination object The positioning of position where position is more accurate.
Action recognition unit 203 on intelligent gateway 201 receives the positional information that data processing unit 202 is sent, according to The variable quantity of the position determines the action message of the destination object in different time piece.Wherein, action message can be target The type of action and action extent index of object, can be by the location variation of predetermined patterns and the deliberate action type of same area Variable quantity reference value is contrasted, and when the variable quantity is more than or equal to variable quantity reference value, determines the action of predetermined patterns Type is deliberate action type, and determines according to the variable quantity of the position action degree of the type of action of the destination object Parameter, the degree of accuracy of identification destination object action message can be improved by the variable quantity reference value of deliberate action type, and enhancing is empty That intends reality realizes details.Wherein, type of action may include that boxing, pin such as kick, embraced at the type, and action extent index can be The distance of work, the time of execution.
In one embodiment, the action recognition unit 203 on intelligent gateway 201 is by 3D depth coordinates in different time piece The kicking variable quantity reference value of the location variation of user's lower extremity and lower extremity is contrasted in system, determines change in location Amount is more than variable quantity reference value, and the action message for judging user be a kick, and analysis obtains completing this and kicked to move The time of work is 3S, and kicking distance is 50cm.
The action for the destination object that game adjustment unit 204 on intelligent gateway 201 obtains according to action recognition unit 203 Information is adjusted to the operational factor for realizing system of virtual reality, and the virtual reality video flowing progress after adjustment is defeated Go out.In one embodiment, action recognition unit 203 has determined that the action message of user is a kick, and distance is 50cm, direction are due east direction, and in this direction, have a vase from user 50cm, so as to the game adjustment on intelligent gateway 201 Unit 204 adjusts the operational factor for realizing system of virtual reality, i.e., in virtual reality according to the kick of this user The scene that user's kicking smashes vase is realized, and the virtual reality video flowing after adjustment is exported.
Specifically, the data processing unit 202 on above-mentioned intelligent gateway 201, action recognition unit 203 and parameter adjustment list Member 204 can be the software service operated on intelligent gateway, so as to reduce the cost of manufacture additional hardware element.
Further, a kind of realization device of virtual reality, may also include:
Mobile phone 301 and virtual reality glasses 401;
Wherein, mobile phone 301 is used to establish with depth camera 101 and connected, and sends data acquisition to depth camera 101 and refer to Order, improve the controlling of virtual reality operation.In one embodiment, the mobile phone 301 is established with depth camera 101 and connected Connect, and data acquisition instructions are sent to depth camera 101.Depth camera 101 obtains after the data acquisition instructions are received In several adjacent timeslices the surface each point of destination object relative to depth camera depth location information.
Wherein, virtual reality glasses 401 are used to receive the virtual reality video flowing after the adjustment of the output of intelligent gateway 201, And shown, it is easy to check the virtual reality video flowing after adjustment.
The realization device of above-mentioned virtual reality, without extra body-sensing peripheral hardware in the implementation process of virtual reality, reduce The complexity of virtual reality operation, also cause virtual reality experience closer to real-life real scene;And this dress Put in the case where not wearing body-sensing peripheral hardware, the type of action and action extent index that can be made with interpretation user, contribute to rich Rich virtual reality realizes details.
The present invention also provides a kind of computer-readable recording medium, is stored thereon with computer program, the program is processed Device realizes the implementation method of the virtual reality in above-mentioned any embodiment when performing.Performed by the computer-readable recording medium Method is identical with the implementation method of the virtual reality in above-described embodiment, and here is omitted.
The more specifically example (non-exhaustive list) of computer-readable medium includes following:Connected up with one or more Electrical connection section (electronic installation), portable computer diskette box (magnetic device), random access memory (RAM), read-only storage (ROM), erasable edit read-only storage (EPROM or flash memory), fiber device, and portable optic disk is read-only deposits Reservoir (CDROM).In addition, computer-readable medium, which can even is that, to print the paper of described program thereon or other are suitable Medium, because can then enter edlin, interpretation or if necessary with it for example by carrying out optical scanner to paper or other media His suitable method is handled electronically to obtain described program, is then stored in computer storage.
The present invention also provides a kind of computer equipment, including memory, processor and storage on a memory and can located The computer program that runs on reason device, the virtual reality in above-mentioned any embodiment is realized during the computing device described program Implementation method.The realization side of the virtual reality in the method and above-described embodiment performed by processor in the computer equipment Method is identical, and here is omitted.
Embodiment described above only expresses the several embodiments of the present invention, and its description is more specific and detailed, but simultaneously Therefore the limitation to the scope of the claims of the present invention can not be interpreted as.It should be pointed out that for one of ordinary skill in the art For, without departing from the inventive concept of the premise, various modifications and improvements can be made, these belong to the guarantor of the present invention Protect scope.Therefore, the protection domain of patent of the present invention should be determined by the appended claims.

Claims (10)

1. a kind of implementation method of virtual reality, it is characterised in that comprise the following steps:
The surface each point of destination object in several adjacent timeslices is obtained relative to the depth location information of preset reference point;
Position according to where the depth location information determines the predetermined patterns of the destination object;
The action message of the destination object is determined according to the variable quantity of the position in different time piece;
The operational factor for realizing system of virtual reality is adjusted according to the action message, and it is virtual after output adjustment Real video flowing.
2. the implementation method of virtual reality according to claim 1, it is characterised in that obtain in several adjacent timeslices The surface each point of destination object relative to preset reference point depth location information the step of performed by depth camera;
The surface each point of destination object in several adjacent timeslices is obtained relative to the depth location information of preset reference point Step, including:
The data acquisition instructions of control terminal transmission are received by depth camera, according to the data acquisition instructions, are obtained adjacent In several timeslices the surface each point of destination object relative to depth camera depth location information;
Or;
If detecting, destination object enters the field range of depth camera, obtains the target pair in several adjacent timeslices The surface each point of elephant relative to depth camera depth location information.
3. the implementation method of virtual reality according to claim 2, it is characterised in that obtaining several adjacent timeslices After the surface each point of interior destination object is relative to the depth location information of preset reference point, including:
Determine the geometric center of destination object;
According to the geometric center, the angle of the depth camera is adjusted, make the depth camera includes institute within sweep of the eye State the geometric center of destination object.
4. the implementation method of virtual reality according to claim 1, it is characterised in that true according to the depth location information The step of position where the predetermined patterns of the fixed destination object, including:
According to the depth location information, the 3D rendering of the predetermined patterns of the destination object is generated in 3D depth coordinates system;
According to the 3D rendering, the position where the predetermined patterns of the destination object is obtained.
5. the implementation method of the virtual reality according to Claims 1-4 any one, it is characterised in that the action letter Breath includes the type of action and action extent index of destination object.
6. the implementation method of virtual reality according to claim 5, it is characterised in that according to institute's rheme in different time piece The variable quantity put determines the step of action message of the destination object, including:
The variable quantity and the variable quantity reference value of corresponding site deliberate action type are contrasted;
If the variable quantity is more than or equal to the variable quantity reference value, it is determined that the type of action of the destination object is default Type of action;
According to the variable quantity of the position in different time piece, the action extent index of the destination object is determined.
7. a kind of virtual reality realizes system, it is characterised in that including:
Information acquisition module, for obtaining the surface each point of destination object in several adjacent timeslices relative to preset reference point Depth location information;
Data processing module, for the position where determining the predetermined patterns of the destination object according to the depth location information Put;
Action recognition module, for determining that the action of the destination object is believed according to the variable quantity of the position in different time piece Breath;
Parameter adjustment module, for being adjusted according to the action message to the operational factor for realizing system of virtual reality, And the virtual reality video flowing after output adjustment.
A kind of 8. realization device of virtual reality, it is characterised in that including:
Depth camera and intelligent gateway;
The depth camera obtains the surface each point of destination object in several adjacent timeslices relative to the depth of depth camera Positional information, and the depth location information is sent to intelligent gateway;
Position of the intelligent gateway according to where the depth location information determines the predetermined patterns of the destination object, according to The variable quantity of the position determines the action message of the destination object in different time piece, according to the action message to virtual Reality realizes that the operational factor of system is adjusted, and the virtual reality video flowing after output adjustment.
9. the realization device of virtual reality according to claim 8, it is characterised in that also include:
Virtual reality glasses;
The virtual reality video flowing that the virtual reality glasses are used for after the adjustment of Rreceive output, and shown.
10. the realization device of virtual reality according to claim 8, it is characterised in that also include:
Control terminal;
The control terminal is used to send data acquisition instructions to the depth camera;
For depth camera after the data acquisition instructions are received, the surface for obtaining destination object in several adjacent timeslices is each Put the depth location information relative to depth camera.
CN201710964798.5A 2017-10-17 2017-10-17 Method, system and device for realizing virtual reality Active CN107783652B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710964798.5A CN107783652B (en) 2017-10-17 2017-10-17 Method, system and device for realizing virtual reality

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710964798.5A CN107783652B (en) 2017-10-17 2017-10-17 Method, system and device for realizing virtual reality

Publications (2)

Publication Number Publication Date
CN107783652A true CN107783652A (en) 2018-03-09
CN107783652B CN107783652B (en) 2020-11-13

Family

ID=61434535

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710964798.5A Active CN107783652B (en) 2017-10-17 2017-10-17 Method, system and device for realizing virtual reality

Country Status (1)

Country Link
CN (1) CN107783652B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114625251A (en) * 2022-03-11 2022-06-14 平安普惠企业管理有限公司 Interaction method and device based on VR, computer equipment and storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103529944A (en) * 2013-10-17 2014-01-22 合肥金诺数码科技股份有限公司 Human body movement identification method based on Kinect
US20150040174A1 (en) * 2013-08-01 2015-02-05 Joiz Ip Ag System and method for synchronizing media platform devices
CN104460962A (en) * 2013-09-18 2015-03-25 天津联合动力信息技术有限公司 4D somatosensory interaction system based on game engine
CN106295479A (en) * 2015-06-05 2017-01-04 上海戏剧学院 Based on body-sensing technology action recognition editing system
CN106601062A (en) * 2016-11-22 2017-04-26 山东科技大学 Interactive method for simulating mine disaster escape training
CN107133984A (en) * 2017-03-24 2017-09-05 深圳奥比中光科技有限公司 The scaling method and system of depth camera and main equipment

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150040174A1 (en) * 2013-08-01 2015-02-05 Joiz Ip Ag System and method for synchronizing media platform devices
CN104460962A (en) * 2013-09-18 2015-03-25 天津联合动力信息技术有限公司 4D somatosensory interaction system based on game engine
CN103529944A (en) * 2013-10-17 2014-01-22 合肥金诺数码科技股份有限公司 Human body movement identification method based on Kinect
CN106295479A (en) * 2015-06-05 2017-01-04 上海戏剧学院 Based on body-sensing technology action recognition editing system
CN106601062A (en) * 2016-11-22 2017-04-26 山东科技大学 Interactive method for simulating mine disaster escape training
CN107133984A (en) * 2017-03-24 2017-09-05 深圳奥比中光科技有限公司 The scaling method and system of depth camera and main equipment

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114625251A (en) * 2022-03-11 2022-06-14 平安普惠企业管理有限公司 Interaction method and device based on VR, computer equipment and storage medium

Also Published As

Publication number Publication date
CN107783652B (en) 2020-11-13

Similar Documents

Publication Publication Date Title
CN107990899A (en) A kind of localization method and system based on SLAM
EP2917754B1 (en) Image processing method, particularly used in a vision-based localization of a device
US10802606B2 (en) Method and device for aligning coordinate of controller or headset with coordinate of binocular system
CN112711249B (en) Robot positioning method and device, intelligent robot and storage medium
CN105094335A (en) Scene extracting method, object positioning method and scene extracting system
CN107665508B (en) Method and system for realizing augmented reality
CN108885487B (en) Gesture control method of wearable system and wearable system
CN106575437A (en) Information-processing device, information processing method, and program
CN107665505A (en) The method and device of augmented reality is realized based on plane monitoring-network
CN110276251B (en) Image recognition method, device, equipment and storage medium
US11127156B2 (en) Method of device tracking, terminal device, and storage medium
CN110456904B (en) Augmented reality glasses eye movement interaction method and system without calibration
CN106504227B (en) Demographic method and its system based on depth image
CN107025661A (en) A kind of method for realizing augmented reality, server, terminal and system
CN109902675B (en) Object pose acquisition method and scene reconstruction method and device
CN109740659A (en) A kind of image matching method and device, electronic equipment, storage medium
CN106652028A (en) Environment three-dimensional mapping method and apparatus
US11010916B2 (en) Method of configuring camera position suitable for localization and robot implementing same
CN110096152A (en) Space-location method, device, equipment and the storage medium of physical feeling
CN111354029A (en) Gesture depth determination method, device, equipment and storage medium
CN109191522B (en) Robot displacement correction method and system based on three-dimensional modeling
CN107783652A (en) Implementation method, system and the device of virtual reality
CN116958452A (en) Three-dimensional reconstruction method and system
KR20160099981A (en) Virtual reality device based on two-way recognition including a three-dimensional marker with a patten for multiple users
CN107831921A (en) A kind of handle space position and the determination method, apparatus and system of coding corresponding relation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant