CN107783652A - Implementation method, system and the device of virtual reality - Google Patents
Implementation method, system and the device of virtual reality Download PDFInfo
- Publication number
- CN107783652A CN107783652A CN201710964798.5A CN201710964798A CN107783652A CN 107783652 A CN107783652 A CN 107783652A CN 201710964798 A CN201710964798 A CN 201710964798A CN 107783652 A CN107783652 A CN 107783652A
- Authority
- CN
- China
- Prior art keywords
- virtual reality
- destination object
- depth
- location information
- action
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- General Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Psychiatry (AREA)
- Social Psychology (AREA)
- Multimedia (AREA)
- User Interface Of Digital Computer (AREA)
- Processing Or Creating Images (AREA)
Abstract
The present invention relates to a kind of implementation method of virtual reality, system and device, computer-readable storage medium and equipment, wherein method includes:The surface each point of destination object in several adjacent timeslices is obtained relative to the depth location information of preset reference point;Position according to where the depth location information determines the predetermined patterns of the destination object;The action message of the destination object is determined according to the variable quantity of the position in different time piece;The operational factor for realizing system of virtual reality is adjusted according to the action message, and the virtual reality video flowing after output adjustment.Such scheme can realize virtual reality technology in the case of without extra body-sensing peripheral hardware, reduce the complexity of virtual reality operation.
Description
Technical field
The present invention relates to technical field of virtual reality, more particularly to a kind of implementation method of virtual reality, system and dress
Put.
Background technology
Virtual reality (Virtual Reality, abbreviation VR) technology is a kind of meter that can be created with the experiencing virtual world
Calculation machine analogue system, it can generate a virtual information environment in Multi information space and be interacted with actual environment, and make
User is on the spot in person, immerses wherein.
However, realization of the conventional art for all kinds of virtual realities supports it is by the peripheral hardware such as somatosensory handle, body-sensing headgear
Realize, some related sensors, such as gyroscope, gravity sensor, pressure inductor etc. are carried on these body-sensing peripheral hardwares.
After the user of virtual reality must wear these peripheral hardwares, the virtual reality video shown in virtual reality glasses can be coordinated to enter
The problem of operation complexity is high be present in the corresponding action of row, such a implementation method.
The content of the invention
Based on this, it is necessary to for above-mentioned technical problem, there is provided a kind of implementation method for operating easy virtual reality, be
System and device.
A kind of implementation method of virtual reality, comprises the following steps:
The surface each point for obtaining destination object in several adjacent timeslices is believed relative to the depth location of preset reference point
Breath;
Position according to where the depth location information determines the predetermined patterns of the destination object;
The action message of the destination object is determined according to the variable quantity of the position in different time piece;
The operational factor for realizing system of virtual reality is adjusted according to the action message, and after output adjustment
Virtual reality video flowing.
Wherein, the surface each point of destination object in several adjacent timeslices is obtained relative to the depth position of preset reference point
The step of confidence ceases is performed by depth camera.
The data acquisition instructions of control terminal transmission are received by depth camera, according to the data acquisition instructions, are obtained
The surface each point of destination object allows users to certainly relative to the depth location information of depth camera in several adjacent timeslices
By the operating process of ground control virtual reality;Or if depth camera detects that destination object enters the visual field model of depth camera
Enclose, then the surface each point for obtaining the destination object in several adjacent timeslices is believed relative to the depth location of depth camera
Breath, by the automatic detection of depth camera, improve the convenience of virtual reality operation.
The surface each point of destination object in several adjacent timeslices is being obtained relative to the depth location of preset reference point
After information, include determining the geometric center of destination object, and according to the geometric center, adjust the angle of the depth camera
Degree, make the geometric center for including the destination object within sweep of the eye of the depth camera, so as to improve virtual reality
The free degree so that the closer reality of virtual reality experience.
The implementation method of the virtual reality, the predetermined patterns of the destination object are determined according to the depth location information
The step of position at place, including according to the depth location information, the destination object is generated in 3D depth coordinates system
The 3D rendering of predetermined patterns, according to the 3D rendering, the position where the predetermined patterns of the destination object is obtained, so that
Positioning to the position where destination object predetermined patterns is more accurate.
Wherein, above-mentioned action message includes the type of action and action extent index of destination object, and destination object is moved
It is more accurate to make information identification.
The implementation method of the virtual reality, the target pair is determined according to the variable quantity of the position in different time piece
The step of action message of elephant, including the variable quantity reference value of the variable quantity and corresponding site deliberate action type is carried out pair
Than if the variable quantity is more than or equal to the variable quantity reference value, it is determined that the type of action of the destination object is default
Type of action, according to the variable quantity of the position in different time piece, the action extent index of the destination object is determined, so as to
Improve the degree of accuracy of identification destination object action message.
A kind of virtual reality realizes system, including:
Information acquisition module, for obtaining the surface each point of destination object in several adjacent timeslices relative to default ginseng
The depth location information of examination point;
Data processing module, where determining the predetermined patterns of the destination object according to the depth location information
Position;
Action recognition module, for determining the dynamic of the destination object according to the variable quantity of the position in different time piece
Make information;
Parameter adjustment module, for being adjusted according to the action message to the operational factor for realizing system of virtual reality
It is whole, and the virtual reality video flowing after output adjustment.
A kind of realization device of virtual reality, including:
Depth camera and intelligent gateway;
The depth camera obtains the surface each point of destination object in several adjacent timeslices relative to depth camera
Depth location information, and the depth location information is sent to intelligent gateway;
Position of the intelligent gateway according to where the depth location information determines the predetermined patterns of the destination object,
The action message of the destination object is determined according to the variable quantity of the position in different time piece, according to the action message pair
The operational factor for realizing system of virtual reality is adjusted, and the virtual reality video flowing after output adjustment.
Wherein, the realization device of virtual reality also includes virtual reality glasses, and the virtual reality glasses are used for Rreceive output
Adjustment after virtual reality video flowing, and shown, be easy to check the virtual reality video flowing after adjustment.
The realization device of virtual reality, in addition to control terminal, the control terminal are used to send number to the depth camera
According to acquisition instructions, depth camera obtains destination object in several adjacent timeslices after the data acquisition instructions are received
Surface each point relative to the depth location information of depth camera, improve the controlling of virtual reality operation.
A kind of computer-readable storage medium, is stored thereon with computer program, it is characterised in that the computer program is located
Reason device realizes the implementation method of virtual reality when performing.
A kind of computer equipment, including memory, processor and storage can be run on a memory and on a processor
Computer program, it is characterised in that the implementation method of virtual reality is realized during the computing device described program.
Implementation method, system and the device of above-mentioned virtual reality, computer-readable storage medium and equipment, according to the target of acquisition
The depth location information of object, the predetermined patterns position of destination object is determined, and according to the position in different time piece
Variable quantity determine the action message of destination object, virtual reality parameter is adjusted according to the action message, in the reality of virtual reality
Without extra body-sensing peripheral hardware during existing, the complexity of virtual reality operation is reduced.
Brief description of the drawings
Fig. 1 is the implementation method flow chart of the virtual reality of one embodiment;
Fig. 2 is that the intelligent gateway of one embodiment identifies the schematic diagram of user action;
Fig. 3 is the structural representation for realizing system of the virtual reality of one embodiment;
Fig. 4 is the schematic diagram of the realization device of the virtual reality of one embodiment.
Embodiment
Technical scheme is described in detail with reference to specific embodiment and accompanying drawing, so that it is more
It is clear.
As shown in figure 1, the present invention provides a kind of implementation method of virtual reality, there can be following steps:
S10, the surface each point of destination object in several adjacent timeslices is obtained relative to the depth position of preset reference point
Confidence ceases;
This step can be performed by depth camera, and depth camera can be infrared Ancillary textures depth camera, flight time
(Time Of Flight, abbreviation TOF) camera or other depth cameras.The number that control terminal is sent can be received by depth camera
According to acquisition instructions, when can also detect that destination object enters the field range of depth camera by depth camera, make depth camera
The surface each point of destination object in several adjacent timeslices is obtained relative to the depth location information of depth camera.Wherein, control
Terminal processed can be the control of intelligent terminal such as mobile phone, tablet personal computer.
In actual applications, destination object can be the user using virtual reality device.For the ease of description, below
Illustrated so that destination object is user as an example.
In one embodiment, user can make TOF camera by controlling terminal to send data acquisition instructions to TOF camera
Start to gather in its visual field in several adjacent timeslices user relative to the depth location information of TOF camera.Pass through peripheral hardware control
The mode of terminal triggering virtual reality processed, can make user more freely control the operating process of virtual reality.
In another embodiment, the depth camera uses infrared Ancillary textures binocular depth camera, works as depth camera
Detect user take virtual reality glasses and enter depth camera within sweep of the eye when, 3D depth cameras are automatically according to the time
The body of piece continuous collecting user relative to depth camera depth location information.By depth camera automatic detection and start to adopt
The mode of collection, improve the convenience of virtual reality operation process.
The surface each point of destination object in several adjacent timeslices is being obtained relative to the depth location of preset reference point
After information, the geometric center of destination object is may further determine that, and the angle of 3D depth cameras is adjusted according to the geometric center, makes depth
Degree camera includes the geometric center within sweep of the eye.
In one embodiment, depth camera gathers in its visual field in several adjacent timeslices user relative to depth phase
The depth location information of machine, and the geometric center of user is determined, according to the movement of this geometric center, adjust acquisition angles, control
Depth camera includes this geometric center within sweep of the eye, so as to improve the free degree of virtual reality operation process so that empty
Intend experience of reality closer to reality.
The depth information is exported, entered after the depth location information of the destination object is obtained by depth camera
Enter step S20.
S20, the position according to where the depth location information determines the predetermined patterns of the destination object;
This step can be performed by intelligent gateway, and the intelligent gateway can establish wireless connection with above-mentioned depth camera, connection
Mode may include Ethernet, USB or WIFI etc., so as to realize wireless telecommunications, improve the Discussing Convenience of virtual reality.Intelligence can be passed through
Gateway receives the depth location information, and according to the depth location information, the pre- of destination object is generated in 3D depth coordinates system
If the 3D rendering at position, and according to the 3D rendering, obtain the position where the predetermined patterns of the destination object.Wherein target pair
The predetermined patterns of elephant can be the positions such as the trunk, arm, lower limb of user.
In one embodiment, the depth position by the user's body that intelligent gateway collects relative to depth camera
Confidence ceases, and in 3D depth coordinates system, is schemed according to the 3D of the trunk of depth location information generation user, arm and lower extremity
Picture, and according to the 3D rendering, obtain the position where user's trunk, arm and lower extremity.By in 3D depth coordinates system,
Rebuild the 3D rendering of destination object predetermined patterns so that the positioning to the position where destination object predetermined patterns is more accurate
Really.
S30, the action message of the destination object is determined according to the variable quantity of the position in different time piece;
This step can be performed by intelligent gateway, and intelligent gateway is according to the positions of different time piece destination object predetermined patterns
Variable quantity, determine the action message of the destination object.Wherein, action message can be the type of action and action degree of destination object
Parameter, the deliberate action Change of types amount reference value of the location variation of predetermined patterns and same area can be contrasted, when
When the variable quantity is more than or equal to variable quantity reference value, the type of action that determines predetermined patterns is deliberate action type, and root
The action extent index of the type of action of the destination object is determined according to the variable quantity of the position, passes through deliberate action type
Variable quantity reference value can improve the degree of accuracy of identification destination object action message, and strengthen virtual reality realizes details.Wherein, move
It may include that boxing, pin are kicked, the type such as embrace as type, action extent index can be the distance of action, the time of execution.
As shown in Fig. 2 the schematic diagram of user action is identified for intelligent gateway.
Specifically, user carries out reality-virtualizing game in the depth camera visual field, and it is each to gather user surface by depth camera
The depth location information relative to camera is put, and this depth location information is returned to intelligent gateway, by intelligent gateway to the depth
Degree positional information handled, determine user body parts, as trunk, arm, lower limb position in the position of different time piece
Variable quantity, location variation and default running parameter are contrasted, so that it is determined that the action message of user, can be once to box
Action, a kick, once embrace action.
In one embodiment, intelligent gateway is by the position of user's lower extremity in 3D depth coordinates system in different time piece
The kicking variable quantity reference value of variable quantity and lower extremity is contrasted, and is determined that location variation is more than variable quantity reference value, is sentenced
The action message for determining user is a kick, and the time that analysis obtains completing this kick is 3S, distance of kicking
For 50cm.
In one of the embodiments, depth camera gathers the right hand position of user in different time piece relative to camera
Depth location information, including the right hand depth location information of timeslice 1, the right hand depth location information of timeslice 2, depth camera pass through
The mode of wireless connection is established with intelligent gateway and connected, and the data of acquisition are returned into intelligent gateway, and intelligent gateway is not according to
With change in location of user's right hand position in 3D coordinate systems in timeslice, judge that user makes right hand punch action, and identify
Go out the action extent index of this right hand boxing, including the completion punch action time is 2S, boxing distance is 20cm.
S40, the operational factor for realizing system of virtual reality is adjusted according to the action message, and output adjustment
Virtual reality video flowing afterwards.
This step can be performed by intelligent gateway, and intelligent gateway is according to the action message of the destination object of acquisition to virtual existing
Real realizes that the operational factor of system is adjusted, and the virtual reality video flowing after adjustment is exported.In a reality
Apply in example, it has been determined that the action message of user is a kick, and distance is 50cm, and direction is due east direction, and side herein
To having a vase from user 50cm, so as to which intelligent gateway is according to the kick of this user, adjust the realization system of virtual reality
The operational factor of system, i.e., the scene that user's kicking smashes vase is realized in virtual reality, and will be virtual existing after adjustment
Real video flowing is exported.
After virtual reality video flowing after output adjustment, step S10 can be returned to.
The implementation method of above-mentioned virtual reality, according to the depth location information of the destination object of acquisition, determine destination object
Predetermined patterns position, and determine according to variable quantity of the position in different time piece the action message of destination object,
Virtual reality parameter is adjusted according to the action message, user need not be worn outside extra body-sensing in the implementation process of virtual reality
If reducing the complexity of virtual reality operation, also cause virtual reality experience closer to real-life real scene;And
And this method the type of action and action extent index that can be made with interpretation user, has in the case where not wearing body-sensing peripheral hardware
Help abundant virtual reality realizes details.
As shown in figure 3, the present invention a kind of virtual reality is also provided realize system, it may include:
Information acquisition module 10, for obtaining the surface each point of destination object in several adjacent timeslices relative to default
The depth location information of reference point;
This functions of modules can be realized that depth camera can be infrared Ancillary textures depth camera, flight by depth camera
Time (Time Of Flight, abbreviation TOF) camera or other depth cameras..Control terminal can be received by depth camera to send out
The data acquisition instructions sent, when can also detect that destination object enters the field range of depth camera by depth camera, make depth
Spend camera and obtain the surface each point of destination object in several adjacent timeslices relative to the depth location information of depth camera.Its
In, control terminal can be the control of intelligent terminal such as mobile phone, tablet personal computer.
In actual applications, destination object can be the user using virtual reality device.For the ease of description, below
Illustrated so that destination object is user as an example.
In one embodiment, user can make TOF camera by controlling terminal to send data acquisition instructions to TOF camera
Start to gather in its visual field in several adjacent timeslices user relative to the depth location information of TOF camera.Pass through peripheral hardware control
The mode of terminal triggering virtual reality processed, can make user more freely control the operating process of virtual reality.
In another embodiment, the depth camera uses infrared Ancillary textures binocular depth camera, works as depth camera
Detect user take virtual reality glasses and enter depth camera within sweep of the eye when, 3D depth cameras are automatically according to the time
The body of piece continuous collecting user relative to depth camera depth location information.By depth camera automatic detection and start to adopt
The mode of collection, improve the convenience of virtual reality operation process.
The surface each point of destination object in several adjacent timeslices is being obtained relative to the depth location of preset reference point
, can also be by determining the geometric center of destination object after information, and the angle of 3D depth cameras is adjusted according to the geometric center,
Make depth camera includes the geometric center within sweep of the eye.
In one embodiment, depth camera gathers in its visual field in several adjacent timeslices user relative to depth phase
The depth location information of machine, and the geometric center of user is determined, according to the movement of this geometric center, adjust acquisition angles, control
Depth camera includes this geometric center within sweep of the eye, so as to improve the free degree of virtual reality operation process so that empty
Intend experience of reality closer to reality.
Information acquisition module 10 is carried out the depth information defeated after the depth location information of the destination object is obtained
Go out, into data processing module 20.
Data processing module 20, for determining the predetermined patterns place of the destination object according to the depth location information
Position;
This functions of modules can realize that the intelligent gateway can be with above-mentioned depth phase by the data processing unit on intelligent gateway
Machine establishes wireless connection, and connected mode may include Ethernet, USB or WIFI etc., so as to realize wireless telecommunications, improves virtual reality
Discussing Convenience.Depth location information can be received by data processing unit on intelligent gateway, according to the depth location information,
The 3D rendering of the predetermined patterns of destination object is generated in 3D depth coordinates system, and according to the 3D rendering, obtains the destination object
Predetermined patterns where position.Wherein the predetermined patterns of destination object can be the positions such as the trunk, arm, lower limb of user.
In one embodiment, the user's body collected by the processing unit on intelligent gateway is relative to depth
The depth location information of camera, in 3D depth coordinates system, according to the depth location information generate the trunk of user, arm and under
The 3D rendering at limb position, and according to the 3D rendering, obtain the position where user's trunk, arm and lower extremity.By in 3D
In depth coordinate system, the 3D rendering of destination object predetermined patterns is rebuild so as to the position where destination object predetermined patterns
The positioning put is more accurate.
Action recognition module 30, for determining the destination object according to the variable quantity of the position in different time piece
Action message;
This functions of modules can realize that action recognition unit is according to different time by the action recognition unit on intelligent gateway
The location variation of piece destination object predetermined patterns, determine the action message of the destination object.Wherein, action message can be target
The type of action and action extent index of object, can be by the location variation of predetermined patterns and the deliberate action type of same area
Variable quantity reference value is contrasted, and when the variable quantity is more than or equal to variable quantity reference value, determines the action of predetermined patterns
Type is deliberate action type, and determines according to the variable quantity of the position action degree of the type of action of the destination object
Parameter, the degree of accuracy of identification destination object action message can be improved by the variable quantity reference value of deliberate action type, and enhancing is empty
That intends reality realizes details.Wherein, type of action may include that boxing, pin such as kick, embraced at the type, and action extent index can be
The distance of work, the time of execution.
Specifically, user carries out reality-virtualizing game in the depth camera visual field, and it is each to gather user surface by depth camera
The depth location information relative to camera is put, and this depth location information is returned to intelligent gateway, by intelligent gateway to the depth
Degree positional information handled, determine user body parts, as trunk, arm, lower limb position in the position of different time piece
Variable quantity, location variation and default running parameter are contrasted, so that it is determined that the action message of user, can be once to box
Action, a kick, once embrace action.
In one embodiment, the action recognition unit on intelligent gateway will be used in different time piece in 3D depth coordinates system
The location variation of family lower extremity and the kicking variable quantity reference value of lower extremity are contrasted, and determine that location variation is more than
Variable quantity reference value, the action message for judging user are a kick, and analysis obtain completing this kick when
Between be 3S, kicking distance be 50cm.
Parameter adjustment module 40, for being carried out according to the action message to the operational factor for realizing system of virtual reality
Adjustment, and the virtual reality video flowing after output adjustment.
This functions of modules can be performed by the parameter adjustment unit on intelligent gateway, and parameter adjustment unit is according to the mesh of acquisition
The action message of mark object is adjusted to the operational factor for realizing system of virtual reality, and by the virtual reality after adjustment
Video flowing is exported.In one embodiment, it has been determined that the action message of user is a kick, and distance is 50cm,
Direction is due east direction, and in this direction, has a vase from user 50cm, so as to parameter adjustment unit kicking according to this user
Leg acts, and adjusts the operational factor for realizing system of virtual reality, i.e., realizes that user's kicking smashes vase in virtual reality
Scene, and the virtual reality video flowing after adjustment is exported.
Said system, without extra body-sensing peripheral hardware, reduces virtual reality operation in the implementation process of virtual reality
Complexity, also cause virtual reality experience closer to real-life real scene;And the system is not being worn outside body-sensing
In the case of if, the type of action and action extent index that can be made with interpretation user, contribute to the realization of abundant virtual reality
Details.
The implementation method for realizing system and the virtual reality of the present invention of the virtual reality of the present invention corresponds, above-mentioned
The technical characteristic and its advantage that the embodiment of the implementation method of virtual reality illustrates are applied to the realization system of virtual reality
In the embodiment of system, hereby give notice that.
As shown in figure 4, the present invention also provides a kind of realization device of virtual reality, it may include:
Depth camera 101 and intelligent gateway 201;
The depth camera 101 obtains the surface each point of destination object in several adjacent timeslices relative to depth camera
Depth location information, and the depth location information is sent to intelligent gateway 201;
Position of the intelligent gateway 201 according to where the depth location information determines the predetermined patterns of the destination object
Put, the action message of the destination object is determined according to the variable quantity of the position in different time piece, believed according to the action
Cease and the operational factor for realizing system of virtual reality is adjusted, and the virtual reality video flowing after output adjustment.
Be the field range of depth camera 101 on the left of Fig. 4, depth camera 101 can within sweep of the eye it is adjacent several when
Between in piece the surface each point of destination object be acquired relative to the depth location information of camera.
Depth camera 101 can be infrared Ancillary textures depth camera, flight time (Time Of Flight, abbreviation
TOF) camera or other depth cameras.The data acquisition instructions that control terminal is sent can be received by depth camera 101, can also be led to
When crossing depth camera 101 and detecting that destination object enters the field range of camera, camera is set to obtain in adjacent several timeslices
The surface each point of destination object relative to camera depth location information.Wherein, it can be mobile phone, tablet personal computer etc. to control terminal
Control of intelligent terminal.
In actual applications, destination object can be the user using virtual reality device.For the ease of description, below
Illustrated so that destination object is user as an example.
In one embodiment, user can make TOF camera by controlling terminal to send data acquisition instructions to TOF camera
Start to gather in its visual field in several adjacent timeslices user relative to the depth location information of TOF camera.Pass through peripheral hardware control
The mode of terminal triggering virtual reality processed, can make user more freely control the operating process of virtual reality.
In another embodiment, the depth camera uses infrared Ancillary textures binocular depth camera, works as depth camera
Detect user take virtual reality glasses and enter depth camera within sweep of the eye when, 3D depth cameras are automatically according to the time
The body of piece continuous collecting user relative to depth camera depth location information.By depth camera automatic detection and start to adopt
The mode of collection, improve the convenience of virtual reality operation process.
The surface each point of destination object in several adjacent timeslices is being obtained relative to the depth location of preset reference point
After information, the geometric center of destination object is may further determine that, and according to the angle of the geometric center percentage regulation camera 101, make
Depth camera 101 includes the geometric center within sweep of the eye.
In one embodiment, depth camera 101 gathers in its visual field in several adjacent timeslices user relative to depth
The depth location information of camera 101 is spent, and determines the geometric center of user, according to the movement of this geometric center, adjusts acquisition angle
Degree, controlling depth camera 101 includes this geometric center within sweep of the eye, so as to improve the freedom of virtual reality operation process
Degree so that the closer reality of virtual reality experience.
Depth camera 101 is returned the depth information after the depth location information of the destination object is obtained
To the intelligent gateway 201 of connection.
It is the intelligent gateway 201 being connected with depth camera 101 on the right side of Fig. 4, the intelligent gateway 201 can be with depth camera 101
Wireless connection is established, connected mode may include Ethernet, USB or WIFI etc., so as to realize wireless telecommunications, improve virtual reality
Discussing Convenience.The depth location information can be received by the data processing unit 202 on intelligent gateway 201, according to the depth position
Confidence is ceased, and the 3D rendering of the predetermined patterns of destination object is generated in 3D depth coordinates system, and according to the 3D rendering, described in acquisition
Position where the predetermined patterns of destination object.Wherein the predetermined patterns of destination object can be trunk, arm, lower limb of user etc.
Position.
In one embodiment, the user's body collected by the data processing unit 202 on intelligent gateway 201
Relative to the depth location information of depth camera, in 3D depth coordinates system, the body of user is generated according to the depth location information
The 3D rendering of dry, arm and lower extremity, and according to the 3D rendering, obtain the position where user's trunk, arm and lower extremity
Put.By in 3D depth coordinates system, rebuilding the 3D rendering of destination object predetermined patterns so that preset portion to destination object
The positioning of position where position is more accurate.
Action recognition unit 203 on intelligent gateway 201 receives the positional information that data processing unit 202 is sent, according to
The variable quantity of the position determines the action message of the destination object in different time piece.Wherein, action message can be target
The type of action and action extent index of object, can be by the location variation of predetermined patterns and the deliberate action type of same area
Variable quantity reference value is contrasted, and when the variable quantity is more than or equal to variable quantity reference value, determines the action of predetermined patterns
Type is deliberate action type, and determines according to the variable quantity of the position action degree of the type of action of the destination object
Parameter, the degree of accuracy of identification destination object action message can be improved by the variable quantity reference value of deliberate action type, and enhancing is empty
That intends reality realizes details.Wherein, type of action may include that boxing, pin such as kick, embraced at the type, and action extent index can be
The distance of work, the time of execution.
In one embodiment, the action recognition unit 203 on intelligent gateway 201 is by 3D depth coordinates in different time piece
The kicking variable quantity reference value of the location variation of user's lower extremity and lower extremity is contrasted in system, determines change in location
Amount is more than variable quantity reference value, and the action message for judging user be a kick, and analysis obtains completing this and kicked to move
The time of work is 3S, and kicking distance is 50cm.
The action for the destination object that game adjustment unit 204 on intelligent gateway 201 obtains according to action recognition unit 203
Information is adjusted to the operational factor for realizing system of virtual reality, and the virtual reality video flowing progress after adjustment is defeated
Go out.In one embodiment, action recognition unit 203 has determined that the action message of user is a kick, and distance is
50cm, direction are due east direction, and in this direction, have a vase from user 50cm, so as to the game adjustment on intelligent gateway 201
Unit 204 adjusts the operational factor for realizing system of virtual reality, i.e., in virtual reality according to the kick of this user
The scene that user's kicking smashes vase is realized, and the virtual reality video flowing after adjustment is exported.
Specifically, the data processing unit 202 on above-mentioned intelligent gateway 201, action recognition unit 203 and parameter adjustment list
Member 204 can be the software service operated on intelligent gateway, so as to reduce the cost of manufacture additional hardware element.
Further, a kind of realization device of virtual reality, may also include:
Mobile phone 301 and virtual reality glasses 401;
Wherein, mobile phone 301 is used to establish with depth camera 101 and connected, and sends data acquisition to depth camera 101 and refer to
Order, improve the controlling of virtual reality operation.In one embodiment, the mobile phone 301 is established with depth camera 101 and connected
Connect, and data acquisition instructions are sent to depth camera 101.Depth camera 101 obtains after the data acquisition instructions are received
In several adjacent timeslices the surface each point of destination object relative to depth camera depth location information.
Wherein, virtual reality glasses 401 are used to receive the virtual reality video flowing after the adjustment of the output of intelligent gateway 201,
And shown, it is easy to check the virtual reality video flowing after adjustment.
The realization device of above-mentioned virtual reality, without extra body-sensing peripheral hardware in the implementation process of virtual reality, reduce
The complexity of virtual reality operation, also cause virtual reality experience closer to real-life real scene;And this dress
Put in the case where not wearing body-sensing peripheral hardware, the type of action and action extent index that can be made with interpretation user, contribute to rich
Rich virtual reality realizes details.
The present invention also provides a kind of computer-readable recording medium, is stored thereon with computer program, the program is processed
Device realizes the implementation method of the virtual reality in above-mentioned any embodiment when performing.Performed by the computer-readable recording medium
Method is identical with the implementation method of the virtual reality in above-described embodiment, and here is omitted.
The more specifically example (non-exhaustive list) of computer-readable medium includes following:Connected up with one or more
Electrical connection section (electronic installation), portable computer diskette box (magnetic device), random access memory (RAM), read-only storage
(ROM), erasable edit read-only storage (EPROM or flash memory), fiber device, and portable optic disk is read-only deposits
Reservoir (CDROM).In addition, computer-readable medium, which can even is that, to print the paper of described program thereon or other are suitable
Medium, because can then enter edlin, interpretation or if necessary with it for example by carrying out optical scanner to paper or other media
His suitable method is handled electronically to obtain described program, is then stored in computer storage.
The present invention also provides a kind of computer equipment, including memory, processor and storage on a memory and can located
The computer program that runs on reason device, the virtual reality in above-mentioned any embodiment is realized during the computing device described program
Implementation method.The realization side of the virtual reality in the method and above-described embodiment performed by processor in the computer equipment
Method is identical, and here is omitted.
Embodiment described above only expresses the several embodiments of the present invention, and its description is more specific and detailed, but simultaneously
Therefore the limitation to the scope of the claims of the present invention can not be interpreted as.It should be pointed out that for one of ordinary skill in the art
For, without departing from the inventive concept of the premise, various modifications and improvements can be made, these belong to the guarantor of the present invention
Protect scope.Therefore, the protection domain of patent of the present invention should be determined by the appended claims.
Claims (10)
1. a kind of implementation method of virtual reality, it is characterised in that comprise the following steps:
The surface each point of destination object in several adjacent timeslices is obtained relative to the depth location information of preset reference point;
Position according to where the depth location information determines the predetermined patterns of the destination object;
The action message of the destination object is determined according to the variable quantity of the position in different time piece;
The operational factor for realizing system of virtual reality is adjusted according to the action message, and it is virtual after output adjustment
Real video flowing.
2. the implementation method of virtual reality according to claim 1, it is characterised in that obtain in several adjacent timeslices
The surface each point of destination object relative to preset reference point depth location information the step of performed by depth camera;
The surface each point of destination object in several adjacent timeslices is obtained relative to the depth location information of preset reference point
Step, including:
The data acquisition instructions of control terminal transmission are received by depth camera, according to the data acquisition instructions, are obtained adjacent
In several timeslices the surface each point of destination object relative to depth camera depth location information;
Or;
If detecting, destination object enters the field range of depth camera, obtains the target pair in several adjacent timeslices
The surface each point of elephant relative to depth camera depth location information.
3. the implementation method of virtual reality according to claim 2, it is characterised in that obtaining several adjacent timeslices
After the surface each point of interior destination object is relative to the depth location information of preset reference point, including:
Determine the geometric center of destination object;
According to the geometric center, the angle of the depth camera is adjusted, make the depth camera includes institute within sweep of the eye
State the geometric center of destination object.
4. the implementation method of virtual reality according to claim 1, it is characterised in that true according to the depth location information
The step of position where the predetermined patterns of the fixed destination object, including:
According to the depth location information, the 3D rendering of the predetermined patterns of the destination object is generated in 3D depth coordinates system;
According to the 3D rendering, the position where the predetermined patterns of the destination object is obtained.
5. the implementation method of the virtual reality according to Claims 1-4 any one, it is characterised in that the action letter
Breath includes the type of action and action extent index of destination object.
6. the implementation method of virtual reality according to claim 5, it is characterised in that according to institute's rheme in different time piece
The variable quantity put determines the step of action message of the destination object, including:
The variable quantity and the variable quantity reference value of corresponding site deliberate action type are contrasted;
If the variable quantity is more than or equal to the variable quantity reference value, it is determined that the type of action of the destination object is default
Type of action;
According to the variable quantity of the position in different time piece, the action extent index of the destination object is determined.
7. a kind of virtual reality realizes system, it is characterised in that including:
Information acquisition module, for obtaining the surface each point of destination object in several adjacent timeslices relative to preset reference point
Depth location information;
Data processing module, for the position where determining the predetermined patterns of the destination object according to the depth location information
Put;
Action recognition module, for determining that the action of the destination object is believed according to the variable quantity of the position in different time piece
Breath;
Parameter adjustment module, for being adjusted according to the action message to the operational factor for realizing system of virtual reality,
And the virtual reality video flowing after output adjustment.
A kind of 8. realization device of virtual reality, it is characterised in that including:
Depth camera and intelligent gateway;
The depth camera obtains the surface each point of destination object in several adjacent timeslices relative to the depth of depth camera
Positional information, and the depth location information is sent to intelligent gateway;
Position of the intelligent gateway according to where the depth location information determines the predetermined patterns of the destination object, according to
The variable quantity of the position determines the action message of the destination object in different time piece, according to the action message to virtual
Reality realizes that the operational factor of system is adjusted, and the virtual reality video flowing after output adjustment.
9. the realization device of virtual reality according to claim 8, it is characterised in that also include:
Virtual reality glasses;
The virtual reality video flowing that the virtual reality glasses are used for after the adjustment of Rreceive output, and shown.
10. the realization device of virtual reality according to claim 8, it is characterised in that also include:
Control terminal;
The control terminal is used to send data acquisition instructions to the depth camera;
For depth camera after the data acquisition instructions are received, the surface for obtaining destination object in several adjacent timeslices is each
Put the depth location information relative to depth camera.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710964798.5A CN107783652B (en) | 2017-10-17 | 2017-10-17 | Method, system and device for realizing virtual reality |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710964798.5A CN107783652B (en) | 2017-10-17 | 2017-10-17 | Method, system and device for realizing virtual reality |
Publications (2)
Publication Number | Publication Date |
---|---|
CN107783652A true CN107783652A (en) | 2018-03-09 |
CN107783652B CN107783652B (en) | 2020-11-13 |
Family
ID=61434535
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710964798.5A Active CN107783652B (en) | 2017-10-17 | 2017-10-17 | Method, system and device for realizing virtual reality |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107783652B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114625251A (en) * | 2022-03-11 | 2022-06-14 | 平安普惠企业管理有限公司 | Interaction method and device based on VR, computer equipment and storage medium |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103529944A (en) * | 2013-10-17 | 2014-01-22 | 合肥金诺数码科技股份有限公司 | Human body movement identification method based on Kinect |
US20150040174A1 (en) * | 2013-08-01 | 2015-02-05 | Joiz Ip Ag | System and method for synchronizing media platform devices |
CN104460962A (en) * | 2013-09-18 | 2015-03-25 | 天津联合动力信息技术有限公司 | 4D somatosensory interaction system based on game engine |
CN106295479A (en) * | 2015-06-05 | 2017-01-04 | 上海戏剧学院 | Based on body-sensing technology action recognition editing system |
CN106601062A (en) * | 2016-11-22 | 2017-04-26 | 山东科技大学 | Interactive method for simulating mine disaster escape training |
CN107133984A (en) * | 2017-03-24 | 2017-09-05 | 深圳奥比中光科技有限公司 | The scaling method and system of depth camera and main equipment |
-
2017
- 2017-10-17 CN CN201710964798.5A patent/CN107783652B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150040174A1 (en) * | 2013-08-01 | 2015-02-05 | Joiz Ip Ag | System and method for synchronizing media platform devices |
CN104460962A (en) * | 2013-09-18 | 2015-03-25 | 天津联合动力信息技术有限公司 | 4D somatosensory interaction system based on game engine |
CN103529944A (en) * | 2013-10-17 | 2014-01-22 | 合肥金诺数码科技股份有限公司 | Human body movement identification method based on Kinect |
CN106295479A (en) * | 2015-06-05 | 2017-01-04 | 上海戏剧学院 | Based on body-sensing technology action recognition editing system |
CN106601062A (en) * | 2016-11-22 | 2017-04-26 | 山东科技大学 | Interactive method for simulating mine disaster escape training |
CN107133984A (en) * | 2017-03-24 | 2017-09-05 | 深圳奥比中光科技有限公司 | The scaling method and system of depth camera and main equipment |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114625251A (en) * | 2022-03-11 | 2022-06-14 | 平安普惠企业管理有限公司 | Interaction method and device based on VR, computer equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN107783652B (en) | 2020-11-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107990899A (en) | A kind of localization method and system based on SLAM | |
EP2917754B1 (en) | Image processing method, particularly used in a vision-based localization of a device | |
US10802606B2 (en) | Method and device for aligning coordinate of controller or headset with coordinate of binocular system | |
CN112711249B (en) | Robot positioning method and device, intelligent robot and storage medium | |
CN105094335A (en) | Scene extracting method, object positioning method and scene extracting system | |
CN107665508B (en) | Method and system for realizing augmented reality | |
CN108885487B (en) | Gesture control method of wearable system and wearable system | |
CN106575437A (en) | Information-processing device, information processing method, and program | |
CN107665505A (en) | The method and device of augmented reality is realized based on plane monitoring-network | |
CN110276251B (en) | Image recognition method, device, equipment and storage medium | |
US11127156B2 (en) | Method of device tracking, terminal device, and storage medium | |
CN110456904B (en) | Augmented reality glasses eye movement interaction method and system without calibration | |
CN106504227B (en) | Demographic method and its system based on depth image | |
CN107025661A (en) | A kind of method for realizing augmented reality, server, terminal and system | |
CN109902675B (en) | Object pose acquisition method and scene reconstruction method and device | |
CN109740659A (en) | A kind of image matching method and device, electronic equipment, storage medium | |
CN106652028A (en) | Environment three-dimensional mapping method and apparatus | |
US11010916B2 (en) | Method of configuring camera position suitable for localization and robot implementing same | |
CN110096152A (en) | Space-location method, device, equipment and the storage medium of physical feeling | |
CN111354029A (en) | Gesture depth determination method, device, equipment and storage medium | |
CN109191522B (en) | Robot displacement correction method and system based on three-dimensional modeling | |
CN107783652A (en) | Implementation method, system and the device of virtual reality | |
CN116958452A (en) | Three-dimensional reconstruction method and system | |
KR20160099981A (en) | Virtual reality device based on two-way recognition including a three-dimensional marker with a patten for multiple users | |
CN107831921A (en) | A kind of handle space position and the determination method, apparatus and system of coding corresponding relation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |