CN102253713B - Towards 3 D stereoscopic image display system - Google Patents

Towards 3 D stereoscopic image display system Download PDF

Info

Publication number
CN102253713B
CN102253713B CN201110171017.XA CN201110171017A CN102253713B CN 102253713 B CN102253713 B CN 102253713B CN 201110171017 A CN201110171017 A CN 201110171017A CN 102253713 B CN102253713 B CN 102253713B
Authority
CN
China
Prior art keywords
unit
user
dimensional
stereoscopic image
image display
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201110171017.XA
Other languages
Chinese (zh)
Other versions
CN102253713A (en
Inventor
陈大炜
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Konka Group Co Ltd
Original Assignee
Konka Group Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Konka Group Co Ltd filed Critical Konka Group Co Ltd
Priority to CN201110171017.XA priority Critical patent/CN102253713B/en
Publication of CN102253713A publication Critical patent/CN102253713A/en
Application granted granted Critical
Publication of CN102253713B publication Critical patent/CN102253713B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Abstract

The invention discloses a kind of towards 3 D stereoscopic image display system, will solve the technical problem that it is to improve man-machine interaction effect, improve the effect of 3 D stereoscopic image.The present invention is by the following technical solutions: a kind of towards 3 D stereoscopic image display system, described includes with lower component towards 3 D stereoscopic image display system: stereoscopic image display unit, depth image sensing unit, depth image processing unit, limb action recognition unit, three-dimensional user interface unit and three-dimensional position indicator unit.Compared with prior art, use the depth image sensing unit towards 3 D stereoscopic image display system and depth image processing unit, all or part of limb action of user under sensing solid space, utilize three-dimensional user interface to interact with system simultaneously, and then arrive the man-machine interaction effect immersed completely, thus improve Consumer's Experience greatly.

Description

Towards 3 D stereoscopic image display system
Technical field
The present invention relates to a kind of field of stereoscopic display, particularly a kind of towards 3 D stereoscopic image display system.
Background technology
Television set enters as a kind of popular leisure household appliances and modal display device are universal already Huge numbers of families, along with the development television set of Display Technique has also gone through black-and-white television, color television, liquid crystal electricity Depending on multiple stages such as, 3 D stereoscopic image TVs.Particularly, nearly 2 years along with various brands 3D TV face successively City sells, and 3 D stereoscopic image TV has moved towards the life of consumer, 3 D stereoscopic image electricity finally from concept Depending on utilizing the eyes of people to observe the angle slightly difference of object, the principle of stereoscopic vision can be produced, by respectively The image that the mode of kind is seen right and left eyes separates, thus makes user pass through by anaglyph spectacles or bore hole body Test stereoscopic vision, and man machine interface is medium and the dialogue transmitting, exchanging information between people and computer Interface, is the important component part of computer system.User can pass through various human-computer interaction devices Interact with computer with mode, such as classical keyboard and mouse pattern, remote controller pattern, touch control, Gesture controls, the most up-to-date action induction human-computer interaction etc..
But current all of user interface is all the form of two dimension, and the most all of operation is all towards display device Screen.As a example by mouse, the movement of mouse, click on, pull etc. all operations and all display can only be set Destination object in standby plane works, and for two-dimensional display device, this is natural thing, Also can guarantee that good man-machine interaction experience, for based on action induction man-machine interaction, user is non-connecing The mode of touch operates user interface, it is envisaged that for aloft by the most invisible touch screen of touch-control Curtain interacts with system, and user operation is fed back still on two-dimensional display device screen, in this case Consumer's Experience will decrease, for 3-dimensional image stereoscopic display device, traditional two Dimension user interface cannot meet actual demand, and how improving this Consumer's Experience is urgently to be resolved hurrily asking Topic.
Summary of the invention
It is an object of the invention to provide a kind of towards 3 D stereoscopic image display system, solve the technical problem that It is to improve man-machine interaction effect, improves the effect of 3 D stereoscopic image.
The present invention is by the following technical solutions: a kind of towards 3 D stereoscopic image display system, described stands towards three-dimensional Body image display system includes with lower component:
Stereoscopic image display unit: for receiving the 3 D stereoscopic image and three of three-dimensional user interface unit transmission Dimension graphical interface of user;
Depth image sensing unit: be used for obtaining and sensing 3 D stereoscopic image display unit place environment, bag Include user in interior deep image information;
Depth image processing unit: the visual range internal field depth of field obtained for analysis depth image sensing unit Degree information, and identify the user object and whole body obtained by limb action recognition unit or part of limb motion;
Limb action recognition unit: be used for following the tracks of and identifying user action, and carry for three-dimensional user interface unit Control interface is inputted for user;
Three-dimensional user interface unit: for sending graphical interfaces to stereoscopic image display unit, with 3 D stereo The form of image is by display unit, and after receiving the user action of limb action recognition unit, moves user Make to feed back on stereoscopic image display unit;
Three-dimensional position indicator unit: be used for receiving scene depth letter in depth image processing unit visual range Identify active user's particular location at three-dimensional user interface after breath, and user's control action is made feedback;
Described depth image sensing unit connects depth image processing unit, and depth image processing unit connects limb
Body action recognition unit, described limb action recognition unit connects three-dimensional user interface unit, three Wesys Family boundary element connects stereoscopic image display unit and three-dimensional position indicator unit respectively.
The present invention compared with prior art, uses the depth image towards 3 D stereoscopic image display system to sense Unit and depth image processing unit, all or part of limb action of user under sensing solid space, simultaneously Utilize three-dimensional user interface to interact with system, and this three-dimensional user interface shows with towards 3 D stereoscopic image Show that the display depth space of system is consistent, and can be carried out with user operation by three-dimensional position indicator Feedback, effective opereating specification of user is also protected with the display depth space towards 3 D stereoscopic image display system Hold consistent, and then arrive the man-machine interaction effect immersed completely, thus improve Consumer's Experience greatly.
Accompanying drawing explanation
Fig. 1 is the structured flowchart of the present invention.
Fig. 2 is the flow chart of the present invention.
Detailed description of the invention
With embodiment, the present invention is described in further detail below in conjunction with the accompanying drawings.
As it is shown in figure 1, the present invention's includes with lower component towards 3 D stereoscopic image display system:
Stereoscopic image display unit: for receiving the 3 D stereoscopic image and three of three-dimensional user interface unit transmission Dimension graphical interface of user;
Depth image sensing unit: be used for obtaining and sensing 3 D stereoscopic image display unit place environment, bag Include user in interior deep image information;
Depth image processing unit: the visual range internal field depth of field obtained for analysis depth image sensing unit Degree information, and identify the user object and whole body obtained by limb action recognition unit or part of limb motion; Limb action recognition unit: be used for following the tracks of and identifying user action, and provide use for three-dimensional user interface unit Family input controls interface;
Three-dimensional user interface unit: for sending graphical interfaces to stereoscopic image display unit, with 3 D stereo The form of image is by display unit, and after receiving the user action of limb action recognition unit, moves user Make to feed back on stereoscopic image display unit;
Three-dimensional position indicator unit: be used for receiving scene depth letter in depth image processing unit visual range Identify active user's particular location at three-dimensional user interface after breath, and user's control action is made feedback.
Described depth image sensing unit connects depth image processing unit, and depth image processing unit connects limb Body action recognition unit, described limb action recognition unit connects three-dimensional user interface unit, three dimensional user circle Face unit connects stereoscopic image display unit and three-dimensional position indicator unit respectively.
Three-dimensional user interface unit of the present invention with the Windows operating system of Microsoft as reference, three dimensional user circle Face unit is patterned, can include the interface elements such as window, icon, file, desktop, towards Under 3 D stereoscopic image display system environment all of interface element all have depth attribute, i.e. interface object from Open the distance of physical display screen it can be understood as a three-dimensional desktop in three dimensions.User can be led to Crossing gesture motion to rotate, scale this three-dimensional desktop, all of window is the most vertically suspended in such a solid In space, and any two is opened the depth difference between window and is maintained at more than the threshold range of setting, user Can be captured by gesture motion, move, pull window interested and as focus window etc., this stands The spatial dimension of body desktop keeps consistent with user's gesture motion effective range, such as extended when user's one hand During relative user self the far-end upper right corner of effective range, three-dimensional position indicator also should move to this The three-dimensional desktop far-end upper right corner, similar, when user's one hand is reduced to the relative user self of effective range During the near-end lower left corner, three-dimensional position indicator also should move to this solid desktop near-end lower left corner.
The three-dimensional position indicator unit of the present invention can be a 3 D stereo in three dimensional user graphical interfaces Little hand state, and corresponding metamorphosis, such as user can be made according to the gesture motion that user is different and doing During propelling movement action, little hands is the five fingers expanded configuration, and user's little hands when doing drag motions is seized condition etc. of clenching fist, User single-handed exercise be display one pinkie show device, when user does double-handed exercise again, be shown as two little Hands indicator etc..Three-dimensional position indicator physical location in three dimensional user graphical interfaces is by user's hand phase Displacement is determined, and direction and the speed of the moving direction of indicator and speed and user's hand motion keep one Cause.
The present invention under the effect of depth image sensing unit, this system can obtain the limb action of user Information, as a example by gesture motion, gesture motion effective range refers to one in physical display device screen and use A specific region between the physical location of family, only after user's hand enters this region, system just can be right User's gesture motion is identified, outside user's hand is positioned at this effective coverage, then and the gesture motion of user Can be ignored by system, the locus scope of the gesture motion effective range application three-dimensional user interface of user is protected Holding consistent, such user operation before 3 D stereo affects display device just can reach the people immersed completely Machine interaction effect.
As in figure 2 it is shown, the present invention's uses following steps to realize towards 3 D stereoscopic image display system: one, Stereoscopic image display unit projects virtual three dimensional user graphical interfaces in entity space;Two, depth image Sensing unit, after stereoscopic image display unit projects virtual three dimensional user graphical interfaces, obtains user institute In environment deep image information, and deep image information is sent to depth image processing unit;Three, the degree of depth Graphics processing unit starts after receiving deep image information identify user object and follow the tracks of user's hand motion, User's hand motion information is fed back to limb action recognition unit;Four, limb action recognition unit judges to use Whether action gesture in family, in useful effect region, is, enters next step operation, otherwise returns previous step; Five, limb action recognition unit identification user's gesture motion judge this action command;Six, limb action is known Order after other unit will identify sends to three-dimensional user interface unit;Seven, three-dimensional user interface unit will letter Breath sends to three-dimensional position indicator unit, updates three-dimensional position indicator post and state;Eight, three Wesy User action control command is made respective feedback and shows on three-dimensional user interface unit by family boundary element.
The limb action recognition unit of the present invention can identify the bimanualness of user, and its step is as follows: one, User's one hand enters gesture motion effective coverage;Two, three-dimensional position indicator unit updates three-dimensional little hand position And form;Three, user moves in effective coverage, and pushes away when three dimensional indicia is positioned at certain window ranges Send action;Four, limb action recognition unit identifies effective gesture motion for clicking on order, three dimensional user circle This window is set to current focus window by face unit, and is placed before current all windows;Five, Another hands is stretched into useful effect region by user;Six, three-dimensional position indicator unit is updated to two vertical Body little hand state, position keeps consistent with two-hand positions;Seven, user's both hands are in current focus window ranges It is relatively distant from action;Eight, limb action recognition unit identifies this effective gesture motion for amplifying window life Order, this window is amplified to both hands by three-dimensional user interface, and to be relatively distant from distance proportion consistent.
The model that the stereoscopic image display unit of the present invention uses KangJia Group Co., Ltd to produce is LC42MS96PD dimensional image display;Depth image sensing unit uses PrimeSense company to produce Imageing sensor;The model that depth image processing unit uses KangJia Group Co., Ltd to produce is The graphics processing unit of KK-Depther;Limb action recognition unit uses KangJia Group Co., Ltd raw The limb action evaluator that model is KK-Montion produced;Three-dimensional user interface unit uses group of Konka stock The three-dimensional user interface unit that model is KK-3DUI that part company limited produces;Three-dimensional position indicator unit Use the three-dimensional position indicator that model is KK-3DCur that KangJia Group Co., Ltd produces;Action Effective coverage is between 0.8 meter to 3.5 meters.

Claims (1)

1. one kind towards 3 D stereoscopic image display system, it is characterised in that: described towards 3 D stereoscopic image Display system includes with lower component:
Stereoscopic image display unit: for receiving the 3 D stereoscopic image and three of three-dimensional user interface unit transmission Dimension graphical interface of user;
Depth image sensing unit: be used for obtaining and sensing 3 D stereoscopic image display unit place environment, bag Include user in interior deep image information;
Depth image processing unit: the visual range internal field depth of field obtained for analysis depth image sensing unit Degree information, and identify the user object and whole body obtained by limb action recognition unit or part of limb motion;
Limb action recognition unit: be used for following the tracks of and identifying user action, and carry for three-dimensional user interface unit Control interface is inputted for user;
Three-dimensional user interface unit: for sending graphical interfaces to stereoscopic image display unit, with 3 D stereo The form of image is by display unit, and after receiving the user action of limb action recognition unit, moves user Make to feed back on stereoscopic image display unit;
Three-dimensional position indicator unit: be used for receiving scene depth letter in depth image processing unit visual range Identify active user's particular location at three-dimensional user interface after breath, and user's control action is made feedback;
Described depth image sensing unit connects depth image processing unit, and depth image processing unit connects limb Body action recognition unit, described limb action recognition unit connects three-dimensional user interface unit, three dimensional user circle Face unit connects stereoscopic image display unit and three-dimensional position indicator unit respectively;
Described towards the employing following steps realization of 3 D stereoscopic image display system: one, stereoscopic image display unit Virtual three dimensional user graphical interfaces is projected in entity space;Two, depth image sensing unit is at three-dimensional shadow After projecting virtual three dimensional user graphical interfaces as display unit, obtain user place environment depth image letter Breath, and deep image information is sent to depth image processing unit;Three, depth image processing unit receives Start identify user object and follow the tracks of user's hand motion after deep image information, user's hand motion is believed Breath feeds back to limb action recognition unit;Four, limb action recognition unit judges that whether user action gesture exists In useful effect region, it is to enter next step operation, otherwise return previous step;Five, limb action identification Unit identification user's gesture motion also judges this action command;Six, after limb action recognition unit will identify Order sends to three-dimensional user interface unit;Seven, three-dimensional user interface unit sends information to three-dimensional position Indicator unit, updates three-dimensional position indicator post and state;Eight, three-dimensional user interface unit is to user Action directive is made respective feedback and shows on three-dimensional user interface unit;
Described limb action recognition unit can identify the bimanualness of user, and its step is as follows: one, user Singlehanded entrance gesture motion effective coverage;Two, three-dimensional position indicator unit updates three-dimensional little hand position and shape State;Three, user moves in effective coverage, and do propelling movement when three dimensional indicia is positioned at certain window ranges dynamic Make;Four, limb action recognition unit identifies effective gesture motion for clicking on order, three-dimensional user interface list This window is set to current focus window by unit, and is placed before current all windows;Five, user Another hands is stretched into useful effect region;Six, to be updated to the solid of two little for three-dimensional position indicator unit Hand state, position keeps consistent with two-hand positions;Seven, user's both hands do phase in current focus window ranges To away from action;Eight, limb action recognition unit identifies this effective gesture motion is to amplify window command, This window is amplified to both hands by three-dimensional user interface, and to be relatively distant from distance proportion consistent;
The model that described stereoscopic image display unit uses KangJia Group Co., Ltd to produce is LC42MS96PD dimensional image display;Depth image sensing unit uses PrimeSense company to produce Imageing sensor;The model that depth image processing unit uses KangJia Group Co., Ltd to produce is The graphics processing unit of KK-Depther;Limb action recognition unit uses KangJia Group Co., Ltd raw The limb action evaluator that model is KK-Montion produced;Three-dimensional user interface unit uses group of Konka The three-dimensional user interface unit that model is KK-3DUI that limited company produces;Three-dimensional position indicator list Unit uses the three-dimensional position indicator that model is KK-3DCur that KangJia Group Co., Ltd produces;Dynamic It is between 0.8 meter to 3.5 meters as effective coverage.
CN201110171017.XA 2011-06-23 2011-06-23 Towards 3 D stereoscopic image display system Active CN102253713B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201110171017.XA CN102253713B (en) 2011-06-23 2011-06-23 Towards 3 D stereoscopic image display system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201110171017.XA CN102253713B (en) 2011-06-23 2011-06-23 Towards 3 D stereoscopic image display system

Publications (2)

Publication Number Publication Date
CN102253713A CN102253713A (en) 2011-11-23
CN102253713B true CN102253713B (en) 2016-10-12

Family

ID=44981016

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201110171017.XA Active CN102253713B (en) 2011-06-23 2011-06-23 Towards 3 D stereoscopic image display system

Country Status (1)

Country Link
CN (1) CN102253713B (en)

Families Citing this family (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106943743B (en) * 2011-02-11 2024-04-26 漳州市阿思星谷电子科技有限公司 Man-machine interaction control method
CN102789312B (en) * 2011-12-23 2016-03-23 苏州触达信息技术有限公司 A kind of user interactive system and method
CN102707878A (en) * 2012-04-06 2012-10-03 深圳创维数字技术股份有限公司 User interface operation control method and device
CN102902355B (en) * 2012-08-31 2015-12-02 中国科学院自动化研究所 The space interaction method of mobile device
TWI494792B (en) 2012-09-07 2015-08-01 Pixart Imaging Inc Gesture recognition system and method
CN103839040B (en) * 2012-11-27 2017-08-25 株式会社理光 Gesture identification method and device based on depth image
US9667883B2 (en) * 2013-01-07 2017-05-30 Eminent Electronic Technology Corp. Ltd. Three-dimensional image sensing device and method of sensing three-dimensional images
CN103067727A (en) * 2013-01-17 2013-04-24 乾行讯科(北京)科技有限公司 Three-dimensional 3D glasses and three-dimensional 3D display system
CN103974008A (en) * 2013-01-30 2014-08-06 联想(北京)有限公司 Information processing method and electronic equipment
CN103543830B (en) * 2013-10-28 2017-02-15 四川大学 Method for mapping human skeleton points to virtual three-dimensional space points in three-dimensional display
CN103530060B (en) * 2013-10-31 2016-06-22 京东方科技集团股份有限公司 Display device and control method, gesture identification method
CN103995620A (en) * 2013-12-02 2014-08-20 深圳市云立方信息科技有限公司 Air touch system
CN104915979A (en) * 2014-03-10 2015-09-16 苏州天魂网络科技有限公司 System capable of realizing immersive virtual reality across mobile platforms
EP2993901A1 (en) * 2014-09-03 2016-03-09 Samsung Electronics Co., Ltd. Method for displaying images and electronic device thereof
CN105511599B (en) * 2014-09-29 2019-06-25 联想(北京)有限公司 Information processing method and device
CN104536575A (en) * 2015-01-04 2015-04-22 苏州易乐展示***工程有限公司 Large screen interaction system realization method based on 3D sensing
CN106681497A (en) * 2016-12-07 2017-05-17 南京仁光电子科技有限公司 Method and device based on somatosensory control application program
CN106846564A (en) * 2016-12-29 2017-06-13 湖南拓视觉信息技术有限公司 A kind of intelligent access control system and control method
CN108388351B (en) * 2018-04-12 2024-03-12 深圳市正图科技有限公司 Mixed reality experience system
CN111045558A (en) * 2018-10-12 2020-04-21 上海博泰悦臻电子设备制造有限公司 Interface control method based on three-dimensional scene, vehicle-mounted equipment and vehicle
CN112925430A (en) * 2019-12-05 2021-06-08 北京芯海视界三维科技有限公司 Method for realizing suspension touch control, 3D display equipment and 3D terminal
CN111722769B (en) * 2020-07-16 2024-03-05 腾讯科技(深圳)有限公司 Interaction method, interaction device, display equipment and storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101952818A (en) * 2007-09-14 2011-01-19 智慧投资控股67有限责任公司 Processing based on the user interactions of attitude
CN101986255A (en) * 2010-11-05 2011-03-16 福州瑞芯微电子有限公司 Semitransparent gradual three-dimensional user interface

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7178111B2 (en) * 2004-08-03 2007-02-13 Microsoft Corporation Multi-planar three-dimensional user interface

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101952818A (en) * 2007-09-14 2011-01-19 智慧投资控股67有限责任公司 Processing based on the user interactions of attitude
CN101986255A (en) * 2010-11-05 2011-03-16 福州瑞芯微电子有限公司 Semitransparent gradual three-dimensional user interface

Also Published As

Publication number Publication date
CN102253713A (en) 2011-11-23

Similar Documents

Publication Publication Date Title
CN102253713B (en) Towards 3 D stereoscopic image display system
US11392212B2 (en) Systems and methods of creating a realistic displacement of a virtual object in virtual reality/augmented reality environments
US20210181856A1 (en) Augmented reality interaction techniques
US11954808B2 (en) Rerendering a position of a hand to decrease a size of a hand to create a realistic virtual/augmented reality environment
JP7092445B2 (en) Methods and systems that provide remote robotic control
Park et al. Real-time 3D pointing gesture recognition for mobile robots with cascade HMM and particle filter
Suarez et al. Hand gesture recognition with depth images: A review
JP5846662B2 (en) Method and system for responding to user selection gestures for objects displayed in three dimensions
US20170235376A1 (en) Systems and methods of direct pointing detection for interaction with a digital device
Stuerzlinger et al. The value of constraints for 3D user interfaces
JP5807686B2 (en) Image processing apparatus, image processing method, and program
JP2022504599A (en) Methods and systems that provide remote robotic control
JP2011022984A (en) Stereoscopic video interactive system
CN103365411A (en) Information input apparatus, information input method, and computer program
CN103246351A (en) User interaction system and method
CN103064514A (en) Method for achieving space menu in immersive virtual reality system
Kulik Building on realism and magic for designing 3D interaction techniques
CN104765454A (en) Human muscle movement perception based menu selection method for human-computer interaction interface
CN105929946B (en) A kind of natural interactive method based on virtual interface
WO2016102948A1 (en) Coherent touchless interaction with stereoscopic 3d images
Gope et al. Interaction with Large Screen Display using Fingertip & Virtual Touch Screen
Park et al. 3D Gesture-based view manipulator for large scale entity model review
Xie et al. Natural Bare-Hand Interaction for Remote Operating Large Touch Screen.
Kim et al. User customizable interaction in coexistence space
Yamada et al. Rebo: A life-like universal remote control

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant