CN102253713B - Towards 3 D stereoscopic image display system - Google Patents
Towards 3 D stereoscopic image display system Download PDFInfo
- Publication number
- CN102253713B CN102253713B CN201110171017.XA CN201110171017A CN102253713B CN 102253713 B CN102253713 B CN 102253713B CN 201110171017 A CN201110171017 A CN 201110171017A CN 102253713 B CN102253713 B CN 102253713B
- Authority
- CN
- China
- Prior art keywords
- unit
- user
- dimensional
- stereoscopic image
- image display
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 230000000694 effects Effects 0.000 claims abstract description 12
- 239000007787 solid Substances 0.000 claims abstract description 5
- 230000033001 locomotion Effects 0.000 claims description 29
- 239000010045 kangjia Substances 0.000 claims description 8
- 230000000007 visual effect Effects 0.000 claims description 6
- 230000005540 biological transmission Effects 0.000 claims description 3
- 230000003993 interaction Effects 0.000 abstract description 9
- 208000032364 Undersensing Diseases 0.000 abstract description 2
- 230000005611 electricity Effects 0.000 description 2
- 230000006698 induction Effects 0.000 description 2
- 230000018109 developmental process Effects 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000029052 metamorphosis Effects 0.000 description 1
- 238000000034 method Methods 0.000 description 1
Abstract
The invention discloses a kind of towards 3 D stereoscopic image display system, will solve the technical problem that it is to improve man-machine interaction effect, improve the effect of 3 D stereoscopic image.The present invention is by the following technical solutions: a kind of towards 3 D stereoscopic image display system, described includes with lower component towards 3 D stereoscopic image display system: stereoscopic image display unit, depth image sensing unit, depth image processing unit, limb action recognition unit, three-dimensional user interface unit and three-dimensional position indicator unit.Compared with prior art, use the depth image sensing unit towards 3 D stereoscopic image display system and depth image processing unit, all or part of limb action of user under sensing solid space, utilize three-dimensional user interface to interact with system simultaneously, and then arrive the man-machine interaction effect immersed completely, thus improve Consumer's Experience greatly.
Description
Technical field
The present invention relates to a kind of field of stereoscopic display, particularly a kind of towards 3 D stereoscopic image display system.
Background technology
Television set enters as a kind of popular leisure household appliances and modal display device are universal already
Huge numbers of families, along with the development television set of Display Technique has also gone through black-and-white television, color television, liquid crystal electricity
Depending on multiple stages such as, 3 D stereoscopic image TVs.Particularly, nearly 2 years along with various brands 3D TV face successively
City sells, and 3 D stereoscopic image TV has moved towards the life of consumer, 3 D stereoscopic image electricity finally from concept
Depending on utilizing the eyes of people to observe the angle slightly difference of object, the principle of stereoscopic vision can be produced, by respectively
The image that the mode of kind is seen right and left eyes separates, thus makes user pass through by anaglyph spectacles or bore hole body
Test stereoscopic vision, and man machine interface is medium and the dialogue transmitting, exchanging information between people and computer
Interface, is the important component part of computer system.User can pass through various human-computer interaction devices
Interact with computer with mode, such as classical keyboard and mouse pattern, remote controller pattern, touch control,
Gesture controls, the most up-to-date action induction human-computer interaction etc..
But current all of user interface is all the form of two dimension, and the most all of operation is all towards display device
Screen.As a example by mouse, the movement of mouse, click on, pull etc. all operations and all display can only be set
Destination object in standby plane works, and for two-dimensional display device, this is natural thing,
Also can guarantee that good man-machine interaction experience, for based on action induction man-machine interaction, user is non-connecing
The mode of touch operates user interface, it is envisaged that for aloft by the most invisible touch screen of touch-control
Curtain interacts with system, and user operation is fed back still on two-dimensional display device screen, in this case
Consumer's Experience will decrease, for 3-dimensional image stereoscopic display device, traditional two
Dimension user interface cannot meet actual demand, and how improving this Consumer's Experience is urgently to be resolved hurrily asking
Topic.
Summary of the invention
It is an object of the invention to provide a kind of towards 3 D stereoscopic image display system, solve the technical problem that
It is to improve man-machine interaction effect, improves the effect of 3 D stereoscopic image.
The present invention is by the following technical solutions: a kind of towards 3 D stereoscopic image display system, described stands towards three-dimensional
Body image display system includes with lower component:
Stereoscopic image display unit: for receiving the 3 D stereoscopic image and three of three-dimensional user interface unit transmission
Dimension graphical interface of user;
Depth image sensing unit: be used for obtaining and sensing 3 D stereoscopic image display unit place environment, bag
Include user in interior deep image information;
Depth image processing unit: the visual range internal field depth of field obtained for analysis depth image sensing unit
Degree information, and identify the user object and whole body obtained by limb action recognition unit or part of limb motion;
Limb action recognition unit: be used for following the tracks of and identifying user action, and carry for three-dimensional user interface unit
Control interface is inputted for user;
Three-dimensional user interface unit: for sending graphical interfaces to stereoscopic image display unit, with 3 D stereo
The form of image is by display unit, and after receiving the user action of limb action recognition unit, moves user
Make to feed back on stereoscopic image display unit;
Three-dimensional position indicator unit: be used for receiving scene depth letter in depth image processing unit visual range
Identify active user's particular location at three-dimensional user interface after breath, and user's control action is made feedback;
Described depth image sensing unit connects depth image processing unit, and depth image processing unit connects limb
Body action recognition unit, described limb action recognition unit connects three-dimensional user interface unit, three Wesys
Family boundary element connects stereoscopic image display unit and three-dimensional position indicator unit respectively.
The present invention compared with prior art, uses the depth image towards 3 D stereoscopic image display system to sense
Unit and depth image processing unit, all or part of limb action of user under sensing solid space, simultaneously
Utilize three-dimensional user interface to interact with system, and this three-dimensional user interface shows with towards 3 D stereoscopic image
Show that the display depth space of system is consistent, and can be carried out with user operation by three-dimensional position indicator
Feedback, effective opereating specification of user is also protected with the display depth space towards 3 D stereoscopic image display system
Hold consistent, and then arrive the man-machine interaction effect immersed completely, thus improve Consumer's Experience greatly.
Accompanying drawing explanation
Fig. 1 is the structured flowchart of the present invention.
Fig. 2 is the flow chart of the present invention.
Detailed description of the invention
With embodiment, the present invention is described in further detail below in conjunction with the accompanying drawings.
As it is shown in figure 1, the present invention's includes with lower component towards 3 D stereoscopic image display system:
Stereoscopic image display unit: for receiving the 3 D stereoscopic image and three of three-dimensional user interface unit transmission
Dimension graphical interface of user;
Depth image sensing unit: be used for obtaining and sensing 3 D stereoscopic image display unit place environment, bag
Include user in interior deep image information;
Depth image processing unit: the visual range internal field depth of field obtained for analysis depth image sensing unit
Degree information, and identify the user object and whole body obtained by limb action recognition unit or part of limb motion;
Limb action recognition unit: be used for following the tracks of and identifying user action, and provide use for three-dimensional user interface unit
Family input controls interface;
Three-dimensional user interface unit: for sending graphical interfaces to stereoscopic image display unit, with 3 D stereo
The form of image is by display unit, and after receiving the user action of limb action recognition unit, moves user
Make to feed back on stereoscopic image display unit;
Three-dimensional position indicator unit: be used for receiving scene depth letter in depth image processing unit visual range
Identify active user's particular location at three-dimensional user interface after breath, and user's control action is made feedback.
Described depth image sensing unit connects depth image processing unit, and depth image processing unit connects limb
Body action recognition unit, described limb action recognition unit connects three-dimensional user interface unit, three dimensional user circle
Face unit connects stereoscopic image display unit and three-dimensional position indicator unit respectively.
Three-dimensional user interface unit of the present invention with the Windows operating system of Microsoft as reference, three dimensional user circle
Face unit is patterned, can include the interface elements such as window, icon, file, desktop, towards
Under 3 D stereoscopic image display system environment all of interface element all have depth attribute, i.e. interface object from
Open the distance of physical display screen it can be understood as a three-dimensional desktop in three dimensions.User can be led to
Crossing gesture motion to rotate, scale this three-dimensional desktop, all of window is the most vertically suspended in such a solid
In space, and any two is opened the depth difference between window and is maintained at more than the threshold range of setting, user
Can be captured by gesture motion, move, pull window interested and as focus window etc., this stands
The spatial dimension of body desktop keeps consistent with user's gesture motion effective range, such as extended when user's one hand
During relative user self the far-end upper right corner of effective range, three-dimensional position indicator also should move to this
The three-dimensional desktop far-end upper right corner, similar, when user's one hand is reduced to the relative user self of effective range
During the near-end lower left corner, three-dimensional position indicator also should move to this solid desktop near-end lower left corner.
The three-dimensional position indicator unit of the present invention can be a 3 D stereo in three dimensional user graphical interfaces
Little hand state, and corresponding metamorphosis, such as user can be made according to the gesture motion that user is different and doing
During propelling movement action, little hands is the five fingers expanded configuration, and user's little hands when doing drag motions is seized condition etc. of clenching fist,
User single-handed exercise be display one pinkie show device, when user does double-handed exercise again, be shown as two little
Hands indicator etc..Three-dimensional position indicator physical location in three dimensional user graphical interfaces is by user's hand phase
Displacement is determined, and direction and the speed of the moving direction of indicator and speed and user's hand motion keep one
Cause.
The present invention under the effect of depth image sensing unit, this system can obtain the limb action of user
Information, as a example by gesture motion, gesture motion effective range refers to one in physical display device screen and use
A specific region between the physical location of family, only after user's hand enters this region, system just can be right
User's gesture motion is identified, outside user's hand is positioned at this effective coverage, then and the gesture motion of user
Can be ignored by system, the locus scope of the gesture motion effective range application three-dimensional user interface of user is protected
Holding consistent, such user operation before 3 D stereo affects display device just can reach the people immersed completely
Machine interaction effect.
As in figure 2 it is shown, the present invention's uses following steps to realize towards 3 D stereoscopic image display system: one,
Stereoscopic image display unit projects virtual three dimensional user graphical interfaces in entity space;Two, depth image
Sensing unit, after stereoscopic image display unit projects virtual three dimensional user graphical interfaces, obtains user institute
In environment deep image information, and deep image information is sent to depth image processing unit;Three, the degree of depth
Graphics processing unit starts after receiving deep image information identify user object and follow the tracks of user's hand motion,
User's hand motion information is fed back to limb action recognition unit;Four, limb action recognition unit judges to use
Whether action gesture in family, in useful effect region, is, enters next step operation, otherwise returns previous step;
Five, limb action recognition unit identification user's gesture motion judge this action command;Six, limb action is known
Order after other unit will identify sends to three-dimensional user interface unit;Seven, three-dimensional user interface unit will letter
Breath sends to three-dimensional position indicator unit, updates three-dimensional position indicator post and state;Eight, three Wesy
User action control command is made respective feedback and shows on three-dimensional user interface unit by family boundary element.
The limb action recognition unit of the present invention can identify the bimanualness of user, and its step is as follows: one,
User's one hand enters gesture motion effective coverage;Two, three-dimensional position indicator unit updates three-dimensional little hand position
And form;Three, user moves in effective coverage, and pushes away when three dimensional indicia is positioned at certain window ranges
Send action;Four, limb action recognition unit identifies effective gesture motion for clicking on order, three dimensional user circle
This window is set to current focus window by face unit, and is placed before current all windows;Five,
Another hands is stretched into useful effect region by user;Six, three-dimensional position indicator unit is updated to two vertical
Body little hand state, position keeps consistent with two-hand positions;Seven, user's both hands are in current focus window ranges
It is relatively distant from action;Eight, limb action recognition unit identifies this effective gesture motion for amplifying window life
Order, this window is amplified to both hands by three-dimensional user interface, and to be relatively distant from distance proportion consistent.
The model that the stereoscopic image display unit of the present invention uses KangJia Group Co., Ltd to produce is
LC42MS96PD dimensional image display;Depth image sensing unit uses PrimeSense company to produce
Imageing sensor;The model that depth image processing unit uses KangJia Group Co., Ltd to produce is
The graphics processing unit of KK-Depther;Limb action recognition unit uses KangJia Group Co., Ltd raw
The limb action evaluator that model is KK-Montion produced;Three-dimensional user interface unit uses group of Konka stock
The three-dimensional user interface unit that model is KK-3DUI that part company limited produces;Three-dimensional position indicator unit
Use the three-dimensional position indicator that model is KK-3DCur that KangJia Group Co., Ltd produces;Action
Effective coverage is between 0.8 meter to 3.5 meters.
Claims (1)
1. one kind towards 3 D stereoscopic image display system, it is characterised in that: described towards 3 D stereoscopic image
Display system includes with lower component:
Stereoscopic image display unit: for receiving the 3 D stereoscopic image and three of three-dimensional user interface unit transmission
Dimension graphical interface of user;
Depth image sensing unit: be used for obtaining and sensing 3 D stereoscopic image display unit place environment, bag
Include user in interior deep image information;
Depth image processing unit: the visual range internal field depth of field obtained for analysis depth image sensing unit
Degree information, and identify the user object and whole body obtained by limb action recognition unit or part of limb motion;
Limb action recognition unit: be used for following the tracks of and identifying user action, and carry for three-dimensional user interface unit
Control interface is inputted for user;
Three-dimensional user interface unit: for sending graphical interfaces to stereoscopic image display unit, with 3 D stereo
The form of image is by display unit, and after receiving the user action of limb action recognition unit, moves user
Make to feed back on stereoscopic image display unit;
Three-dimensional position indicator unit: be used for receiving scene depth letter in depth image processing unit visual range
Identify active user's particular location at three-dimensional user interface after breath, and user's control action is made feedback;
Described depth image sensing unit connects depth image processing unit, and depth image processing unit connects limb
Body action recognition unit, described limb action recognition unit connects three-dimensional user interface unit, three dimensional user circle
Face unit connects stereoscopic image display unit and three-dimensional position indicator unit respectively;
Described towards the employing following steps realization of 3 D stereoscopic image display system: one, stereoscopic image display unit
Virtual three dimensional user graphical interfaces is projected in entity space;Two, depth image sensing unit is at three-dimensional shadow
After projecting virtual three dimensional user graphical interfaces as display unit, obtain user place environment depth image letter
Breath, and deep image information is sent to depth image processing unit;Three, depth image processing unit receives
Start identify user object and follow the tracks of user's hand motion after deep image information, user's hand motion is believed
Breath feeds back to limb action recognition unit;Four, limb action recognition unit judges that whether user action gesture exists
In useful effect region, it is to enter next step operation, otherwise return previous step;Five, limb action identification
Unit identification user's gesture motion also judges this action command;Six, after limb action recognition unit will identify
Order sends to three-dimensional user interface unit;Seven, three-dimensional user interface unit sends information to three-dimensional position
Indicator unit, updates three-dimensional position indicator post and state;Eight, three-dimensional user interface unit is to user
Action directive is made respective feedback and shows on three-dimensional user interface unit;
Described limb action recognition unit can identify the bimanualness of user, and its step is as follows: one, user
Singlehanded entrance gesture motion effective coverage;Two, three-dimensional position indicator unit updates three-dimensional little hand position and shape
State;Three, user moves in effective coverage, and do propelling movement when three dimensional indicia is positioned at certain window ranges dynamic
Make;Four, limb action recognition unit identifies effective gesture motion for clicking on order, three-dimensional user interface list
This window is set to current focus window by unit, and is placed before current all windows;Five, user
Another hands is stretched into useful effect region;Six, to be updated to the solid of two little for three-dimensional position indicator unit
Hand state, position keeps consistent with two-hand positions;Seven, user's both hands do phase in current focus window ranges
To away from action;Eight, limb action recognition unit identifies this effective gesture motion is to amplify window command,
This window is amplified to both hands by three-dimensional user interface, and to be relatively distant from distance proportion consistent;
The model that described stereoscopic image display unit uses KangJia Group Co., Ltd to produce is
LC42MS96PD dimensional image display;Depth image sensing unit uses PrimeSense company to produce
Imageing sensor;The model that depth image processing unit uses KangJia Group Co., Ltd to produce is
The graphics processing unit of KK-Depther;Limb action recognition unit uses KangJia Group Co., Ltd raw
The limb action evaluator that model is KK-Montion produced;Three-dimensional user interface unit uses group of Konka
The three-dimensional user interface unit that model is KK-3DUI that limited company produces;Three-dimensional position indicator list
Unit uses the three-dimensional position indicator that model is KK-3DCur that KangJia Group Co., Ltd produces;Dynamic
It is between 0.8 meter to 3.5 meters as effective coverage.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201110171017.XA CN102253713B (en) | 2011-06-23 | 2011-06-23 | Towards 3 D stereoscopic image display system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201110171017.XA CN102253713B (en) | 2011-06-23 | 2011-06-23 | Towards 3 D stereoscopic image display system |
Publications (2)
Publication Number | Publication Date |
---|---|
CN102253713A CN102253713A (en) | 2011-11-23 |
CN102253713B true CN102253713B (en) | 2016-10-12 |
Family
ID=44981016
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201110171017.XA Active CN102253713B (en) | 2011-06-23 | 2011-06-23 | Towards 3 D stereoscopic image display system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN102253713B (en) |
Families Citing this family (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106943743B (en) * | 2011-02-11 | 2024-04-26 | 漳州市阿思星谷电子科技有限公司 | Man-machine interaction control method |
CN102789312B (en) * | 2011-12-23 | 2016-03-23 | 苏州触达信息技术有限公司 | A kind of user interactive system and method |
CN102707878A (en) * | 2012-04-06 | 2012-10-03 | 深圳创维数字技术股份有限公司 | User interface operation control method and device |
CN102902355B (en) * | 2012-08-31 | 2015-12-02 | 中国科学院自动化研究所 | The space interaction method of mobile device |
TWI494792B (en) | 2012-09-07 | 2015-08-01 | Pixart Imaging Inc | Gesture recognition system and method |
CN103839040B (en) * | 2012-11-27 | 2017-08-25 | 株式会社理光 | Gesture identification method and device based on depth image |
US9667883B2 (en) * | 2013-01-07 | 2017-05-30 | Eminent Electronic Technology Corp. Ltd. | Three-dimensional image sensing device and method of sensing three-dimensional images |
CN103067727A (en) * | 2013-01-17 | 2013-04-24 | 乾行讯科(北京)科技有限公司 | Three-dimensional 3D glasses and three-dimensional 3D display system |
CN103974008A (en) * | 2013-01-30 | 2014-08-06 | 联想(北京)有限公司 | Information processing method and electronic equipment |
CN103543830B (en) * | 2013-10-28 | 2017-02-15 | 四川大学 | Method for mapping human skeleton points to virtual three-dimensional space points in three-dimensional display |
CN103530060B (en) * | 2013-10-31 | 2016-06-22 | 京东方科技集团股份有限公司 | Display device and control method, gesture identification method |
CN103995620A (en) * | 2013-12-02 | 2014-08-20 | 深圳市云立方信息科技有限公司 | Air touch system |
CN104915979A (en) * | 2014-03-10 | 2015-09-16 | 苏州天魂网络科技有限公司 | System capable of realizing immersive virtual reality across mobile platforms |
EP2993901A1 (en) * | 2014-09-03 | 2016-03-09 | Samsung Electronics Co., Ltd. | Method for displaying images and electronic device thereof |
CN105511599B (en) * | 2014-09-29 | 2019-06-25 | 联想(北京)有限公司 | Information processing method and device |
CN104536575A (en) * | 2015-01-04 | 2015-04-22 | 苏州易乐展示***工程有限公司 | Large screen interaction system realization method based on 3D sensing |
CN106681497A (en) * | 2016-12-07 | 2017-05-17 | 南京仁光电子科技有限公司 | Method and device based on somatosensory control application program |
CN106846564A (en) * | 2016-12-29 | 2017-06-13 | 湖南拓视觉信息技术有限公司 | A kind of intelligent access control system and control method |
CN108388351B (en) * | 2018-04-12 | 2024-03-12 | 深圳市正图科技有限公司 | Mixed reality experience system |
CN111045558A (en) * | 2018-10-12 | 2020-04-21 | 上海博泰悦臻电子设备制造有限公司 | Interface control method based on three-dimensional scene, vehicle-mounted equipment and vehicle |
CN112925430A (en) * | 2019-12-05 | 2021-06-08 | 北京芯海视界三维科技有限公司 | Method for realizing suspension touch control, 3D display equipment and 3D terminal |
CN111722769B (en) * | 2020-07-16 | 2024-03-05 | 腾讯科技(深圳)有限公司 | Interaction method, interaction device, display equipment and storage medium |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101952818A (en) * | 2007-09-14 | 2011-01-19 | 智慧投资控股67有限责任公司 | Processing based on the user interactions of attitude |
CN101986255A (en) * | 2010-11-05 | 2011-03-16 | 福州瑞芯微电子有限公司 | Semitransparent gradual three-dimensional user interface |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7178111B2 (en) * | 2004-08-03 | 2007-02-13 | Microsoft Corporation | Multi-planar three-dimensional user interface |
-
2011
- 2011-06-23 CN CN201110171017.XA patent/CN102253713B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101952818A (en) * | 2007-09-14 | 2011-01-19 | 智慧投资控股67有限责任公司 | Processing based on the user interactions of attitude |
CN101986255A (en) * | 2010-11-05 | 2011-03-16 | 福州瑞芯微电子有限公司 | Semitransparent gradual three-dimensional user interface |
Also Published As
Publication number | Publication date |
---|---|
CN102253713A (en) | 2011-11-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN102253713B (en) | Towards 3 D stereoscopic image display system | |
US11392212B2 (en) | Systems and methods of creating a realistic displacement of a virtual object in virtual reality/augmented reality environments | |
US20210181856A1 (en) | Augmented reality interaction techniques | |
US11954808B2 (en) | Rerendering a position of a hand to decrease a size of a hand to create a realistic virtual/augmented reality environment | |
JP7092445B2 (en) | Methods and systems that provide remote robotic control | |
Park et al. | Real-time 3D pointing gesture recognition for mobile robots with cascade HMM and particle filter | |
Suarez et al. | Hand gesture recognition with depth images: A review | |
JP5846662B2 (en) | Method and system for responding to user selection gestures for objects displayed in three dimensions | |
US20170235376A1 (en) | Systems and methods of direct pointing detection for interaction with a digital device | |
Stuerzlinger et al. | The value of constraints for 3D user interfaces | |
JP5807686B2 (en) | Image processing apparatus, image processing method, and program | |
JP2022504599A (en) | Methods and systems that provide remote robotic control | |
JP2011022984A (en) | Stereoscopic video interactive system | |
CN103365411A (en) | Information input apparatus, information input method, and computer program | |
CN103246351A (en) | User interaction system and method | |
CN103064514A (en) | Method for achieving space menu in immersive virtual reality system | |
Kulik | Building on realism and magic for designing 3D interaction techniques | |
CN104765454A (en) | Human muscle movement perception based menu selection method for human-computer interaction interface | |
CN105929946B (en) | A kind of natural interactive method based on virtual interface | |
WO2016102948A1 (en) | Coherent touchless interaction with stereoscopic 3d images | |
Gope et al. | Interaction with Large Screen Display using Fingertip & Virtual Touch Screen | |
Park et al. | 3D Gesture-based view manipulator for large scale entity model review | |
Xie et al. | Natural Bare-Hand Interaction for Remote Operating Large Touch Screen. | |
Kim et al. | User customizable interaction in coexistence space | |
Yamada et al. | Rebo: A life-like universal remote control |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant |