US9734582B2 - Remote pointing method - Google Patents

Remote pointing method Download PDF

Info

Publication number
US9734582B2
US9734582B2 US14/769,562 US201314769562A US9734582B2 US 9734582 B2 US9734582 B2 US 9734582B2 US 201314769562 A US201314769562 A US 201314769562A US 9734582 B2 US9734582 B2 US 9734582B2
Authority
US
United States
Prior art keywords
image
user
coordinates
region
interest
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US14/769,562
Other versions
US20160005173A1 (en
Inventor
Kamin LEE
Sungjin Kim
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
LG Electronics Inc
Original Assignee
LG Electronics Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by LG Electronics Inc filed Critical LG Electronics Inc
Assigned to LG ELECTRONICS INC. reassignment LG ELECTRONICS INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: LEE, Kamin, KIM, SUNGJIN
Publication of US20160005173A1 publication Critical patent/US20160005173A1/en
Application granted granted Critical
Publication of US9734582B2 publication Critical patent/US9734582B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • G06T7/0042
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/271Image signal generators wherein the generated image signals comprise depth maps or disparity maps
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01BMEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
    • G01B11/00Measuring arrangements characterised by the use of optical techniques
    • G01B11/24Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
    • G01B11/245Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures using a plurality of fixed, simultaneously operating transducers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/002Specific input/output arrangements not covered by G06F3/01 - G06F3/16
    • G06F3/005Input arrangements through a video camera
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • H04N13/0271
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/4223Cameras
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • G06T2207/10021Stereoscopic video; Stereoscopic image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30241Trajectory
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N2013/0074Stereoscopic image analysis
    • H04N2013/0081Depth or disparity estimation from stereoscopic image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N2013/0074Stereoscopic image analysis
    • H04N2013/0085Motion estimation from stereoscopic image signals

Definitions

  • the present invention is related to a remote pointing method.
  • Display devices such as broadcasting signal receivers (e.g., television) equipped with a network function receiving a data stream according to a real-time streaming scheme through a network such as the Internet, decoding the received data stream into an audio and a video signal, and processing the audio and the video signal are now widely used.
  • broadcasting signal receivers e.g., television
  • a network function receiving a data stream according to a real-time streaming scheme through a network such as the Internet, decoding the received data stream into an audio and a video signal, and processing the audio and the video signal.
  • Display devices can carry out such functions as selecting a predetermined object displayed on a screen according to a user's gesture and moving a pointer.
  • the aforementioned method can be called a remote pointing method.
  • Remote pointing methods used in the conventional display devices pose a problem that a computational process for detecting a user's gesture is complicated.
  • the present invention has been made in an effort to provide a remote pointing method simplifying a computational process for detecting a user's gesture.
  • a remote pointing method comprises capturing images by a first and a second camera disposed being separated spatially from each other; detecting a pointing part in a first image captured by the first camera; determining a region of interest including the pointing part in a second image captured by the second camera; and extracting stereoscopic coordinates of the pointing part within the region of interest.
  • the remote pointing method can further comprise tracking a movement trajectory of the pointing part by using the stereoscopic coordinates.
  • the detecting the pointing part further comprises detecting first plane coordinates of the pointing part in the first image, and the determining a region of interest can further comprise extracting second plane coordinates of the pointing part.
  • the extracting stereoscopic coordinates can detect the stereoscopic coordinates by using a disparity between the first plane coordinates and the second plane coordinates.
  • the first plane coordinates include coordinates of a first central point of the pointing part in the first image; the second plane coordinates include coordinates of a second central point of the pointing part; and the detecting stereoscopic coordinates detects stereoscopic coordinates about the central point of the pointing part by using a disparity between the first central coordinates and the second central coordinates.
  • the size of the region of interest in the second image can be larger than the size of the region occupied by the pointing part in the first image.
  • Another remote pointing method comprises capturing images by a first and a second camera disposed being separated spatially from each other; determining a first region of interest including a pointing part in a first image captured by the first camera; determining a second region of interest including the pointing part in a second image captured by the second camera and corresponding to the first region of interest; extracting a depth map by using the first and the second region of interest; and tracking a movement trajectory of the pointing part in the depth map.
  • the size of the first region of interest can be the same as the size of the second region of interest.
  • the present invention simplifies a computational process for detecting a user's gesture, thereby simplifying the structure of a driver and reducing manufacturing costs.
  • a remote pointing method can reduce resources required for detecting a user's gesture.
  • FIG. 1 illustrates a structure of a display device to which a remote pointing method according to the present invention is applied
  • FIG. 2 illustrates a motion of a pointing part of a user and a control method according to the motion
  • FIGS. 3 to 6 illustrate motion commands
  • FIGS. 7 to 21 illustrate a method for detecting a motion command in detail.
  • the broadcasting signal receiver described in this document as one example of a display device is a kind of intelligent image display device which adds a computer-supporting function to the broadcasting signal receiving function.
  • a function related to utilizing the Internet is supported in addition to the original broadcasting signal receiving function, more user-friendly interfaces such as a manual-type input device, a touch screen, and a spatial remote controller can be implemented.
  • the broadcasting signal receiver can connect to the Internet or a computer to carry out functions for e-mail, web-browsing, banking, or game.
  • a standard general-purpose OS can be used.
  • the broadcasting signal receiver allows various applications to be added or removed freely, for example, on a general-purpose OS kernel, which enables various user-friendly functions to be carried out.
  • Examples of the broadcasting signal receiver can include a network TV, HBBTV, and smart TV; depending on the situation, the broadcasting signal receiver can also be realized in a smart phone.
  • a broadcasting signal receiver is described as an example of a display device to which the present invention is applied.
  • a broadcasting signal receiver can be interpreted as a display device.
  • FIG. 1 illustrates a structure of a display device to which a remote pointing method according to the present invention is applied.
  • a display device 10 comprises a display unit 100 , a receiver 105 Q, an external device interfacing unit 135 Q, a storage unit 140 Q, a user input interface unit 150 Q, a controller 170 Q, and a command input unit 160 Q.
  • the display unit 100 can be regarded as a display panel.
  • the display device 10 can comprise a voice output unit 185 Q for outputting a voice, a sound sensing unit 191 Q for sensing a voice, and a motion sensing unit 192 Q for sensing a user's motion.
  • a display device 10 according to the present invention will be described in more detail below.
  • a receiver 105 Q comprises a tuner 110 Q, a demodulator 120 Q, and a network interface unit 130 Q.
  • the receiver 105 Q it is possible to design the receiver 105 Q not to include the network interface unit 130 Q while incorporating the tuner 110 Q and the demodulator 120 Q; on the other hand, it is also possible to design the receiver 105 Q not to include the tuner 110 Q and the demodulator 120 Q while incorporating the network interface unit 130 Q.
  • the tuner 110 Q selects a channel chosen by the user or Radio Frequency (RF) broadcasting signals corresponding to all of the preset channels among RF broadcasting signals received through an antenna. Also, the tuner 110 Q converts the chosen RF broadcasting signal to an intermediate frequency signal, a baseband video signal, or a voice signal.
  • RF Radio Frequency
  • the demodulator 120 Q receives a digital IF signal (DIF) converted by the tuner 110 Q and demodulates the received DIF.
  • DIF digital IF signal
  • the demodulator 120 Q can output a stream signal TS after carrying out demodulation and channel decoding.
  • the stream signal can be a signal multiplexing a video signal, a voice signal, or a data signal.
  • the stream signal output from the demodulator 120 Q can be input to the controller 170 Q.
  • the controller 170 Q carries out demultiplexing and video/voice signal processing, displays video on the display unit 100 , and outputs a voice through the voice output unit 185 Q.
  • the external device interface unit 135 Q can connect an external device to the display device 10 of the present invention.
  • the external device interface unit 135 Q can include an A/V input and output unit (not shown) or a wireless communication unit (not shown).
  • the network interface unit 130 Q provides an interface for connecting the display unit 10 to a wired/wireless network including the Internet.
  • the network interface unit 130 Q can transmit and receive data to and from another user or another electronic device through a connected network or another network linked to the connected network.
  • the storage unit 140 Q can store a program for processing and controlling various signals within the controller 170 Q or store a signal-processed video, audio, or data signal.
  • the display unit 10 can play a contents file (a video file, a still image file, a music file, a document file, an application file, and so on) stored in the storage unit 140 Q and provide contents to the user.
  • a contents file a video file, a still image file, a music file, a document file, an application file, and so on
  • the command input unit 160 Q can include an input key for receiving a user's command.
  • the command input unit 160 Q can comprise a wired input unit 190 Q for receiving a command in a wired manner and a wireless input unit 200 Q for receiving a command wirelessly.
  • the user can input various commands such as power on/off, channel selection, display setting, volume control, movement of a cursor or a pointer on a screen, and menu selection.
  • a wireless input unit 200 Q can be a remote controller.
  • the user input interface unit 150 Q can deliver a signal that the user provides through the command input unit 160 Q to the controller 170 Q or deliver a signal from the controller 170 Q to the command input unit 160 Q.
  • the controller 170 Q can demultiplex streams input through the tuner 110 Q, the demodulator 120 Q, or the external device interface unit 135 Q; or process demultiplexed signals to generate and output a video or a voice signal.
  • the video signal processed in the controller 170 Q is fed to the display unit 100 to be displayed as a video according to the video signal. Also, the video signal processed in the controller 170 Q can be input to an external display device through the external device interface unit 135 .
  • the voice signal processed in the controller 170 Q can be output to the voice output unit 185 Q. Also, the voice signal processed in the controller 170 Q can be input to an external output device through the external device interface unit 135 Q.
  • the controller 170 Q can control the overall operation within the display device. For example, the controller 170 Q controls the tuner 110 Q to tune to an RF broadcasting program corresponding to the channel selected by the user or a preset channel.
  • controller 170 Q can control the display device 10 by using a user command or an internal program received through the user input interface unit 150 Q.
  • the controller 170 Q can control the display unit 100 to display images.
  • the controller 170 Q can control playing contents.
  • the contents in this case may correspond to the contents stored within the display unit 10 , received broadcasting contents, or contents input from the outside.
  • the contents can be at least one of a broadcasting image, an externally input image, an audio file, a still image, a web page accessed, and a document file.
  • the display unit 100 converts an image signal, a data signal, and an OSD signal processed in the controller 170 Q; or an image signal, a data signal, and so on received from the external device interface unit 135 Q to the respective RGB signals and generates a driving signal.
  • the display unit 100 can employ PDP, LCD, OLED, flexible display, 3D display, and so on.
  • the display unit 100 can be composed of a touch screen, thereby also being used as an input device in addition to being used as an output device.
  • the voice output unit 185 Q receives a voice-processed signal from the controller 170 Q, for example, a stereo signal, a 3.1 channel signal, or a 5.1 channel signal; and outputs the received voice-processed signal as a voice.
  • the voice output unit 185 Q can be implemented by various forms of speakers.
  • the sound sensing unit 191 Q can sense a sound generated inside or outside the display device 10 .
  • the sound sensing unit 191 Q can sense various kinds of sounds generated from the outside, such as a telephone ring sound, a doorbell sound, a sound of water, and a user's voice.
  • the sound sensing unit 191 Q can include a voice sensor or a microphone.
  • the motion sensing unit 192 Q can sense the motion of a user.
  • the motion sensing unit 192 Q can include a sensor unit (not shown) equipped with at least one of a touch sensor, a voice sensor, a position sensor, and a motion sensor.
  • the signal sensed by the sensing unit can be delivered to the controller 170 Q.
  • the motion sensing unit 192 Q can include a camera 200 .
  • the image information captured by the camera 200 can be input to the controller 170 Q.
  • the motion sensing unit 192 Q can include the sensor unit and the camera 200 together.
  • the controller 170 Q may detect the user's gesture by using the image captured by the camera 200 or a signal sensed by the sensing unit (not shown) separately or through a combination thereof.
  • the motion sensing unit 192 Q can sense the motion of a predetermined pointing part of the user's body according to the control of the controller 170 Q.
  • controller 170 Q can move the pointer displayed on the display unit 100 , select an object displayed on the display unit 100 , or carry out a predetermined function according to the motion of the pointing part.
  • the motion sensing unit 192 Q can also sense such information as the shape of the user's face and a viewing direction of the user's face as well as the user's motion.
  • the camera 200 can comprise multiple cameras. Multiple cameras can be configured to operate as a stereo camera.
  • FIG. 2 illustrates a motion of a pointing part of a user and a control method according to the motion.
  • FIG. 2 illustrates an example where the user inputs a motion command by using his/her hand 300 .
  • the user's hand 300 is set as a pointing part.
  • the present invention is not limited to the aforementioned example, however; the present invention can also include a case where the user gives a motion command by using other body part such as the user's arm or leg.
  • the pointing part according to the present invention can be the wireless input unit 200 Q.
  • the pointer C displayed on the display unit 100 can move up and down, side to side in accordance to the movement of the user's hand 300 .
  • the pointer C displayed on the display unit 100 can also rotate accordingly.
  • the motion sensing unit 192 Q can detect the motion.
  • the motion sensing unit 192 Q can use a camera to detect the motion of the user.
  • the information about the motion of the user's hand sensed by the motion sensing unit 192 Q is sent to the controller 170 Q of the display unit 10 .
  • the controller 170 Q can calculate coordinates of the position at which to display the pointer C from the information about the motion of the user's hand 300 . Moreover, the display unit 100 of the display device 10 can display the pointer C so that the pointer C corresponds to the coordinates calculated by the controller 170 Q.
  • a selected region within the display unit 100 corresponding to the pointer C can be zoomed in to be viewed bigger and closer or zoomed out to be viewed smaller and further away in accordance with the user's particular motion.
  • FIGS. 3 to 6 illustrate motion commands.
  • the motion sensing unit 192 Q described above senses the motion of a pointing part described below.
  • the motion command can also be generated by the motion of other object such as the wireless input unit 200 Q and a rod.
  • the display device 10 when the display device 10 enters a remote pointing mode S 300 , the display device 10 can display a pointer C on the display unit 100 , S 310 .
  • a remote key input mode can be set when the user directly enters a command through an input means such as the wireless input unit 200 Q commanding entrance to the remote pointing mode.
  • the remote pointing mode can be set by the user's motion command.
  • Setting the remote pointing mode can also be done when the user selects a predetermined icon or an object displayed on the display unit 100 .
  • the present invention is not limited to the cases above, and the method for setting the remote pointing mode can be changed in various ways.
  • the pointer C can be moved on the display unit 100 according to the motion command, a predetermined object displayed on the display unit 100 is selected, or a predetermined function is carried out S 330 .
  • the motion command is regarded as being received when a predetermined pointing part among the user's body moves according to a predetermined pattern in the remote pointing mode.
  • the pointer C displayed on the display unit 100 moves from left to right in accordance with the movement of the user's hand 300 as shown in FIG. 4(A) .
  • FIG. 4 corresponds to the case where the user's hand 300 is configured as the pointing part.
  • movement of the user's hand 300 can correspond to the motion command for moving the pointer C.
  • FIG. 4 is based on an assumption that the user's hand 300 , namely, the pointing port moves along a horizontal direction; however, the present invention is not limited to the assumption.
  • the pointer 300 can move along a vertical direction (DRV) on the display unit 100 .
  • the pointer C can be rotated or moved along a diagonal direction on the display unit 100 .
  • an object 400 corresponding to the pointer C from among the objects displayed on the display unit 100 can be selected as shown in FIG. 5(A) .
  • registration information corresponding to the object 400 can be displayed on the display unit 100 .
  • FIG. 5 corresponds to the case where the user's hand 300 is set as the pointing part or the index finger of the user's hand 300 is set as the pointing part.
  • motion commands are generated as the user folds his/her index finger or the user moves his/her hand 300 , but the present invention is not limited to the examples above.
  • a motion command corresponding to a function of moving the pointer C can be generated.
  • a motion command can be input when the distance between the pointing part (hand 300 ) and the display unit 100 is changed.
  • a motion command corresponding to a function of selecting a predetermined object displayed on the display unit 100 can be generated.
  • FIGS. 7 to 21 illustrate a method for detecting a motion command in detail.
  • the pointing part is the user's hand 300 .
  • Detection of a motion command can be regarded as detecting the user's gesture. Detection of a motion command can also be regarded as detecting a motion of a pointing part.
  • a camera can capture the image of a subject S 400 .
  • a first camera 200 a and a second camera 200 b spatially disposed from each other by a predetermined distance L 1 can capture an image of a subject, for example, an image of the user 500 .
  • the first camera 200 a and the second camera 200 b can be included in the motion sensing part 192 Q described above.
  • an image captured by the first camera 200 a is called a first image while an image captured by the second camera 200 b is called a second image.
  • FIG. 9(A) shows a first image 600 captured by the first camera 200 a
  • FIG. 9(B) shows a second image 610 captured by the second camera 200 b.
  • the first image 600 and the second image 610 may look the same, but they can be different from each other.
  • the position of the user 500 on the first image 600 can be different slightly from that on the second image 610 .
  • the disparity above can be generated since the imaging position of the first camera 200 a with respect to the user 500 differs from that of the second camera 200 b.
  • the pointing part can be detected from either of the first 600 and the second image 610 .
  • the pointing part for example, the hand 300 of the user 500 can be detected from either of the first 600 and the second image 610 .
  • the controller 170 Q of the display device 10 can detect the hand 300 of the user 500 by comparing the first image 600 with the information about the hand stored in the memory and detecting the part corresponding to the pre-stored information from the first image 600 .
  • the user ( 500 )'s hand 300 can belong to the part denoted by the symbol 700 .
  • the area of the symbol 700 is called a first area.
  • image plane coordinates of the pointing part on the first image 600 can be extracted.
  • the plane coordinates extracted from the first image 600 are called first plane coordinates.
  • x-axis coordinates (Xa-Xb) and y-axis coordinates (Ya-Yb) of the first area 700 including the user ( 500 )'s hand 300 can be extracted.
  • the pointing part extracted from the first image 600 namely, a region of interest including the user ( 500 )'s hand 300 can be determined on the second image 610 , S 420 .
  • image plane coordinates of the pointing part in the region of interest 710 can be extracted.
  • the plane coordinates extracted from the second image 610 are called second plane coordinates.
  • x-axis coordinates ranging from Xc to Xd and y-axis coordinates ranging from Yc to Yd of the region of interest including the user ( 500 )'s hand 300 can be extracted.
  • FIGS. 12(A) and (B) provide examples of extracting plane coordinates of the first region 700 and the region of interest 710 including the user ( 500 )'s hand 300 , respectively; however, it is also possible to extract plane coordinates of the user ( 500 )'s hand 300 within the first region 700 and plane coordinates of the user ( 500 )'s hand 300 within the region of interest 710 .
  • the region of interest 710 can be determined based on the pointing part extracted from the first image 600 , namely, coordinates of the user ( 500 )'s hand 300 , in other words, first plane coordinates.
  • the coordinates of the hand 300 within the second image 610 can differ from the coordinates of the hand 300 within the first image 600 .
  • the size of the region of interest 710 in the second image 610 is larger than the pointing part in the first image 600 , namely, the size of a region occupied by the hand 300 .
  • the x-axis coordinates of the first region 700 including the hand 300 in the first image 600 range from Xa to Xb, while the y-axis coordinates range from Ya to Yb.
  • the x-axis coordinates of the region of interest 710 including the hand 300 in the second image 610 range from Xc to Xd, and contain the coordinates in the range of Xa to Xb, while the y-axis coordinates range from Yc to Yd, and contain the coordinates in the range of Ya to Yb.
  • the size of the region of interest 710 can be larger than that of the first region 700 .
  • the first region 700 is an arbitrary region including the hand 300 on the first image 600 ; however, the first region 700 can be the part of the first image 600 occupied by the hand 300 .
  • the pointing part included in the region of interest 710 for example, stereoscopic coordinates of the hand 300 can be extracted S 430 .
  • stereoscopic coordinates can be extracted.
  • the first plane coordinates can include the pointing part in the first image 600 , namely, coordinates of a central point of the user ( 500 )'s hand ( 300 ), while the second plane coordinates can include coordinates of a central point of the pointing part in the second image 610 .
  • the coordinates of a central point of the user ( 500 )'s hand 300 in the first image are called first central point coordinates
  • the coordinates of a central point of the user ( 500 )'s hand in the second image are called second central point coordinates.
  • stereoscopic coordinates about a central point of the pointing part can be extracted by using the spatial distance L 1 between the first 200 a and the second camera 200 b ; and the difference between the first and the second central point coordinates.
  • the plane coordinates of a central point of the user ( 500 )'s hand 300 namely, first central point coordinates (Xc 1 , Yc 1 ) are extracted from the first image 600
  • the plane coordinates of a central point of the user ( 500 )'s hand 300 namely, second central point coordinates (Xc 2 , Yc 2 ) are extracted from the second image 610 .
  • Yc 1 and Yc 2 can be the same coordinate values.
  • the difference between the first (Xc 1 , Yc 1 ) and the second central point coordinates (Xc 2 , Yc 2 ), for example, the disparity L 2 between the x-axis coordinates of the first (Xc 1 , Yc 1 ) and the second central point coordinates (Xc 2 , Yc 2 ) can be extracted.
  • the distance L 3 between the user 500 and the first 200 a and the second camera 200 b can be extracted.
  • the disparity L 2 between x-axis coordinates of the first (Xc 1 , Yc 1 ) and the second central point coordinates (Xc 2 , Yc 2 ) becomes larger; on the other hand, if the user 500 gets further away from the first 200 a and the second camera 200 b , the disparity L 2 between x-axis coordinates of the first (Xc 1 , Yc 1 ) and the second central point coordinates (Xc 2 , Yc 2 ) becomes relatively small.
  • the stereoscopic coordinates of the user ( 500 )'s hand 300 can be obtained.
  • the stereoscopic coordinates of the central point of the user ( 500 )'s hand 300 can be determined approximately as ((Xc 1 +Xc 2 )/2, Yc 1 (or Yc 2 ), L 3 ).
  • the trajectory of the user ( 500 )'s hand 300 can be extracted S 440 , as shown in FIG. 16 .
  • a motion of the user ( 500 )'s hand 300 can be eventually extracted and a motion command can be detected accordingly.
  • the example above describes a method for extracting a trajectory by using the central point of the user ( 500 )'s hand 300 , but the present invention is not limited to the example above.
  • first to fifth part P 1 a -P 5 a of the hand 300 can be selected in the first image 600
  • sixth to tenth part P 1 b -P 5 b of the hand 300 corresponding to the first to fifth part P 1 a -P 5 a can be selected in the second image 610 .
  • the first part P 1 a can correspond to the sixth part P 1 b , the second part P 2 a to the seventh part P 2 b , the third part P 3 a to the eighth part P 3 b , the fourth part P 4 a to the ninth part P 4 b , and the fifth part P 5 a to the tenth part P 5 b.
  • the trajectory of the hand 300 can be extracted.
  • a pointing part is detected in the first image 600 and a region of interest including the pointing part is determined in the second image 610 .
  • the present invention is not limited to the assumption above.
  • a pointing part can be extracted from the second image 610 , and a region of interest including the pointing part can be determined in the first image 600 .
  • a first camera 200 a and a second camera 200 b can capture a first 600 and a second image 610 .
  • a depth map can be constructed by using the first image 600 and the second image 610 , S 510 .
  • a depth map is constructed based on disparities between the first 600 and the second image 610 for all the pixel positions.
  • Depth map is a well-known method in the field of stereopsis, of which detailed descriptions will be omitted.
  • a predetermined pointing part can be extracted from a constructed depth map.
  • Using a depth map to extract a pointing part may require an additional computing chip since construction of the depth map consists of a series of complicated computational processes.
  • a pointing part is extracted in the first image 600 , a region of interest corresponding to the pointing part is then determined in the second image 610 , and a method for extracting stereoscopic coordinates of the pointing part with respect to the region of interest is finally used, manufacturing costs can be reduced since there will be no need to add a computing chip.
  • a region of interest including the pointing part can be determined for each of the first 600 and the second image 610 , a depth map can be constructed based on the regions of interest to extract a motion trajectory of the pointing part.
  • this scheme is described.
  • an image of a subject for example, an image of the user 500 can be captured S 600 by using a first 200 a and a second camera 200 b disposed spatially separated from each other by a predetermined distance L 1 .
  • a pointing part for example, a first region of interest including the user ( 500 )'s hand 300 can be determined S 610 from the first image 600 captured by the first camera 200 a.
  • a pointing part for example, a second region of interest including the user ( 500 )'s hand 300 can be determined S 620 from the second image 610 captured by the second camera 200 b.
  • a pointing part namely, the user ( 500 )'s hand 300 can be extracted from the first image 600 .
  • the controller 170 Q of the display device 10 can compare the information about the hand stored in a memory with the first image 600 and extract the user ( 500 )'s hand 300 by using a method for extracting a part corresponding to pre-stored information from the first image 600 .
  • a first region of interest 800 can be determined so that the hand 300 can be included therein.
  • the pointing part extracted from the first image 600 namely, the second region of interest 810 including the user ( 500 )'s hand 300 can be determined in the second image 610 .
  • the size of the first region of interest 800 can be the same as that of the second region of interest 810 .
  • the plane coordinates of the first region of interest 800 can be the same as those of the second region of interest 810 .
  • the x-axis coordinates of the first region of interest 800 on the first image 600 range from X 1 to X 2 and y-axis coordinates range from Y 1 to Y 2
  • the x-axis coordinates of the first region of interest 800 on the second image 610 can range from X 1 to X 2 and y-axis coordinates can range from Y 1 to Y 2 .
  • the reason why the size of the first region of interest 800 is kept to be the same as that of the second region of interest 810 is to construct a depth map based on the first 800 and the second region of interest 810 .
  • a depth map can be constructed based on the first 800 and the second region of interest 810 , S 630 .
  • image data of the first region of interest 800 are compared with the image data of the second region of interest 810 , and a depth map is constructed, which includes information about distance of the predetermined part in the first region of interest 800 , namely, the hand 300 and information about distance of the hand 300 in the second region of interest 810 .
  • a motion trajectory of the pointing part can be extracted by using depth maps constructed continuously S 640 .
  • a motion trajectory of the pointing part can be extracted by constructing a depth map of a relatively small size compared with a method based on a depth map of a full-scale image size utilizing the first 600 and the second image 610 . Accordingly, computational processes involved can be made to be relatively simple.

Abstract

The present invention relates to a remote point method. A remote pointing method according to the present invention comprises capturing images by a first and a second camera disposed being separated spatially from each other; detecting a pointing part in a first image captured by the first camera; determining a region of interest including the pointing part in a second image captured by the second camera; and extracting stereoscopic coordinates of the pointing part within the region of interest.

Description

CROSS-REFERENCE TO RELATED PATENT APPLICATION
This application is a U.S. National Stage Application under 35 U.S.C. §371 of PCT Application No. PCT/KR2013/001392, filed Feb. 21, 2013, whose entire disclosure is hereby incorporated by reference.
TECHNICAL FIELD
The present invention is related to a remote pointing method.
BACKGROUND ART
Display devices such as broadcasting signal receivers (e.g., television) equipped with a network function receiving a data stream according to a real-time streaming scheme through a network such as the Internet, decoding the received data stream into an audio and a video signal, and processing the audio and the video signal are now widely used.
Display devices can carry out such functions as selecting a predetermined object displayed on a screen according to a user's gesture and moving a pointer. The aforementioned method can be called a remote pointing method.
Remote pointing methods used in the conventional display devices pose a problem that a computational process for detecting a user's gesture is complicated.
DISCLOSURE Technical Problem
The present invention has been made in an effort to provide a remote pointing method simplifying a computational process for detecting a user's gesture.
Technical Solution
A remote pointing method according to the present invention comprises capturing images by a first and a second camera disposed being separated spatially from each other; detecting a pointing part in a first image captured by the first camera; determining a region of interest including the pointing part in a second image captured by the second camera; and extracting stereoscopic coordinates of the pointing part within the region of interest.
The remote pointing method can further comprise tracking a movement trajectory of the pointing part by using the stereoscopic coordinates.
The detecting the pointing part further comprises detecting first plane coordinates of the pointing part in the first image, and the determining a region of interest can further comprise extracting second plane coordinates of the pointing part.
The extracting stereoscopic coordinates can detect the stereoscopic coordinates by using a disparity between the first plane coordinates and the second plane coordinates.
The first plane coordinates include coordinates of a first central point of the pointing part in the first image; the second plane coordinates include coordinates of a second central point of the pointing part; and the detecting stereoscopic coordinates detects stereoscopic coordinates about the central point of the pointing part by using a disparity between the first central coordinates and the second central coordinates.
The size of the region of interest in the second image can be larger than the size of the region occupied by the pointing part in the first image.
Another remote pointing method according to the present invention comprises capturing images by a first and a second camera disposed being separated spatially from each other; determining a first region of interest including a pointing part in a first image captured by the first camera; determining a second region of interest including the pointing part in a second image captured by the second camera and corresponding to the first region of interest; extracting a depth map by using the first and the second region of interest; and tracking a movement trajectory of the pointing part in the depth map.
The size of the first region of interest can be the same as the size of the second region of interest.
Advantageous Effects
The present invention simplifies a computational process for detecting a user's gesture, thereby simplifying the structure of a driver and reducing manufacturing costs.
Furthermore, a remote pointing method according to the present invention can reduce resources required for detecting a user's gesture.
DESCRIPTION OF DRAWINGS
The accompanying drawings, which are included to provide a further understanding of the invention and are incorporated in and constitute a part of this application, illustrate embodiment(s) of the invention and together with the description serve to explain the principle of the invention.
FIG. 1 illustrates a structure of a display device to which a remote pointing method according to the present invention is applied;
FIG. 2 illustrates a motion of a pointing part of a user and a control method according to the motion;
FIGS. 3 to 6 illustrate motion commands; and
FIGS. 7 to 21 illustrate a method for detecting a motion command in detail.
BEST MODE
The technical object, characteristics, and advantages of the present invention will now be more clearly understood from detailed descriptions given below. In what follows, preferred embodiments according to the present invention will be given with reference to appended drawings. Throughout the document, the same reference number represents the same component. In the following description, if it is decided that the detailed description of known function or configuration related to the invention obscures the technical principles of the present invention, the corresponding description is omitted.
In what follows, a mobile terminal according to the present invention will be described in more detail with reference to appended drawings. A suffix such as “module” and “unit” may be assigned or used interchangeably to refer to elements or components. Use of such a suffix herein is merely intended to facilitate description of the specification, and the suffix itself is not intended to give any special meaning or function.
The broadcasting signal receiver described in this document as one example of a display device is a kind of intelligent image display device which adds a computer-supporting function to the broadcasting signal receiving function. As a function related to utilizing the Internet is supported in addition to the original broadcasting signal receiving function, more user-friendly interfaces such as a manual-type input device, a touch screen, and a spatial remote controller can be implemented. And thanks to the wired or wireless Internet capability, the broadcasting signal receiver can connect to the Internet or a computer to carry out functions for e-mail, web-browsing, banking, or game. To implement various functions above, a standard general-purpose OS can be used.
Therefore, the broadcasting signal receiver according to the present invention allows various applications to be added or removed freely, for example, on a general-purpose OS kernel, which enables various user-friendly functions to be carried out. Examples of the broadcasting signal receiver can include a network TV, HBBTV, and smart TV; depending on the situation, the broadcasting signal receiver can also be realized in a smart phone.
In what follows, a broadcasting signal receiver is described as an example of a display device to which the present invention is applied. In what follows, a broadcasting signal receiver can be interpreted as a display device.
FIG. 1 illustrates a structure of a display device to which a remote pointing method according to the present invention is applied.
With reference to FIG. 1, a display device 10 comprises a display unit 100, a receiver 105Q, an external device interfacing unit 135Q, a storage unit 140Q, a user input interface unit 150Q, a controller 170Q, and a command input unit 160Q.
The display unit 100 can be regarded as a display panel.
Furthermore, the display device 10 according to the present invention can comprise a voice output unit 185Q for outputting a voice, a sound sensing unit 191Q for sensing a voice, and a motion sensing unit 192Q for sensing a user's motion.
A display device 10 according to the present invention will be described in more detail below.
A receiver 105Q comprises a tuner 110Q, a demodulator 120Q, and a network interface unit 130Q. Depending on the needs, it is possible to design the receiver 105Q not to include the network interface unit 130Q while incorporating the tuner 110Q and the demodulator 120Q; on the other hand, it is also possible to design the receiver 105Q not to include the tuner 110Q and the demodulator 120Q while incorporating the network interface unit 130Q.
The tuner 110Q selects a channel chosen by the user or Radio Frequency (RF) broadcasting signals corresponding to all of the preset channels among RF broadcasting signals received through an antenna. Also, the tuner 110Q converts the chosen RF broadcasting signal to an intermediate frequency signal, a baseband video signal, or a voice signal.
The demodulator 120Q receives a digital IF signal (DIF) converted by the tuner 110Q and demodulates the received DIF.
The demodulator 120Q can output a stream signal TS after carrying out demodulation and channel decoding. At this time, the stream signal can be a signal multiplexing a video signal, a voice signal, or a data signal.
The stream signal output from the demodulator 120Q can be input to the controller 170Q.
The controller 170Q carries out demultiplexing and video/voice signal processing, displays video on the display unit 100, and outputs a voice through the voice output unit 185Q.
The external device interface unit 135Q can connect an external device to the display device 10 of the present invention. To this purpose, the external device interface unit 135Q can include an A/V input and output unit (not shown) or a wireless communication unit (not shown).
The network interface unit 130Q provides an interface for connecting the display unit 10 to a wired/wireless network including the Internet.
The network interface unit 130Q can transmit and receive data to and from another user or another electronic device through a connected network or another network linked to the connected network.
The storage unit 140Q can store a program for processing and controlling various signals within the controller 170Q or store a signal-processed video, audio, or data signal.
The display unit 10 can play a contents file (a video file, a still image file, a music file, a document file, an application file, and so on) stored in the storage unit 140Q and provide contents to the user.
The command input unit 160Q can include an input key for receiving a user's command. The command input unit 160Q can comprise a wired input unit 190Q for receiving a command in a wired manner and a wireless input unit 200Q for receiving a command wirelessly.
Through the command input unit 160Q, the user can input various commands such as power on/off, channel selection, display setting, volume control, movement of a cursor or a pointer on a screen, and menu selection.
At this time, a wireless input unit 200Q can be a remote controller.
The user input interface unit 150Q can deliver a signal that the user provides through the command input unit 160Q to the controller 170Q or deliver a signal from the controller 170Q to the command input unit 160Q.
The controller 170Q can demultiplex streams input through the tuner 110Q, the demodulator 120Q, or the external device interface unit 135Q; or process demultiplexed signals to generate and output a video or a voice signal.
The video signal processed in the controller 170Q is fed to the display unit 100 to be displayed as a video according to the video signal. Also, the video signal processed in the controller 170Q can be input to an external display device through the external device interface unit 135.
The voice signal processed in the controller 170Q can be output to the voice output unit 185Q. Also, the voice signal processed in the controller 170Q can be input to an external output device through the external device interface unit 135Q.
Moreover, the controller 170Q can control the overall operation within the display device. For example, the controller 170Q controls the tuner 110Q to tune to an RF broadcasting program corresponding to the channel selected by the user or a preset channel.
Also, the controller 170Q can control the display device 10 by using a user command or an internal program received through the user input interface unit 150Q.
The controller 170Q can control the display unit 100 to display images.
The controller 170Q can control playing contents. The contents in this case may correspond to the contents stored within the display unit 10, received broadcasting contents, or contents input from the outside. The contents can be at least one of a broadcasting image, an externally input image, an audio file, a still image, a web page accessed, and a document file.
The display unit 100 converts an image signal, a data signal, and an OSD signal processed in the controller 170Q; or an image signal, a data signal, and so on received from the external device interface unit 135Q to the respective RGB signals and generates a driving signal.
The display unit 100 can employ PDP, LCD, OLED, flexible display, 3D display, and so on.
Meanwhile, the display unit 100 can be composed of a touch screen, thereby also being used as an input device in addition to being used as an output device.
The voice output unit 185Q receives a voice-processed signal from the controller 170Q, for example, a stereo signal, a 3.1 channel signal, or a 5.1 channel signal; and outputs the received voice-processed signal as a voice. The voice output unit 185Q can be implemented by various forms of speakers.
The sound sensing unit 191Q can sense a sound generated inside or outside the display device 10.
For example, the sound sensing unit 191Q can sense various kinds of sounds generated from the outside, such as a telephone ring sound, a doorbell sound, a sound of water, and a user's voice. To this purpose, the sound sensing unit 191Q can include a voice sensor or a microphone.
The motion sensing unit 192Q can sense the motion of a user. To this purpose, the motion sensing unit 192Q can include a sensor unit (not shown) equipped with at least one of a touch sensor, a voice sensor, a position sensor, and a motion sensor. The signal sensed by the sensing unit can be delivered to the controller 170Q.
Moreover, the motion sensing unit 192Q can include a camera 200. The image information captured by the camera 200 can be input to the controller 170Q.
Or the motion sensing unit 192Q can include the sensor unit and the camera 200 together.
The controller 170Q may detect the user's gesture by using the image captured by the camera 200 or a signal sensed by the sensing unit (not shown) separately or through a combination thereof.
To be more specific, the motion sensing unit 192Q can sense the motion of a predetermined pointing part of the user's body according to the control of the controller 170Q.
Moreover, the controller 170Q can move the pointer displayed on the display unit 100, select an object displayed on the display unit 100, or carry out a predetermined function according to the motion of the pointing part.
Furthermore, the motion sensing unit 192Q can also sense such information as the shape of the user's face and a viewing direction of the user's face as well as the user's motion.
Though not shown in this document, the camera 200 can comprise multiple cameras. Multiple cameras can be configured to operate as a stereo camera.
In what follows, how multiples cameras are used will be described in more detail.
FIG. 2 illustrates a motion of a pointing part of a user and a control method according to the motion. In what follows, those descriptions given above will not be repeated.
FIG. 2 illustrates an example where the user inputs a motion command by using his/her hand 300. In other words, in the example of FIG. 2, the user's hand 300 is set as a pointing part. The present invention is not limited to the aforementioned example, however; the present invention can also include a case where the user gives a motion command by using other body part such as the user's arm or leg.
Also, the pointing part according to the present invention can be the wireless input unit 200Q.
With reference to FIG. 2, if the user moves his/her hand up and down, side to side, the pointer C displayed on the display unit 100 can move up and down, side to side in accordance to the movement of the user's hand 300.
Moreover, if the user rotates the hand 300, the pointer C displayed on the display unit 100 can also rotate accordingly.
If the user makes a predetermined motion, the motion sensing unit 192Q can detect the motion. The motion sensing unit 192Q can use a camera to detect the motion of the user.
The information about the motion of the user's hand sensed by the motion sensing unit 192Q is sent to the controller 170Q of the display unit 10.
The controller 170Q can calculate coordinates of the position at which to display the pointer C from the information about the motion of the user's hand 300. Moreover, the display unit 100 of the display device 10 can display the pointer C so that the pointer C corresponds to the coordinates calculated by the controller 170Q.
As described above, even for the case where the user controls the display device 10 by using his/her motion rather than a remote control device such as the wireless input unit 200Q, a selected region within the display unit 100 corresponding to the pointer C can be zoomed in to be viewed bigger and closer or zoomed out to be viewed smaller and further away in accordance with the user's particular motion.
Moreover, it is possible to move the pointer C displayed on the display unit 100 or select an object displayed on the display unit 100 according to the user's motion; it is still possible to select a predetermined key from an input menu displayed on the display unit 100.
FIGS. 3 to 6 illustrate motion commands. In what follows, descriptions about those parts already described above will be omitted. For example, it is preferable to regard that the motion sensing unit 192Q described above senses the motion of a pointing part described below.
Furthermore, though the following descriptions are based on an assumption that a motion command corresponds to the user's body motion, namely, gesture, the motion command can also be generated by the motion of other object such as the wireless input unit 200Q and a rod.
Referring to FIG. 3, when the display device 10 enters a remote pointing mode S300, the display device 10 can display a pointer C on the display unit 100, S310.
A remote key input mode can be set when the user directly enters a command through an input means such as the wireless input unit 200Q commanding entrance to the remote pointing mode. Similarly, the remote pointing mode can be set by the user's motion command.
Setting the remote pointing mode can also be done when the user selects a predetermined icon or an object displayed on the display unit 100. The present invention is not limited to the cases above, and the method for setting the remote pointing mode can be changed in various ways.
While the pointer C is displayed on the display unit 100, whether a motion command is received can be checked S320.
From the checking result, if it is determined that a motion command has been input, the pointer C can be moved on the display unit 100 according to the motion command, a predetermined object displayed on the display unit 100 is selected, or a predetermined function is carried out S330.
At this time, the motion command is regarded as being received when a predetermined pointing part among the user's body moves according to a predetermined pattern in the remote pointing mode.
For example, as shown in FIG. 4(B), if the user's hand 300 moves along a horizontal direction (DRH), more specifically, from left to right, the pointer C displayed on the display unit 100 moves from left to right in accordance with the movement of the user's hand 300 as shown in FIG. 4(A).
The example of FIG. 4 corresponds to the case where the user's hand 300 is configured as the pointing part.
As shown in FIG. 4, movement of the user's hand 300 can correspond to the motion command for moving the pointer C.
The example of FIG. 4 is based on an assumption that the user's hand 300, namely, the pointing port moves along a horizontal direction; however, the present invention is not limited to the assumption. For example, in case the hand 300 moves along a vertical direction (DRV), the pointer 300 can move along a vertical direction (DRV) on the display unit 100.
Moreover, in case the user's hand 300 rotates or moves along a diagonal direction, too, it is regarded that a motion command has been received, and the pointer C can be rotated or moved along a diagonal direction on the display unit 100.
A motion command following a pattern different from FIG. 4 will be described with reference to FIG. 5.
If the index finger of the user's hand 300 is folded while the index finger is stretched as shown in FIG. 5(B), an object 400 corresponding to the pointer C from among the objects displayed on the display unit 100 can be selected as shown in FIG. 5(A). Similarly, it is also possible to carry out a predetermined function corresponding to the object 400. Or registration information corresponding to the object 400 can be displayed on the display unit 100.
In this way, if the user performs a folding motion while the user's index finger is stretched, a function such as selecting the pointer C corresponding to the object 400 is possible. In this case, the movement of the user's index finger can correspond to a motion command.
The example of FIG. 5 corresponds to the case where the user's hand 300 is set as the pointing part or the index finger of the user's hand 300 is set as the pointing part.
In the examples of FIGS. 4 and 5, motion commands are generated as the user folds his/her index finger or the user moves his/her hand 300, but the present invention is not limited to the examples above.
For example, though not shown, in case the user moves his/her hand along a predetermined direction with the palm unfolded, a motion command corresponding to a function of moving the pointer C can be generated.
Also, a motion command corresponding to the function of selecting a predetermined object when the user opens and closes his/her hand,
Also, as shown in FIG. 6, a motion command can be input when the distance between the pointing part (hand 300) and the display unit 100 is changed.
For example, in case the user's hand 300 is set as the pointing part and the user's hand 300 approaches the display unit 100 closer than a predetermined threshold, a motion command corresponding to a function of selecting a predetermined object displayed on the display unit 100 can be generated.
FIGS. 7 to 21 illustrate a method for detecting a motion command in detail. In what follows, descriptions about those parts already described above will be omitted. In what follows, it is assumed that the pointing part is the user's hand 300.
Detection of a motion command can be regarded as detecting the user's gesture. Detection of a motion command can also be regarded as detecting a motion of a pointing part.
Referring to FIG. 7, a camera can capture the image of a subject S400.
To detect the motion of the pointing part, as shown in FIG. 8, a first camera 200 a and a second camera 200 b spatially disposed from each other by a predetermined distance L1 can capture an image of a subject, for example, an image of the user 500. At this time, the first camera 200 a and the second camera 200 b can be included in the motion sensing part 192Q described above.
In what follows, an image captured by the first camera 200 a is called a first image while an image captured by the second camera 200 b is called a second image.
FIG. 9(A) shows a first image 600 captured by the first camera 200 a, and FIG. 9(B) shows a second image 610 captured by the second camera 200 b.
At a first glance, the first image 600 and the second image 610 may look the same, but they can be different from each other.
For example, if the first 600 and the second image 610 are superimposed on each other, as shown in FIG. 10, the position of the user 500 on the first image 600 can be different slightly from that on the second image 610.
As shown in FIG. 8, the disparity above can be generated since the imaging position of the first camera 200 a with respect to the user 500 differs from that of the second camera 200 b.
After image capture S400, the pointing part can be detected from either of the first 600 and the second image 610.
For example, as shown in FIG. 11(A), the pointing part, for example, the hand 300 of the user 500 can be detected from either of the first 600 and the second image 610. More specifically, the controller 170Q of the display device 10 can detect the hand 300 of the user 500 by comparing the first image 600 with the information about the hand stored in the memory and detecting the part corresponding to the pre-stored information from the first image 600.
In FIG. 11(A), the user (500)'s hand 300 can belong to the part denoted by the symbol 700. In what follows, the area of the symbol 700 is called a first area.
As described above, in the step of extracting the pointing part on the first image 600, image plane coordinates of the pointing part on the first image 600 can be extracted. In what follows, the plane coordinates extracted from the first image 600 are called first plane coordinates.
For example, as shown in FIG. 12(A), x-axis coordinates (Xa-Xb) and y-axis coordinates (Ya-Yb) of the first area 700 including the user (500)'s hand 300 can be extracted.
Also, as shown in FIG. 11(B), the pointing part extracted from the first image 600, namely, a region of interest including the user (500)'s hand 300 can be determined on the second image 610, S420.
As described above, in the step of determining the region of interest 710 on the second image 610, image plane coordinates of the pointing part in the region of interest 710 can be extracted. In what follows, the plane coordinates extracted from the second image 610 are called second plane coordinates.
For example, as shown in FIG. 12(B), x-axis coordinates ranging from Xc to Xd and y-axis coordinates ranging from Yc to Yd of the region of interest including the user (500)'s hand 300 can be extracted.
FIGS. 12(A) and (B) provide examples of extracting plane coordinates of the first region 700 and the region of interest 710 including the user (500)'s hand 300, respectively; however, it is also possible to extract plane coordinates of the user (500)'s hand 300 within the first region 700 and plane coordinates of the user (500)'s hand 300 within the region of interest 710.
The region of interest 710 can be determined based on the pointing part extracted from the first image 600, namely, coordinates of the user (500)'s hand 300, in other words, first plane coordinates.
Since the second image 610 is captured by the second camera 200 b, and the first image 610 is captured by the first camera 200 a, the coordinates of the hand 300 within the second image 610 can differ from the coordinates of the hand 300 within the first image 600.
Accordingly, in order to make the hand 300 to belong to the region of interest 710 of the second image 610, it may be preferable that the size of the region of interest 710 in the second image 610 is larger than the pointing part in the first image 600, namely, the size of a region occupied by the hand 300.
For example, as shown in FIG. 13, suppose the x-axis coordinates of the first region 700 including the hand 300 in the first image 600 range from Xa to Xb, while the y-axis coordinates range from Ya to Yb. Then the x-axis coordinates of the region of interest 710 including the hand 300 in the second image 610 range from Xc to Xd, and contain the coordinates in the range of Xa to Xb, while the y-axis coordinates range from Yc to Yd, and contain the coordinates in the range of Ya to Yb.
As described above, the size of the region of interest 710 can be larger than that of the first region 700.
In the example above, it is assumed that the first region 700 is an arbitrary region including the hand 300 on the first image 600; however, the first region 700 can be the part of the first image 600 occupied by the hand 300.
Meanwhile, after the region of interest 710 is extracted from the second image 610, the pointing part included in the region of interest 710, for example, stereoscopic coordinates of the hand 300 can be extracted S430.
To be specific, by using the spatial distance L1 between the first 200 a and the second camera 200 b; and the difference between the first plane coordinates and the second plane coordinates, stereoscopic coordinates can be extracted.
To be more specific, the first plane coordinates can include the pointing part in the first image 600, namely, coordinates of a central point of the user (500)'s hand (300), while the second plane coordinates can include coordinates of a central point of the pointing part in the second image 610. At this time, the coordinates of a central point of the user (500)'s hand 300 in the first image are called first central point coordinates, while the coordinates of a central point of the user (500)'s hand in the second image are called second central point coordinates.
Also, in the step of extracting stereoscopic coordinates, stereoscopic coordinates about a central point of the pointing part can be extracted by using the spatial distance L1 between the first 200 a and the second camera 200 b; and the difference between the first and the second central point coordinates.
For example, as shown in FIG. 14(A), the plane coordinates of a central point of the user (500)'s hand 300, namely, first central point coordinates (Xc1, Yc1) are extracted from the first image 600, and as shown in FIG. 14(B), the plane coordinates of a central point of the user (500)'s hand 300, namely, second central point coordinates (Xc2, Yc2) are extracted from the second image 610. In this case, Yc1 and Yc2 can be the same coordinate values.
Referring to FIG. 15, now that the spatial distance L1 between the first 200 a and the second camera 200 b is known, the difference between the first (Xc1, Yc1) and the second central point coordinates (Xc2, Yc2), for example, the disparity L2 between the x-axis coordinates of the first (Xc1, Yc1) and the second central point coordinates (Xc2, Yc2) can be extracted.
Then by using the ratio of the spatial distance L1 between the first 200 a and the second camera 200 b to the disparity L2 between the x-axis coordinates of the first (Xc1, Yc1) and the second central point coordinates (Xc2, Yc2), the distance L3 between the user 500 and the first 200 a and the second camera 200 b can be extracted.
When the distance L3 between the user 500 and the first 200 a and the second camera 200 b is extracted, viewing angles of the first 200 a and the second camera 200 b can be taken into account.
If the user 500 approaches the first 200 a and the second camera 200 b, the disparity L2 between x-axis coordinates of the first (Xc1, Yc1) and the second central point coordinates (Xc2, Yc2) becomes larger; on the other hand, if the user 500 gets further away from the first 200 a and the second camera 200 b, the disparity L2 between x-axis coordinates of the first (Xc1, Yc1) and the second central point coordinates (Xc2, Yc2) becomes relatively small.
In this way, if the distance L3 between the user 500 and the first 200 a and the second camera 200 b is extracted while the first central point coordinates (Xc1, Yc1) of the user (500)'s hand 300 in the first image 600 and the second central point coordinates (Xc2, Yc2) of the user (500)'s hand 300 in the second image 610 are known, the stereoscopic coordinates of the user (500)'s hand 300 can be obtained.
For example, in the case of FIG. 15, the stereoscopic coordinates of the central point of the user (500)'s hand 300 can be determined approximately as ((Xc1+Xc2)/2, Yc1 (or Yc2), L3).
If the stereoscopic coordinates of the central point of the user (500)'s hand 300 are extracted continuously according to the method described above, the trajectory of the user (500)'s hand 300 can be extracted S440, as shown in FIG. 16.
As described above, if the trajectory of the user (500)'s hand 300 is extracted, a motion of the user (500)'s hand 300 can be eventually extracted and a motion command can be detected accordingly.
The example above describes a method for extracting a trajectory by using the central point of the user (500)'s hand 300, but the present invention is not limited to the example above.
For example, as shown in FIG. 17, first to fifth part P1 a-P5 a of the hand 300 can be selected in the first image 600, and sixth to tenth part P1 b-P5 b of the hand 300 corresponding to the first to fifth part P1 a-P5 a can be selected in the second image 610.
At this time, the first part P1 a can correspond to the sixth part P1 b, the second part P2 a to the seventh part P2 b, the third part P3 a to the eighth part P3 b, the fourth part P4 a to the ninth part P4 b, and the fifth part P5 a to the tenth part P5 b.
In this way, by extracting stereoscopic coordinates of a plurality of parts selected from the hand 300, the trajectory of the hand 300 can be extracted.
In the examples above, for the convenience of description, it has been assumed that a pointing part is detected in the first image 600 and a region of interest including the pointing part is determined in the second image 610. However, the present invention is not limited to the assumption above. For example, a pointing part can be extracted from the second image 610, and a region of interest including the pointing part can be determined in the first image 600.
Differently from the above, described will be a method for extracting a motion trajectory of a pointing part by using a depth map from the first 600 and the second image 610.
Referring to FIG. 18, a first camera 200 a and a second camera 200 b can capture a first 600 and a second image 610.
Afterwards, a depth map can be constructed by using the first image 600 and the second image 610, S510.
At this time, a depth map is constructed based on disparities between the first 600 and the second image 610 for all the pixel positions. Depth map is a well-known method in the field of stereopsis, of which detailed descriptions will be omitted.
Now, a predetermined pointing part can be extracted from a constructed depth map.
Next, by extracting the pointing part from depth maps constructed continuously, the trajectory of the pointing part can be detected eventually.
Using a depth map to extract a pointing part may require an additional computing chip since construction of the depth map consists of a series of complicated computational processes.
In case an additional computing chip is not used, a considerable part of a processor's resources has to be dedicated to construct a depth map, and a program for managing the depth map may also become considerably heavy.
On the other hand, if a pointing part is extracted in the first image 600, a region of interest corresponding to the pointing part is then determined in the second image 610, and a method for extracting stereoscopic coordinates of the pointing part with respect to the region of interest is finally used, manufacturing costs can be reduced since there will be no need to add a computing chip.
Moreover, since a depth map of a full-scale image size is not required, computational processes can be carried out in a relatively simple manner.
Meanwhile, according to the present invention, a region of interest including the pointing part can be determined for each of the first 600 and the second image 610, a depth map can be constructed based on the regions of interest to extract a motion trajectory of the pointing part. In what follows, this scheme is described.
Referring to FIG. 19, an image of a subject, for example, an image of the user 500 can be captured S600 by using a first 200 a and a second camera 200 b disposed spatially separated from each other by a predetermined distance L1.
Afterwards, a pointing part, for example, a first region of interest including the user (500)'s hand 300 can be determined S610 from the first image 600 captured by the first camera 200 a.
Also, a pointing part, for example, a second region of interest including the user (500)'s hand 300 can be determined S620 from the second image 610 captured by the second camera 200 b.
For example, as shown in FIG. 20(A), a pointing part, namely, the user (500)'s hand 300 can be extracted from the first image 600.
At this time, the controller 170Q of the display device 10 can compare the information about the hand stored in a memory with the first image 600 and extract the user (500)'s hand 300 by using a method for extracting a part corresponding to pre-stored information from the first image 600.
Afterwards, a first region of interest 800 can be determined so that the hand 300 can be included therein.
As in the case of FIG. 20(B), the pointing part extracted from the first image 600, namely, the second region of interest 810 including the user (500)'s hand 300 can be determined in the second image 610.
At this time, as shown in FIG. 21, the size of the first region of interest 800 can be the same as that of the second region of interest 810. Preferably, the plane coordinates of the first region of interest 800 can be the same as those of the second region of interest 810.
For example, in case x-axis coordinates of the first region of interest 800 on the first image 600 range from X1 to X2 and y-axis coordinates range from Y1 to Y2, the x-axis coordinates of the first region of interest 800 on the second image 610 can range from X1 to X2 and y-axis coordinates can range from Y1 to Y2.
The reason why the size of the first region of interest 800 is kept to be the same as that of the second region of interest 810 is to construct a depth map based on the first 800 and the second region of interest 810.
Afterwards, a depth map can be constructed based on the first 800 and the second region of interest 810, S630.
To be specific, image data of the first region of interest 800 are compared with the image data of the second region of interest 810, and a depth map is constructed, which includes information about distance of the predetermined part in the first region of interest 800, namely, the hand 300 and information about distance of the hand 300 in the second region of interest 810.
Next, a motion trajectory of the pointing part can be extracted by using depth maps constructed continuously S640.
As described above, in case the first 800 and the second region of interest 810 are used to construct a depth map, as shown in FIG. 18, a motion trajectory of the pointing part can be extracted by constructing a depth map of a relatively small size compared with a method based on a depth map of a full-scale image size utilizing the first 600 and the second image 610. Accordingly, computational processes involved can be made to be relatively simple.
Though the present invention has been described with reference to a limited number of embodiments and drawings, the present invention is not limited to the embodiments, and it should be understood that various modifications and changes can be implemented from the descriptions by those skilled in the art.
Therefore, the technical scope of the present invention should not be limited to the embodiments of this document, but it should be defined by the appended claims or equivalents of the appended claims.

Claims (4)

The invention claimed is:
1. A remote pointing method, comprising:
capturing images by a first and a second camera disposed being separated spatially from each other;
detecting a pointing part in a first image captured by the first camera and first plane coordinates of the pointing part in the first image;
determining a region of interest including the pointing part in a second image captured by the second camera and second plane coordinates of the pointing part; and
extracting stereoscopic coordinates of the pointing part within a region of interest by using a disparity between the first plane coordinates and the second plane coordinates,
wherein the first plane coordinates include coordinates of a first central point of the pointing part in the first image, and the second plane coordinates include coordinates of a second central point of the pointing part, and
the extracting stereoscopic coordinates detects stereoscopic coordinates about the central point of the pointing part by using a disparity between the first central coordinates and the second central coordinates.
2. The method of claim 1, further comprising
tracking a movement trajectory of the pointing part by using the stereoscopic coordinates.
3. The method of claim 1, wherein a size of the region of interest in the second image is larger than a size of the region occupied by the pointing part in the first image.
4. A remote pointing method, comprising:
capturing images by a first and a second camera disposed being separated spatially from each other;
determining a first region of interest including a pointing part in a first image captured by the first camera;
determining a second region of interest including the pointing part in a second image captured by the second camera and corresponding to the first region of interest;
extracting a depth map by using the first and the second region of interest; and
tracking a movement trajectory of the pointing part in the depth map,
wherein a size of the first region of interest is the same as a size of the second region of interest.
US14/769,562 2013-02-21 2013-02-21 Remote pointing method Active 2033-02-24 US9734582B2 (en)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/KR2013/001392 WO2014129683A1 (en) 2013-02-21 2013-02-21 Remote pointing method

Publications (2)

Publication Number Publication Date
US20160005173A1 US20160005173A1 (en) 2016-01-07
US9734582B2 true US9734582B2 (en) 2017-08-15

Family

ID=51391455

Family Applications (1)

Application Number Title Priority Date Filing Date
US14/769,562 Active 2033-02-24 US9734582B2 (en) 2013-02-21 2013-02-21 Remote pointing method

Country Status (2)

Country Link
US (1) US9734582B2 (en)
WO (1) WO2014129683A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11121334B2 (en) * 2014-06-26 2021-09-14 Trustees Of Tufts College 3D graphene transistor
KR20170024715A (en) * 2015-08-26 2017-03-08 삼성전자주식회사 Object detection apparatus and object detection method thereof

Citations (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20010043719A1 (en) 1997-03-21 2001-11-22 Kenichi Harakawa Hand pointing device
US20030156756A1 (en) * 2002-02-15 2003-08-21 Gokturk Salih Burak Gesture recognition system using depth perceptive sensors
JP2004265222A (en) 2003-03-03 2004-09-24 Nippon Telegr & Teleph Corp <Ntt> Interface method, system, and program
KR20090048197A (en) 2007-11-09 2009-05-13 광주과학기술원 Calculation method and system of pointing locations, and collaboration system comprising it
US20100194741A1 (en) * 2009-01-30 2010-08-05 Microsoft Corporation Depth map movement tracking via optical flow and velocity prediction
US20100289826A1 (en) * 2009-05-12 2010-11-18 Samsung Electronics Co., Ltd. Method and apparatus for display speed improvement of image
US20100329511A1 (en) * 2009-06-25 2010-12-30 Samsung Electronics Co., Ltd. Apparatus and method for detecting hands of subject in real time
US20110083106A1 (en) * 2009-10-05 2011-04-07 Seiko Epson Corporation Image input system
KR20110099176A (en) 2010-03-01 2011-09-07 이문기 Pointing device of augmented reality
US20110298698A1 (en) 2010-06-04 2011-12-08 Acer Incorporated Manual human machine interface operation system and method thereof
US20120134536A1 (en) * 2010-11-26 2012-05-31 Myokan Yoshihiro Image Processing Apparatus and Method, and Program
US20120200494A1 (en) * 2009-10-13 2012-08-09 Haim Perski Computer vision gesture based control of a device
US20120319945A1 (en) * 2010-05-21 2012-12-20 Mccarthy John System and method for reporting data in a computer vision system
US20130120250A1 (en) * 2011-11-16 2013-05-16 Chunghwa Picture Tubes, Ltd. Gesture recognition system and method
US20130293679A1 (en) * 2012-05-02 2013-11-07 Primesense Ltd. Upper-Body Skeleton Extraction from Depth Maps
US20130295539A1 (en) * 2012-05-03 2013-11-07 Microsoft Corporation Projected visual cues for guiding physical movement
US20130343606A1 (en) * 2012-06-25 2013-12-26 Imimtek, Inc. Systems and methods for tracking human hands by performing parts based template matching using images from multiple viewpoints
US20140009384A1 (en) * 2012-07-04 2014-01-09 3Divi Methods and systems for determining location of handheld device within 3d environment
US20140104168A1 (en) * 2012-10-12 2014-04-17 Microsoft Corporation Touchless input
US20140177909A1 (en) * 2012-12-24 2014-06-26 Industrial Technology Research Institute Three-dimensional interactive device and operation method thereof
US20140184496A1 (en) * 2013-01-03 2014-07-03 Meta Company Extramissive spatial imaging digital eye glass apparatuses, methods and systems for virtual or augmediated vision, manipulation, creation, or interaction with objects, materials, or other entities
US20140292723A1 (en) * 2013-04-02 2014-10-02 Fujitsu Limited Information processing device and information processing method
US20140375547A1 (en) * 2012-03-13 2014-12-25 Eyesight Mobile Technologies Ltd. Touch free user interface
US20150116353A1 (en) * 2013-10-30 2015-04-30 Morpho, Inc. Image processing device, image processing method and recording medium
US20160124513A1 (en) * 2014-01-07 2016-05-05 Softkinetic Software Human-to-Computer Natural Three-Dimensional Hand Gesture Based Navigation Method
US9465444B1 (en) * 2014-06-30 2016-10-11 Amazon Technologies, Inc. Object recognition for gesture tracking

Patent Citations (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20010043719A1 (en) 1997-03-21 2001-11-22 Kenichi Harakawa Hand pointing device
US20030156756A1 (en) * 2002-02-15 2003-08-21 Gokturk Salih Burak Gesture recognition system using depth perceptive sensors
JP2004265222A (en) 2003-03-03 2004-09-24 Nippon Telegr & Teleph Corp <Ntt> Interface method, system, and program
KR20090048197A (en) 2007-11-09 2009-05-13 광주과학기술원 Calculation method and system of pointing locations, and collaboration system comprising it
US20100194741A1 (en) * 2009-01-30 2010-08-05 Microsoft Corporation Depth map movement tracking via optical flow and velocity prediction
US20100289826A1 (en) * 2009-05-12 2010-11-18 Samsung Electronics Co., Ltd. Method and apparatus for display speed improvement of image
US20100329511A1 (en) * 2009-06-25 2010-12-30 Samsung Electronics Co., Ltd. Apparatus and method for detecting hands of subject in real time
US20110083106A1 (en) * 2009-10-05 2011-04-07 Seiko Epson Corporation Image input system
US20120200494A1 (en) * 2009-10-13 2012-08-09 Haim Perski Computer vision gesture based control of a device
KR20110099176A (en) 2010-03-01 2011-09-07 이문기 Pointing device of augmented reality
US20120319945A1 (en) * 2010-05-21 2012-12-20 Mccarthy John System and method for reporting data in a computer vision system
US20110298698A1 (en) 2010-06-04 2011-12-08 Acer Incorporated Manual human machine interface operation system and method thereof
US20120134536A1 (en) * 2010-11-26 2012-05-31 Myokan Yoshihiro Image Processing Apparatus and Method, and Program
US20130120250A1 (en) * 2011-11-16 2013-05-16 Chunghwa Picture Tubes, Ltd. Gesture recognition system and method
US20140375547A1 (en) * 2012-03-13 2014-12-25 Eyesight Mobile Technologies Ltd. Touch free user interface
US20130293679A1 (en) * 2012-05-02 2013-11-07 Primesense Ltd. Upper-Body Skeleton Extraction from Depth Maps
US20130295539A1 (en) * 2012-05-03 2013-11-07 Microsoft Corporation Projected visual cues for guiding physical movement
US20130343606A1 (en) * 2012-06-25 2013-12-26 Imimtek, Inc. Systems and methods for tracking human hands by performing parts based template matching using images from multiple viewpoints
US20140009384A1 (en) * 2012-07-04 2014-01-09 3Divi Methods and systems for determining location of handheld device within 3d environment
US20140104168A1 (en) * 2012-10-12 2014-04-17 Microsoft Corporation Touchless input
US20140177909A1 (en) * 2012-12-24 2014-06-26 Industrial Technology Research Institute Three-dimensional interactive device and operation method thereof
US20140184496A1 (en) * 2013-01-03 2014-07-03 Meta Company Extramissive spatial imaging digital eye glass apparatuses, methods and systems for virtual or augmediated vision, manipulation, creation, or interaction with objects, materials, or other entities
US20140292723A1 (en) * 2013-04-02 2014-10-02 Fujitsu Limited Information processing device and information processing method
US20150116353A1 (en) * 2013-10-30 2015-04-30 Morpho, Inc. Image processing device, image processing method and recording medium
US20160124513A1 (en) * 2014-01-07 2016-05-05 Softkinetic Software Human-to-Computer Natural Three-Dimensional Hand Gesture Based Navigation Method
US9465444B1 (en) * 2014-06-30 2016-10-11 Amazon Technologies, Inc. Object recognition for gesture tracking

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
International Search Report dated Nov. 6, 2013 issued in Application No. PCT/KR2013/001392.

Also Published As

Publication number Publication date
US20160005173A1 (en) 2016-01-07
WO2014129683A1 (en) 2014-08-28

Similar Documents

Publication Publication Date Title
US10891012B2 (en) Mobile terminal, image display device and user interface provision method using the same
US9250707B2 (en) Image display apparatus and method for operating the same
KR101711828B1 (en) Image display apparatus, and method for operating the same
CN107925791B (en) Image display device and mobile terminal
EP3048799B1 (en) Image display apparatus
US9414125B2 (en) Remote control device
US9733718B2 (en) Display apparatus and display method thereof
CN107770604B (en) Electronic device and method of operating the same
KR20170124811A (en) Image display apparatus and operating method for the same
US20160054805A1 (en) Mobile input device and command input method using the same
CN111010508B (en) Shooting method and electronic equipment
KR102501713B1 (en) Method for displaying an image and an electronic device thereof
US20110119631A1 (en) Method and apparatus for operating user interface based on user&#39;s visual perspective in electronic display device
US9734582B2 (en) Remote pointing method
CN116711316A (en) Electronic device and operation method thereof
KR20150008769A (en) Image display apparatus, and method for operating the same
US9785248B2 (en) Display device and method for driving the same
KR102444150B1 (en) Image display apparatus
KR20120075952A (en) Method for controlling multimedia device by using remote controller and multimedia device thereof
KR20130072638A (en) Method for operating an image display apparatus
CN110493511B (en) Panoramic image generation method and mobile terminal
US20160062479A1 (en) Image display apparatus and method for operating the same
KR20230116662A (en) Image display apparatus
KR102014149B1 (en) Image display apparatus, and method for operating the same
KR20130123679A (en) Video confenrece apparatus, and method for operating the same

Legal Events

Date Code Title Description
AS Assignment

Owner name: LG ELECTRONICS INC., KOREA, REPUBLIC OF

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LEE, KAMIN;KIM, SUNGJIN;SIGNING DATES FROM 20150825 TO 20150827;REEL/FRAME:036833/0657

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4