WO2023040288A1 - Display device and device control method - Google Patents

Display device and device control method Download PDF

Info

Publication number
WO2023040288A1
WO2023040288A1 PCT/CN2022/090560 CN2022090560W WO2023040288A1 WO 2023040288 A1 WO2023040288 A1 WO 2023040288A1 CN 2022090560 W CN2022090560 W CN 2022090560W WO 2023040288 A1 WO2023040288 A1 WO 2023040288A1
Authority
WO
WIPO (PCT)
Prior art keywords
gesture
image data
display
camera
instruction
Prior art date
Application number
PCT/CN2022/090560
Other languages
French (fr)
Chinese (zh)
Inventor
孟祥菲
Original Assignee
海信视像科技股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 海信视像科技股份有限公司 filed Critical 海信视像科技股份有限公司
Priority to CN202280058406.8A priority Critical patent/CN117918057A/en
Publication of WO2023040288A1 publication Critical patent/WO2023040288A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/4223Cameras
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/4104Peripherals receiving signals from specially adapted client devices
    • H04N21/4126The peripheral being portable, e.g. PDAs or mobile phones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/441Acquiring end-user identification, e.g. using personal code sent by the remote control or by inserting a card
    • H04N21/4415Acquiring end-user identification, e.g. using personal code sent by the remote control or by inserting a card using biometric characteristics of the user, e.g. by voice recognition or fingerprint scanning
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Definitions

  • the present application relates to the technical field of intelligent display, and in particular to a display device and a device control method.
  • a display device equipped with a camera can provide users with functions such as taking pictures and mirrors.
  • the camera captures external images, and then transmits the captured images to the display for presentation.
  • the display device will process the images captured by the camera, and then display them on the display for users to browse, such as image processing technology.
  • the range of the camera’s shooting picture may be inconsistent with the display content range of the display.
  • the shooting picture range is large and the display only displays a part of the shooting picture. Seeing the corresponding gesture, the display device executes the gesture instruction action.
  • An embodiment of the present application provides a display device, including: a camera configured to collect image data; a display configured to present a user interface, wherein, when an application related to the camera is opened, the image data collected by the camera is presented; A controller connected to the camera and the display, the controller is configured to: respond to an instruction to open a camera-related application, control the display to present image data collected by the camera; Instructions for image data, enlarge the image data, and present at least a part of the enlarged image data on the display; in response to the preset gesture collected by the camera, acquire a gesture rectangle area, and generate A gesture instruction corresponding to the preset gesture; if the gesture rectangle area is within the user interface presented by the display, execute the gesture instruction; if the gesture rectangle area is not presented within the user interface, Then the gesture command is not executed.
  • the embodiment of the present application also provides another display device, including: a display; a camera; and a controller respectively connected to the camera and the display, and the controller is configured to: respond to a control signal for starting the camera , displaying the image collected by the camera on the display; in response to the control signal indicating that the face is focused, acquiring a rectangular area of the face position based on the image; determining an adjustment center point based on the rectangular area of the face position; The image is adjusted according to the adjustment center point and a preset adjustment value, a target image is acquired, and the target image is displayed on the display.
  • the embodiment of the present application also provides a focal length control method, including: controlling the display to present the image data collected by the camera; responding to an instruction input by the user for triggering the enlargement of the image data, enlarging the image data, and at least a part of the The enlarged image data is presented on the display, and based on the image data presented on the display, the focal length display range is obtained; in response to the preset gesture collected by the camera, the gesture rectangular area is obtained, and the predetermined gesture is generated.
  • Set the gesture instruction corresponding to the gesture if the gesture rectangle area is in the user interface presented by the display, execute the gesture instruction; if the gesture rectangle area is not presented in the user interface, then do not execute the gesture instruction gesture commands.
  • Fig. 1 is an operation scene between a display device and a control device according to some embodiments
  • FIG. 2 is a block diagram of a hardware configuration of a control device 100 according to some embodiments.
  • FIG. 3 is a block diagram of a hardware configuration of a display device 200 according to some embodiments.
  • FIG. 4 is a software configuration diagram in a display device 200 according to some embodiments.
  • Fig. 5a is an exemplary diagram of an image displayed in a display device 200 according to some embodiments.
  • Fig. 5b is an example diagram comparing the camera range and the actual display range according to some embodiments.
  • Fig. 6 is a flowchart of a focus control method according to some embodiments.
  • Fig. 7a is an example diagram of device control in a display device 200 according to some embodiments.
  • Fig. 7b is an example diagram of gesture control in another display device 200 according to some embodiments.
  • FIG. 8 is a flowchart of another device control method according to some embodiments.
  • FIG. 9 is an example diagram of a gesture area in a display device 200 according to some embodiments.
  • Fig. 10a is an example diagram of a rectangular area of a human face position in a display device 200 according to some embodiments.
  • Fig. 10b is an example diagram of determining an adjustment center point in a rectangular area of a face position in a display device 200 according to some embodiments;
  • Fig. 10c is an example diagram of determining an adjustment center point for a rectangular area of a face position in another display device 200 according to some embodiments;
  • Fig. 10d is an example diagram of displaying a target image in a display device 200 according to some embodiments.
  • FIG. 11 is a flow chart of yet another focus control method according to some embodiments.
  • FIG. 12 is an example diagram of determining an adjustment center point in another display device 200 in a rectangular area of a human face according to some embodiments;
  • Fig. 13 is a flow chart of yet another focus control method according to some embodiments.
  • Fig. 14 is an example diagram of displaying an object image in another display device 200 according to some embodiments.
  • Fig. 1 is a schematic diagram of an operation scene between a display device and a control device according to an embodiment. As shown in FIG. 1 , the user can operate the display device 200 through the smart device 300 or the control device 100 .
  • control device 100 may be a remote control, and the communication between the remote control and the display device includes infrared protocol communication, bluetooth protocol communication, and other short-distance communication methods, and the display device 200 is controlled wirelessly or wiredly.
  • the user can control the display device 200 by inputting user commands through buttons on the remote control, voice input, control panel input, and the like.
  • the smart device 300 (such as a mobile terminal, a tablet computer, a computer, a notebook computer, etc.) can also be used to control the display device 200 .
  • the display device 200 is controlled using an application program running on the smart device.
  • the display device 200 can also be controlled in a manner other than the control device 100 and the smart device 300.
  • the module for obtaining voice commands configured inside the display device 200 can directly receive the user's voice command control , the user's voice command control can also be received through the voice control device installed outside the display device 200.
  • the display device 200 also performs data communication with the server 400 .
  • the display device 200 may be allowed to communicate via a local area network (LAN), a wireless local area network (WLAN), and other networks.
  • the server 400 may provide various contents and interactions to the display device 200 .
  • FIG. 2 is a configuration block diagram of the control device 100 according to some embodiments.
  • the control device 100 includes a controller 110 , a communication interface 130 , a user input/output interface 140 , a memory, and a power supply.
  • the control device 100 can receive the user's input operation instructions, and convert the operation instructions into instructions that the display device 200 can recognize and respond to, and play an intermediary role between the user and the display device 200 .
  • FIG. 3 is a block diagram of a hardware configuration of a display device 200 according to some embodiments.
  • the display device 200 includes a tuner and demodulator 210, a communicator 220, a detector 230, an external device interface 240, a controller 250, a display 260, an audio output interface 270, a memory, a power supply, and a user interface. at least one.
  • the controller includes a processor, a video processor, an audio processor, a graphics processor, a RAM, a ROM, a first interface to an nth interface for input/output.
  • the display 260 includes a display screen component for presenting images, and a drive component for driving focus control, for receiving image signals output from the controller, and displaying video content, image content, and a menu control interface. Components and users manipulate the UI interface.
  • the display 260 can be a liquid crystal display, an OLED display, and a projection display, and can also be a projection device and a projection screen.
  • the communicator 220 is a component for communicating with external devices or servers according to various communication protocol types.
  • the communicator may include at least one of a Wifi module, a Bluetooth module, a wired Ethernet module and other network communication protocol chips or near field communication protocol chips, and an infrared receiver.
  • the display device 200 may establish transmission and reception of control signals and data signals with the external control device 100 or the server 400 through the communicator 220 .
  • the user interface may be used to receive a control signal from the control device 100 (such as an infrared remote controller, etc.).
  • the detector 230 may include a light receiver, an image acquisition module, a sound acquisition module, and the like.
  • the external device interface 240 may include, but is not limited to, the following: high-definition multimedia interface (HDMI), analog or data high-definition component input interface (component), composite video input interface (CVBS), USB input interface (USB), Any one or more interfaces such as RGB ports. It may also be a composite input/output interface formed by the above-mentioned multiple interfaces.
  • HDMI high-definition multimedia interface
  • component analog or data high-definition component input interface
  • CVBS composite video input interface
  • USB input interface USB
  • Any one or more interfaces such as RGB ports. It may also be a composite input/output interface formed by the above-mentioned multiple interfaces.
  • the tuner-demodulator 210 receives broadcast TV signals through wired or wireless reception, and demodulates audio and video signals, such as EPG data signals, from multiple wireless or cable broadcast TV signals.
  • the controller 250 and the tuner-demodulator 210 may be located in different split devices, that is, the tuner-demodulator 210 may also be located in an external device of the main device where the controller 250 is located, such as an external set-top box wait.
  • the controller 250 controls the operation of the display device and responds to user operations through various software control programs stored in the memory.
  • the controller 250 controls the overall operations of the display device 200 . For example, in response to receiving a user command for selecting a UI object to be displayed on the display 260, the controller 250 may perform an operation related to the object selected by the user command.
  • the controller includes a central processing unit (Central Processing Unit, CPU), a video processor, an audio processor, a graphics processing unit (Graphics Processing Unit, GPU), RAM Random Access Memory, RAM), ROM (Read- Only Memory, ROM), at least one of the first interface to the nth interface for input/output, a communication bus (Bus), and the like.
  • CPU Central Processing Unit
  • video processor video processor
  • audio processor audio processor
  • graphics processing unit Graphics Processing Unit, GPU
  • RAM Random Access Memory
  • ROM Read- Only Memory
  • CPU processor It is used to execute the operating system and application program instructions stored in the memory, and to execute various application programs, data and content according to various interactive instructions received from the outside, so as to finally display and play various audio and video content.
  • a CPU processor may include multiple processors. For example, including a main processor and one or more sub-processors.
  • the user can input user commands through a graphical user interface (GUI) displayed on the display 260, and the user input interface receives user input commands through the graphical user interface (GUI).
  • GUI graphical user interface
  • the user may input a user command by inputting a specific sound or gesture, and the user input interface recognizes the sound or gesture through a sensor to receive the user input command.
  • the system of the display device may include a kernel (Kernel), a command parser (shell), a file system and application programs.
  • Kernel kernel
  • shell command parser
  • file system file system
  • application programs application programs
  • the system is divided into four layers, from top to bottom are respectively the application (Applications) layer (abbreviated as “application layer”), application framework (Application Framework) layer (abbreviated as “framework layer”) "), the system library layer (referred to as the “system runtime layer”), and the kernel layer.
  • application layer application layer
  • application framework Application Framework
  • framework layer application framework
  • system library layer system library layer
  • kernel layer kernel layer
  • there is at least one application program running in the application program layer and these application programs can be window (Window) program, system setting program or clock program etc. that come with the operating system; they can also be developed by third-party developers. s application.
  • the application program packages in the application program layer are not limited to the above examples.
  • the kernel layer is a layer between hardware and software. As shown in Figure 4, the kernel layer at least includes at least one of the following drivers: audio driver, display driver, bluetooth driver, camera driver, WIFI driver, USB driver, HDMI driver, sensor driver (such as fingerprint sensor, temperature sensor, pressure sensors, etc.), and power drives, etc.
  • the kernel layer at least includes at least one of the following drivers: audio driver, display driver, bluetooth driver, camera driver, WIFI driver, USB driver, HDMI driver, sensor driver (such as fingerprint sensor, temperature sensor, pressure sensors, etc.), and power drives, etc.
  • taking the magic mirror application (photographing application) in FIG. Call the interface of the application framework layer, start the magic mirror application, and call the kernel layer to start the camera driver, so as to obtain images through the camera.
  • the applications that have the function of calling the camera in the display device are not only shooting applications, but also video call applications, mirror applications, audio and video recording applications, fitness application teaching applications, and so on.
  • shooting applications only the shooting application is taken as an example for development.
  • any application in this application that has the function of calling a camera is applicable to the technical solution claimed in this application.
  • the gesture recognition function can also be provided. For example, in starting a shooting application, when the user gestures a finger related to "yeah", the camera function may be triggered.
  • the display device involved in the present application when the display device involved in the present application is in a running state, its position in a certain physical space is relatively fixed, and the comfortable viewing distance selected by the user is relatively far away, or the distance from the display device is relatively long. So that when the camera application is opened to take an image, the captured portrait or user's face only accounts for a small part of the entire image. In addition, if multiple windows are presented on the display simultaneously in the display device, the portraits in the small windows will be further reduced, seriously affecting the user's enjoyment of interacting with the portraits on the display interface.
  • the display device involved in the present application can activate the "portrait follow" function when invoking the application related to the camera function.
  • the "portrait following” effect can be achieved by directly adjusting the focal length of the camera to enlarge the captured portrait; it is also possible to control the image captured by the camera centered on the recognized portrait or face without adjusting the focal length of the camera. Zoom in and display only a part of the camera image on the monitor to achieve the effect of "portrait following".
  • the camera application when the user uses the display device 200 to take a picture, the camera application has a function of adjusting the focal length of the camera/or zooming in on the captured image, and a function of taking pictures with gestures. Capable of taking pictures automatically.
  • the focal length adjustment function refers to the ability to adjust the focal length of the camera by executing a certain instruction from the user, so that the range of images collected by the camera can be changed.
  • the function of zooming in on the captured image refers to enlarging the display range of the image captured by the camera on the user interface under the condition that the range of the image captured by the camera remains unchanged by executing a certain instruction from the user.
  • the focal length adjustment function of the camera can change the size of the subject in the shooting picture, and synchronously, the display range of the subject on the user interface will also change accordingly.
  • the function of zooming in on the captured image does not substantially change the focal length of the camera, it can also zoom in on the captured image to change its display range on the user interface.
  • the relevant instruction controls are recorded as focus adjustment, they actually perform an action of zooming in on the captured image. In this case, in this application, it is considered that what is actually triggered is the function of zooming in on the captured image.
  • the entire camera shooting area will be divided into a part displayed on the user interface (referred to as the displayed part for short) and a part not displayed on the user interface (referred to as the non-displayed part for short).
  • the part that the user can actually watch is the display part, which will also be directly judged by the user as the image captured by the camera is the display part.
  • the user generally also controls the camera according to the display part.
  • the way to trigger the function of zooming in and shooting images includes: when the portrait following mode is turned on, the display part can automatically move and zoom dynamically as the user moves within the camera shooting range.
  • the image can also be zoomed manually through the user's selection, and the image on the user interface can also be moved and zoomed according to the user's selection.
  • the part that is not displayed is also a part of the picture captured by the camera. If the controller generally recognizes the entire captured picture when performing gesture recognition, it is very likely that the gesture that appears on the part that is not displayed on the display will also be recognized as The camera is triggered by the gesture camera command. At this time, the user does not see the corresponding gesture in the user interface, but the display device gives feedback on taking pictures, which affects the user experience.
  • FIG. 5a shows an example diagram of an image displayed in a display device 200 according to some embodiments.
  • an image is displayed on the display, and the image is an enlarged image captured by a camera and then presented on the display.
  • the image displayed on the display is different from the image corresponding to the shooting range of the camera.
  • the shooting range of the corresponding image captured by the camera is larger than the corresponding shooting range of the display image shown in FIG. 5a, that is, the display display range described in the embodiment of the present application.
  • the focal length display range may be equivalent to the actual image range displayed on the monitor.
  • the user opens the camera application program installed on the display device 200 by operating the control device, and then displays the original image on the display, that is, the image corresponding to the shooting range of the camera, and then triggers an instruction to zoom in on the captured image, and then displays the original image on the display.
  • the magnified captured image is displayed.
  • Fig. 6 shows a flowchart of an instruction control method according to some embodiments.
  • the controller when executing a gesture instruction according to the gesture position and the focus display range, the controller is configured to execute the instruction control method described in FIG. 6 , including:
  • the image data is enlarged, and at least a part of the enlarged image data is presented on the display, as in step 601 .
  • the command to trigger zooming in on an image may be input by the user through an operating control device (such as a remote control), or may be input by the user through a sound collector (such as a microphone) of the display device, depending on the application.
  • the scene selection setting is not limited in this application. It may also be an automatic zoom-in triggered according to the display size of the portrait in the captured image after the portrait following function is activated.
  • the initial display range presented on the display is the shooting range of the camera, and in response to an instruction to zoom in on the captured image, the captured image is enlarged and then cropped by a cropping frame, so
  • the image content in the cropping frame is the picture presented on the display (hereinafter referred to as the second image for short).
  • the cropping frame has the same aspect ratio as that of the display window in the display in this application.
  • the position of the cropping frame is adjusted so that the target object is included in the cropping frame, and the adjusted image frame is output to the display, wherein the
  • the adjusted image frame is the image data corresponding to the image data of the adjusted cropping frame input by the camera, the distance between the center of the target object and the center of the cropping frame is the smallest, and no black border appears on the display .
  • a coordinate system is created, the coordinate information of the enlarged image is acquired, the area of the cropping frame in the enlarged image is determined, and the display range of the display is updated according to the coordinate information of the cropping frame.
  • a gesture rectangle area is acquired in response to a gesture command detected by a camera, step 602 .
  • the gesture command refers to a special gesture detected by the camera, such as a command delivery method for completing a series of photographing actions through gesture recognition.
  • Gesture recognition can be understood as the user can use simple gestures to control or interact with
  • the camera collects images and detects that images include special gestures, and it is considered that a gesture command is detected.
  • the user can make different gestures within the shooting range of the camera to send different commands to the display device.
  • the camera detects the gesture command, processes the image including the gesture command, obtains the gesture rectangle area, more specifically, recognizes all key points of the gesture in the image area, and determines the minimum horizontal and vertical coordinates according to the coordinate information of all key points value and the maximum horizontal and vertical coordinate values to draw a rectangle to obtain the gesture rectangle area.
  • the gesture command is executed. Such as step 603. If it is determined that the gesture rectangle area is not within the display range of the user interface, the gesture command is not executed, as in step 604 .
  • the gesture command is executed in the gesture rectangle area within the display range of the second image, and the gesture command is not executed when the gesture rectangle area is not within the display range of the cropping frame. Only the gesture instructions within the display range of the captured image visible to the user are executed, so as to improve the effectiveness and accuracy of display device control.
  • the gesture rectangle area is within the display range of the second image, and the gesture instruction is executed. More specifically, the gesture instruction is a photographing instruction, and the photographing operation is executed, and the image is acquired and stored, further , the user can also edit the captured image through the control device, voice command or gesture command. experience.
  • the gesture rectangle area is not within the display range of the second image, and the gesture command is not executed, that is, although the gesture command is acquired within the shooting range of the camera, it is not within the visible range of the user , the gesture command is not executed to further improve the user experience.
  • the shooting area of the camera has a certain range, some collected images may not contain a complete target object.
  • some collected images may not contain a complete target object.
  • FIG. 7b only one hand is photographed at the upper left corner, and the completed character corresponding to the complete hand is not photographed.
  • the completed character corresponding to this hand cannot be captured even with the face following.
  • the actual display range includes a complete person
  • the camera shooting range includes the complete person and the gesture in the upper left corner. Therefore, since the gesture in the upper left corner is not within the actual display range, the corresponding gesture is not corresponding.
  • the gesture rectangle area is within the range displayed on the user interface through the coordinate information of the second image display range and the coordinate information of the gesture rectangle area, more specifically, the lower left corner of the camera shooting range
  • Fig. 8 shows a flowchart of another instruction processing method according to some embodiments. This embodiment further optimizes the foregoing method on the basis of the foregoing embodiments. As shown in Figure 8, the method includes:
  • step 801 it is determined whether an instruction to enlarge the captured image is received.
  • the display range of the captured image on the user interface is adjusted through a control device such as a remote control related button.
  • the non-display part as shown in FIG. 7a, is a display part within the display range on the user interface, and other parts are non-display parts.
  • the camera detects the gesture instruction, and executes the gesture instruction, such as step 802 .
  • the gesture command can be directly executed , where the corresponding gesture command can be set according to the requirements of the application scenario, for example, gesture command A is to take a picture, gesture command B is to store, etc.
  • the gesture command will not be executed.
  • the user inputs an instruction to zoom in on the captured image by operating a control device (such as a remote control) or the user inputs through a sound collector (such as a microphone) of the display device.
  • a control device such as a remote control
  • a sound collector such as a microphone
  • the display device does not substantially adjust the focal length of the camera, that is to say, the camera maintains the original shooting range. For example, as shown in FIG. 7 a , the entire camera shooting range remains unchanged. Instead, after the image captured by the camera is acquired, the image is enlarged and displayed on the display, so that the effect of zooming in on the object can also be achieved from the user's perspective.
  • the actual camera captures a larger image range, while the monitor only displays a part of the enlarged image.
  • the actual display range on the monitor shows the enlarged part of the image captured by the camera, so the actual range of the displayed image on the monitor is inconsistent with the range of the image captured by the camera.
  • the entire camera shooting area will be divided into For the display part and the non-display part, the display range of the second image is the actual display range on the user interface, that is, the display part, and the other parts are non-display parts.
  • the gesture command is executed in the gesture rectangle area within the second image display range, and the gesture command is not executed in the gesture rectangle area not in the second image display range.
  • the user's gesture is still within the range captured by the camera, if it is no longer within the range displayed by the display, then even if the gesture can be captured by the camera, the instruction is not recognized, or the instruction can be recognized but Does not respond to commands corresponding to gestures.
  • the user can perform adjustment and control operations according to the gestures intuitively seen on the display, improving the effectiveness and accuracy of display device control.
  • the coordinate information of the cropping frame is acquired, and the second image display range of the display is updated based on the coordinate information of the cropping frame, such as step 803 .
  • the coordinate axis is established with the lower left corner of the shooting range of the camera as the original coordinate point (0, 0), the lower boundary of the shooting range of the camera is the horizontal axis, and the left boundary is the vertical axis, creating a coordinate system, read Get the coordinate information of the cropping frame after zooming in and adjusting, for example, four coordinate points (x1, y1), (x1, y2), (x2, y1), (x2, y2), so as to determine the second Image display range.
  • the cropping frame is only used to locate the display range of the enlarged captured image on the display, so that the display will display according to the framed range, rather than actually cropping the enlarged captured image.
  • the controller directly presents the image in the cropping frame on the display, where the cropping frame is not visible.
  • the camera detects the gesture command, acquires multiple key points of the gesture command, and acquires the key point coordinate information corresponding to the multiple key points to determine the minimum horizontal and vertical coordinate values and the maximum horizontal and vertical coordinate values, based on the minimum The horizontal and vertical coordinate values and the maximum horizontal and vertical coordinate values are used to obtain the gesture rectangle area, as in step 804 .
  • a plurality of key points of the gesture command are obtained, and the key point coordinate information corresponding to the plurality of key points is obtained, and the lower left corner of the shooting range of the camera is taken as the original coordinate point (0, 0 ) to establish a coordinate axis, the lower boundary of the camera shooting range is the horizontal axis, and the left boundary is the vertical axis, create a coordinate system, read the coordinate information of key points, and determine the minimum horizontal and vertical coordinate values and maximum horizontal and vertical coordinate values, such as four coordinates Points (m1, n1), (m1, n2), (m2, n1), (m2, n2), so as to determine the gesture rectangle area according to the four coordinate points.
  • the target coordinate information of the second image display range is acquired, and the gesture coordinate information of the gesture rectangle area is acquired, such as step 805 .
  • the gesture instruction is executed, such as step 806 .
  • the gesture rectangle area is not within the display range of the second image, and the gesture command is not executed, as in step 807 .
  • the target coordinate information of the second image display range is obtained such as (x1, y1), (x1, y2), (x2, y1), (x2, y2) described above, and the gesture rectangle is obtained
  • the gesture coordinate information of the area is as described above (m1, n1), (m1, n2), (m2, n1), (m2, n2), according to the target coordinate information and gesture coordinate information, determine the gesture rectangle area in the second image Within the display range, the gesture instruction is executed, and based on the target coordinate information and the gesture coordinate information, it is determined that the gesture rectangle area is not within the display range of the second image, and the gesture instruction is not executed.
  • the camera interface can adjust the photographing display range through the remote controller and other control devices.
  • the photograph can be taken, and when the gesture is not within the focal length display range, the photograph cannot be taken.
  • the artificial intelligence makeup test is a makeup trial function on the camera application, by calculating the position of the human face, and drawing the makeup to the corresponding position of the human face
  • the makeup test function includes a portrait follow effect, or a face follow effect (also known as face follow focus).
  • face follow focus refers to when shooting a shot, Vary the range of the captured image displayed on the monitor as people or objects move closer to or away from the camera to keep them in precise focus.
  • Fig. 11 shows a flow chart of yet another instruction control method according to some embodiments.
  • the controller when following a portrait, the controller is configured to execute the instruction control method described in FIG. 11 , including:
  • the image captured by the camera is displayed on the display, as in step 1101 .
  • a control signal for starting the camera is input through buttons on the remote control, voice input, control panel input, etc., and the images acquired by the camera are displayed on the display.
  • the image displayed on the display may or may not include a human face image.
  • This application is mainly aimed at processing including a human face image, and the human face image may be in any position of the displayed image.
  • the lower right corner of the image shown in FIG. 10 above may also be the center position as shown in FIG. 12 , which is specifically determined according to the location of the user in the application scenario.
  • step 1102 In response to the control signal indicating that the face is in focus, acquire a rectangular area of the face position based on the image, as in step 1102 .
  • the face focus control signal is input through buttons on the remote control, voice input, control panel input, etc.
  • the face area refers to the area where the face is located, so as to perform face recognition on the image and obtain
  • the rectangular area of the face position the method of recognizing the face area of the image, includes but is not limited to the following steps, extracting the image features in the image, inputting the image features into the pre-trained face recognition model for judgment, and determining that the face recognition model is judged as human
  • the position area corresponding to the image feature of the face is the face image area, such as the rectangular area of the face position as shown in FIG. 12 .
  • the image is adjusted according to the adjustment center point and the preset adjustment value, the target image is acquired, and the target image is displayed on the display, such as step 1104 .
  • the adjustment center point refers to a coordinate point in the rectangular area of the face position
  • the adjustment center point is obtained by calculating the coordinate information of the rectangular area of the face position, specifically, obtaining the coordinates of the rectangular area of the face position
  • the information is calculated and processed based on the coordinate information of the rectangular area of the face position, and the coordinates of the center point are obtained as the adjustment center point; where the coordinate information refers to one or more coordinate points on the rectangular area of the face position, and the coordinates corresponding to the coordinate points value.
  • the preset adjustment value can be selected and set according to the needs of the application scenario.
  • the adjustment center point is used as the center point, and the pixels of the image are adjusted according to the preset adjustment value to acquire the target image and display the target image on the display.
  • the calculated adjustment center point is O as shown in Figure 10c, and the preset adjustment value is 1.5 times, and the image in Figure 10c is adjusted, and the obtained target image is shown in Figure 10d , to ensure the follow-focus effect.
  • the face area is used to calculate and adjust the center point to perform face follow focus to ensure the follow focus effect.
  • Fig. 13 shows a flow chart of another instruction control method according to some embodiments.
  • the foregoing focal length control method is further optimized.
  • the method includes:
  • the image captured by the camera is displayed on the display, and in response to a control signal indicating that the face follows focus, a rectangular area of the face position is acquired based on the image, such as step 1301 .
  • the image displayed on the display may or may not include a human face image.
  • This application is mainly aimed at processing including a human face image, and the human face image may be in any position of the displayed image. .
  • the face focus control signal is input through buttons on the remote control, voice input, control panel input, etc., so as to perform face recognition on the image and obtain the rectangular area of the face position.
  • the image captured by the camera is displayed on the display, and in response to the control signal indicating that the face is in focus, after obtaining the rectangular area of the human face position based on the image, the acquisition of the rectangular area of the human body position based on the image can be performed. Determining the adjustment center point based on the rectangular area of the face position or acquiring the rectangular area of the human body position based on the image, determining the adjustment center point based on the rectangular area of the face position and the rectangular area of the human body position, such as step 1303 .
  • the adjustment center point is determined based on the human face position rectangular area based on the fact that the rectangular area of the human body position cannot be obtained from the image, such as step 1302 .
  • the rectangular area of the human body position is the body area where the limbs and torso are located.
  • a human body detection algorithm can be used to determine it, or a human body detection model can be established. Furthermore, detection is performed based on the human body detection model, and the rectangular area of the human body position is obtained.
  • the human body position rectangle area cannot be obtained from the image, it means that the image only includes the human face position rectangle area, and the adjustment center point is determined based on the human face position rectangle area. That is to say, when only the rectangular area of human face position can be detected but the rectangular area of human body position is not detected, only the rectangular area of human face position is used as the adjustment center point for calculating the rectangular scaling area.
  • the coordinate information of the rectangular area of the face position is obtained, and the calculation process is performed based on the coordinate information of the rectangular area of the face position, and the coordinates of the center point are obtained as the adjustment center point. More specifically, in When the face follow focus function is turned on, the rectangular area of the face position and the rectangular area of the human body position are detected, and the coordinate axis is established with the lower left corner of the image as the original coordinate point (0, 0).
  • the lower boundary of the camera shooting range is the horizontal axis
  • the left boundary is For the vertical axis, create a coordinate system and read the coordinate information of the rectangular area of the face position, such as four coordinate points (a1, b1), (a1, b2), (a2, b1), (a2, b2), so that according to The four coordinate points determine the rectangular area of the face position, and the coordinates of the center point are further calculated as ((a1+a2)/2, (b1+b2)/2) as the adjustment center point.
  • the human body position rectangular area is acquired based on the image, and the adjustment center point is determined based on the human face position rectangular area and the human body position rectangular area.
  • the coordinate information of the rectangular area of the human face position is obtained based on the rectangular area of the human face position
  • the coordinate information of the rectangular area of the human body position is obtained based on the rectangular area of the human body position
  • the calculation process is performed based on the coordinate information of the rectangular area of the human face position
  • the face focus function when the face focus function is turned on, the rectangular area of the face position and the rectangular area of the human body position are detected, and the coordinate axis is established with the lower left corner of the image as the original coordinate point (0, 0), and the lower boundary of the camera shooting range is horizontal Axis, the left boundary is the vertical axis, create a coordinate system, read the coordinate information of the face position rectangle area and the human body position rectangle area coordinate information, for example, four coordinate points (a1, b1), (a1, b2), (a2 , b1), (a2, b2), so as to determine the rectangular area of the face position according to the four coordinate points, and the four coordinate points (c1, d1), (c1, d2), (c2, d1), (c2, d2), so as to determine the rectangular area of the human body position according to the four coordinate points.
  • the image is adjusted according to the adjustment center point and the preset adjustment value, the target image is acquired, and the target image is displayed on the display, such as step 1304.
  • the adjustment center point is determined based on the rectangular area of the face position, specifically, the coordinate information of the rectangular area of the face position is obtained, and the calculation process is performed based on the coordinate information of the rectangular area of the face position, and the coordinates of the center point are obtained as the adjustment center point.
  • the preset adjustment value can be selected and set according to the needs of the application scenario.
  • the adjustment center point is used as the center point, and the pixels of the image are adjusted according to the preset adjustment value to acquire the target image and display the target image on the display.
  • the calculated adjustment center point is O as shown in Figure 10a, and the preset adjustment value is 1.5 times, and the image in Figure 10a is adjusted, and the obtained image is shown in Figure 10c, The follow-focus effect is guaranteed.
  • the calculated adjustment center point is O as shown in Figure 12, and the preset adjustment value is 1.5 times, and the image in Figure 12 is adjusted, and the obtained image is shown in Figure 14. The follow-focus effect is guaranteed.
  • the face focus effect of the camera on the display interface of the monitor is based on the double inspection standard of the face position rectangle area and the body position rectangle area.
  • the face position rectangle area and the body position rectangle area are detected, the face position The rectangular area and the rectangular area of the human body position calculate the rectangular zoom area.
  • the rectangular area of the human body position cannot be detected, only the rectangular area of the face position is used to calculate the rectangular zoom area.
  • the rectangle scaling area is set as empty.
  • the face area is used to calculate and adjust the center point to perform face follow focus to ensure the follow focus effect.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Theoretical Computer Science (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

A display device and a device control method, the method comprising: in response to an instruction to open a camera-related application, controlling a display to show image data collected by a camera; in response to an instruction inputted by a user for use in triggering the enlargement of the image data, enlarging the image data, and showing at least a portion of the enlarged image data on the display; in response to a preset gesture collected by the camera, obtaining a gesture rectangle region, and generating a gesture instruction corresponding to the preset gesture; if the gesture rectangle region is in a user interface shown on the display, executing the gesture instruction; and if the gesture rectangle region is not shown in the user interface, the gesture instruction is not executed.

Description

显示设备及设备控制方法Display device and device control method
相关申请的交叉引用Cross References to Related Applications
本申请要求申请日为2021年9月18日,申请号为202111104825.4,和申请日为2021年9月27日,申请号为202111134859.8的中国申请的优先权,其全部内容引用于此。This application claims the priority of the Chinese application with the filing date of September 18, 2021, the application number is 202111104825.4, and the filing date is September 27, 2021, and the application number is 202111134859.8, the entire contents of which are cited here.
技术领域technical field
本申请涉及智能显示技术领域,尤其涉及一种显示设备及设备控制方法。The present application relates to the technical field of intelligent display, and in particular to a display device and a device control method.
背景技术Background technique
随着显示设备的快速发展,显示设备的功能越来越多,例如显示设备配备摄像头可以为用户提供照相、镜子等功能。With the rapid development of display devices, the display devices have more and more functions. For example, a display device equipped with a camera can provide users with functions such as taking pictures and mirrors.
一般情况下,摄像头拍摄外界图像,再将拍摄到的图像传输至显示器上呈现。在实际应用中,有可能显示设备会对摄像头拍摄到的画面进行处理后,再显示到显示器上供用户浏览,如图像处理技术等。Generally, the camera captures external images, and then transmits the captured images to the display for presentation. In practical applications, it is possible that the display device will process the images captured by the camera, and then display them on the display for users to browse, such as image processing technology.
然而,在一些图像处理过程中,有可能会导致摄像头拍摄画面的范围与显示器显示内容范围不一致的情况,如拍摄画面范围较大而显示器只显示拍摄画面一部分,用户在显示设备的显示界面中没有看到相应的手势,显示设备却执行手势指令动作。However, in some image processing processes, the range of the camera’s shooting picture may be inconsistent with the display content range of the display. For example, the shooting picture range is large and the display only displays a part of the shooting picture. Seeing the corresponding gesture, the display device executes the gesture instruction action.
发明内容Contents of the invention
本申请实施方式提供了一种显示设备,包括:摄像头,被配置为采集图像数据;显示器,被配置为呈现用户界面,其中,当打开摄像头相关应用时,呈现所述摄像头采集的图像数据;分别与所述摄像头和所述显示器连接的控制器,所述控制器被配置为:响应于打开摄像头相关应用的指令,控制显示器呈现摄像头采集的图像数据;响应于用户输入的用于触发放大所述图像数据的指令,将所述图像数据放大,并将至少一部分的所述放大后的图像数据呈现在所述显示器上;响应于所述摄像头采集的到预设手势,获取手势矩形区域,以及生成与所述预设手势相对应的手势指令;如果所述手势矩形区域在所述显示器呈现的用户界面内,则执行所述手势指令;如果所述手势矩形区域未呈现在所述用户界面内,则不执行所述手势指令。An embodiment of the present application provides a display device, including: a camera configured to collect image data; a display configured to present a user interface, wherein, when an application related to the camera is opened, the image data collected by the camera is presented; A controller connected to the camera and the display, the controller is configured to: respond to an instruction to open a camera-related application, control the display to present image data collected by the camera; Instructions for image data, enlarge the image data, and present at least a part of the enlarged image data on the display; in response to the preset gesture collected by the camera, acquire a gesture rectangle area, and generate A gesture instruction corresponding to the preset gesture; if the gesture rectangle area is within the user interface presented by the display, execute the gesture instruction; if the gesture rectangle area is not presented within the user interface, Then the gesture command is not executed.
本申请实施方式还提供了另一种显示设备,包括:显示器;摄像头;分别与所述摄像头和所述显示器连接的控制器,所述控制器被配置为:响应于启动所述摄像头的控制信号,在所述显示器中显示所述摄像头采集的图像;响应于指示人脸跟焦的控制信号,基于所述图像获取人脸位置矩形区域;基于所述人脸位置矩形区域,确定调整中心点;根据所述调整中心点和预设调整值对所述图像进行调整,获取目标图像,以及在所述显示器上显示所述目标图像。The embodiment of the present application also provides another display device, including: a display; a camera; and a controller respectively connected to the camera and the display, and the controller is configured to: respond to a control signal for starting the camera , displaying the image collected by the camera on the display; in response to the control signal indicating that the face is focused, acquiring a rectangular area of the face position based on the image; determining an adjustment center point based on the rectangular area of the face position; The image is adjusted according to the adjustment center point and a preset adjustment value, a target image is acquired, and the target image is displayed on the display.
本申请实施方式还提供一种焦距控制方法,包括:控制显示器呈现摄像头采集的图像数据;响应于用户输入的用于触发放大所述图像数据的指令,将所述图像数据放大,并将至少一部分的所述放大后的图像数据呈现在所述显示器上,基于显示器上呈现图像数据是获取焦距显示范围;响应于所述摄像头采集的到预设手势,获取手势矩形区域,以及生成与所述预设手势相对应的手势指令;如果所述手势矩形区域在所述显示器呈现的用户界面 内,则执行所述手势指令;如果所述手势矩形区域未呈现在所述用户界面内,则不执行所述手势指令。The embodiment of the present application also provides a focal length control method, including: controlling the display to present the image data collected by the camera; responding to an instruction input by the user for triggering the enlargement of the image data, enlarging the image data, and at least a part of the The enlarged image data is presented on the display, and based on the image data presented on the display, the focal length display range is obtained; in response to the preset gesture collected by the camera, the gesture rectangular area is obtained, and the predetermined gesture is generated. Set the gesture instruction corresponding to the gesture; if the gesture rectangle area is in the user interface presented by the display, execute the gesture instruction; if the gesture rectangle area is not presented in the user interface, then do not execute the gesture instruction gesture commands.
附图说明Description of drawings
图1为根据一些实施例的显示设备与控制装置之间操作场景;Fig. 1 is an operation scene between a display device and a control device according to some embodiments;
图2为根据一些实施例的控制设备100的硬件配置框图;FIG. 2 is a block diagram of a hardware configuration of a control device 100 according to some embodiments;
图3为根据一些实施例的显示设备200的硬件配置框图;FIG. 3 is a block diagram of a hardware configuration of a display device 200 according to some embodiments;
图4为根据一些实施例的显示设备200中软件配置图;FIG. 4 is a software configuration diagram in a display device 200 according to some embodiments;
图5a为根据一些实施例的一种显示设备200中显示图像的示例图;Fig. 5a is an exemplary diagram of an image displayed in a display device 200 according to some embodiments;
图5b为根据一些实施例的摄像头范围与实际显示范围对比的示例图;Fig. 5b is an example diagram comparing the camera range and the actual display range according to some embodiments;
图6为根据一些实施例的一种焦距控制方法的流程图;Fig. 6 is a flowchart of a focus control method according to some embodiments;
图7a为根据一些实施例的一种显示设备200中设备控制的示例图;Fig. 7a is an example diagram of device control in a display device 200 according to some embodiments;
图7b为根据一些实施例的另一种显示设备200中手势控制的示例图;Fig. 7b is an example diagram of gesture control in another display device 200 according to some embodiments;
图8为根据一些实施例的另一种设备控制方法的流程图;FIG. 8 is a flowchart of another device control method according to some embodiments;
图9为根据一些实施例的一种显示设备200中手势区域的示例图;FIG. 9 is an example diagram of a gesture area in a display device 200 according to some embodiments;
图10a为根据一些实施例的一种显示设备200中人脸位置矩形区域的示例图;Fig. 10a is an example diagram of a rectangular area of a human face position in a display device 200 according to some embodiments;
图10b为根据一些实施例的一种显示设备200中人脸位置矩形区域确定调整中心点的示例图;Fig. 10b is an example diagram of determining an adjustment center point in a rectangular area of a face position in a display device 200 according to some embodiments;
图10c为根据一些实施例的另一种显示设备200中人脸位置矩形区域确定调整中心点的示例图;Fig. 10c is an example diagram of determining an adjustment center point for a rectangular area of a face position in another display device 200 according to some embodiments;
图10d为根据一些实施例的一种显示设备200中显示目标图像的示例图;Fig. 10d is an example diagram of displaying a target image in a display device 200 according to some embodiments;
图11为根据一些实施例的又一种焦距控制方法的流程图;FIG. 11 is a flow chart of yet another focus control method according to some embodiments;
图12为根据一些实施例的另一种显示设备200中人脸位置矩形区域确定调整中心点的示例图;FIG. 12 is an example diagram of determining an adjustment center point in another display device 200 in a rectangular area of a human face according to some embodiments;
图13为根据一些实施例的再一种焦距控制方法的流程图;Fig. 13 is a flow chart of yet another focus control method according to some embodiments;
图14为根据一些实施例的另一种显示设备200中显示目标图像的示例图。Fig. 14 is an example diagram of displaying an object image in another display device 200 according to some embodiments.
具体实施方式Detailed ways
为使本申请的目的和实施方式更加清楚,下面将结合本申请示例性实施例中的附图,对本申请示例性实施方式进行清楚、完整地描述,显然,描述的示例性实施例仅是本申请一部分实施例,而不是全部的实施例。In order to make the purpose and implementation of the application clearer, the following will clearly and completely describe the exemplary implementation of the application in conjunction with the accompanying drawings in the exemplary embodiment of the application. Obviously, the described exemplary embodiment is only the present application. Claim some of the examples, not all of them.
图1为根据实施例中显示设备与控制装置之间操作场景的示意图。如图1所示,用户可通过智能设备300或控制装置100操作显示设备200。Fig. 1 is a schematic diagram of an operation scene between a display device and a control device according to an embodiment. As shown in FIG. 1 , the user can operate the display device 200 through the smart device 300 or the control device 100 .
在一些实施例中,控制装置100可以是遥控器,遥控器和显示设备的通信包括红外协议通信或蓝牙协议通信,及其他短距离通信方式,通过无线或有线方式来控制显示设备200。用户可以通过遥控器上按键、语音输入、控制面板输入等输入用户指令,来控制显示设备200。In some embodiments, the control device 100 may be a remote control, and the communication between the remote control and the display device includes infrared protocol communication, bluetooth protocol communication, and other short-distance communication methods, and the display device 200 is controlled wirelessly or wiredly. The user can control the display device 200 by inputting user commands through buttons on the remote control, voice input, control panel input, and the like.
在一些实施例中,也可以使用智能设备300(如移动终端、平板电脑、计算机、笔记本电脑等)以控制显示设备200。例如,使用在智能设备上运行的应用程序控制显示设备200。In some embodiments, the smart device 300 (such as a mobile terminal, a tablet computer, a computer, a notebook computer, etc.) can also be used to control the display device 200 . For example, the display device 200 is controlled using an application program running on the smart device.
在一些实施例中,显示设备200还可以采用除了控制装置100和智能设备300之外的方式进行控制,例如,可以通过显示设备200设备内部配置的获取语音指令的模块直接接收用户的语音指令控制,也可以通过显示设备200设备外部设置的语音控制设备来接收用户的语音指令控制。In some embodiments, the display device 200 can also be controlled in a manner other than the control device 100 and the smart device 300. For example, the module for obtaining voice commands configured inside the display device 200 can directly receive the user's voice command control , the user's voice command control can also be received through the voice control device installed outside the display device 200.
在一些实施例中,显示设备200还与服务器400进行数据通信。可允许显示设备200通过局域网(LAN)、无线局域网(WLAN)和其他网络进行通信连接。服务器400可以向显示设备200提供各种内容和互动。In some embodiments, the display device 200 also performs data communication with the server 400 . The display device 200 may be allowed to communicate via a local area network (LAN), a wireless local area network (WLAN), and other networks. The server 400 may provide various contents and interactions to the display device 200 .
图2为根据一些实施例的控制装置100的配置框图。如图2所示,控制装置100包括控制器110、通信接口130、用户输入/输出接口140、存储器、供电电源。控制装置100可接收用户的输入操作指令,且将操作指令转换为显示设备200可识别和响应的指令,起到用户与显示设备200之间交互中介作用。FIG. 2 is a configuration block diagram of the control device 100 according to some embodiments. As shown in FIG. 2 , the control device 100 includes a controller 110 , a communication interface 130 , a user input/output interface 140 , a memory, and a power supply. The control device 100 can receive the user's input operation instructions, and convert the operation instructions into instructions that the display device 200 can recognize and respond to, and play an intermediary role between the user and the display device 200 .
图3为根据一些实施例的显示设备200的硬件配置框图。FIG. 3 is a block diagram of a hardware configuration of a display device 200 according to some embodiments.
在一些实施例中,显示设备200包括调谐解调器210、通信器220、检测器230、外部装置接口240、控制器250、显示器260、音频输出接口270、存储器、供电电源、用户接口中的至少一种。In some embodiments, the display device 200 includes a tuner and demodulator 210, a communicator 220, a detector 230, an external device interface 240, a controller 250, a display 260, an audio output interface 270, a memory, a power supply, and a user interface. at least one.
在一些实施例中控制器包括处理器,视频处理器,音频处理器,图形处理器,RAM,ROM,用于输入/输出的第一接口至第n接口。In some embodiments the controller includes a processor, a video processor, an audio processor, a graphics processor, a RAM, a ROM, a first interface to an nth interface for input/output.
在一些实施例中,显示器260包括用于呈现画面的显示屏组件,以及驱动焦距控制的驱动组件,用于接收源自控制器输出的图像信号,进行显示视频内容、图像内容以及菜单操控界面的组件以及用户操控UI界面。In some embodiments, the display 260 includes a display screen component for presenting images, and a drive component for driving focus control, for receiving image signals output from the controller, and displaying video content, image content, and a menu control interface. Components and users manipulate the UI interface.
在一些实施例中,显示器260可为液晶显示器、OLED显示器、以及投影显示器,还可以为一种投影装置和投影屏幕。In some embodiments, the display 260 can be a liquid crystal display, an OLED display, and a projection display, and can also be a projection device and a projection screen.
在一些实施例中,通信器220是用于根据各种通信协议类型与外部设备或服务器进行通信的组件。例如:通信器可以包括Wifi模块,蓝牙模块,有线以太网模块等其他网络通信协议芯片或近场通信协议芯片,以及红外接收器中的至少一种。显示设备200可以通过通信器220与外部控制设备100或服务器400建立控制信号和数据信号的发送和接收。In some embodiments, the communicator 220 is a component for communicating with external devices or servers according to various communication protocol types. For example, the communicator may include at least one of a Wifi module, a Bluetooth module, a wired Ethernet module and other network communication protocol chips or near field communication protocol chips, and an infrared receiver. The display device 200 may establish transmission and reception of control signals and data signals with the external control device 100 or the server 400 through the communicator 220 .
在一些实施例中,用户接口,可用于接收控制装置100(如:红外遥控器等)的控制信号。In some embodiments, the user interface may be used to receive a control signal from the control device 100 (such as an infrared remote controller, etc.).
在一些实施例中,检测器230可包括光接收器、图像获取模块以及声音获取模块等。In some embodiments, the detector 230 may include a light receiver, an image acquisition module, a sound acquisition module, and the like.
在一些实施例中,外部装置接口240可以包括但不限于如下:高清多媒体接口(HDMI)、模拟或数据高清分量输入接口(分量)、复合视频输入接口(CVBS)、USB输入接口(USB)、RGB端口等任一个或多个接口。也可以是上述多个接口形成的复合性的输入/输出接口。In some embodiments, the external device interface 240 may include, but is not limited to, the following: high-definition multimedia interface (HDMI), analog or data high-definition component input interface (component), composite video input interface (CVBS), USB input interface (USB), Any one or more interfaces such as RGB ports. It may also be a composite input/output interface formed by the above-mentioned multiple interfaces.
在一些实施例中,调谐解调器210通过有线或无线接收方式接收广播电视信号,以及从多个无线或有线广播电视信号中解调出音视频信号,如以及EPG数据信号。In some embodiments, the tuner-demodulator 210 receives broadcast TV signals through wired or wireless reception, and demodulates audio and video signals, such as EPG data signals, from multiple wireless or cable broadcast TV signals.
在一些实施例中,控制器250和调谐解调器210可以位于不同的分体设备中,即调谐解调器210也可在控制器250所在的主体设备的外置设备中,如外置机顶盒等。In some embodiments, the controller 250 and the tuner-demodulator 210 may be located in different split devices, that is, the tuner-demodulator 210 may also be located in an external device of the main device where the controller 250 is located, such as an external set-top box wait.
在一些实施例中,控制器250,通过存储在存储器上中各种软件控制程序,来控制显示设备的工作和响应用户的操作。控制器250控制显示设备200的整体操作。例如:响应于接收到用于选择在显示器260上显示UI对象的用户命令,控制器250便可以执行与由用户命令选择的对象有关的操作。In some embodiments, the controller 250 controls the operation of the display device and responds to user operations through various software control programs stored in the memory. The controller 250 controls the overall operations of the display device 200 . For example, in response to receiving a user command for selecting a UI object to be displayed on the display 260, the controller 250 may perform an operation related to the object selected by the user command.
在一些实施例中控制器包括中央处理器(Central Processing Unit,CPU),视频处理器,音频处理器,图形处理器(Graphics Processing Unit,GPU),RAM Random Access Memory,RAM),ROM(Read-Only Memory,ROM),用于输入/输出的第一接口至第n接口,通信总线(Bus)等中的至少一种。In some embodiments, the controller includes a central processing unit (Central Processing Unit, CPU), a video processor, an audio processor, a graphics processing unit (Graphics Processing Unit, GPU), RAM Random Access Memory, RAM), ROM (Read- Only Memory, ROM), at least one of the first interface to the nth interface for input/output, a communication bus (Bus), and the like.
CPU处理器。用于执行存储在存储器中操作***和应用程序指令,以及根据接收外部输入的各种交互指令,来执行各种应用程序、数据和内容,以便最终显示和播放各种音视频内容。CPU处理器,可以包括多个处理器。如,包括一个主处理器以及一个或多个子处理器。CPU processor. It is used to execute the operating system and application program instructions stored in the memory, and to execute various application programs, data and content according to various interactive instructions received from the outside, so as to finally display and play various audio and video content. A CPU processor may include multiple processors. For example, including a main processor and one or more sub-processors.
在一些实施例中,用户可在显示器260上显示的图形用户界面(GUI)输入用户命令,则用户输入接口通过图形用户界面(GUI)接收用户输入命令。或者,用户可通过输入特定的声音或手势进行输入用户命令,则用户输入接口通过传感器识别出声音或手势,来接收用户输入命令。In some embodiments, the user can input user commands through a graphical user interface (GUI) displayed on the display 260, and the user input interface receives user input commands through the graphical user interface (GUI). Alternatively, the user may input a user command by inputting a specific sound or gesture, and the user input interface recognizes the sound or gesture through a sensor to receive the user input command.
在一些实施例中,显示设备的***可以包括内核(Kernel)、命令解析器(shell)、文件***和应用程序。In some embodiments, the system of the display device may include a kernel (Kernel), a command parser (shell), a file system and application programs.
参见图4,在一些实施例中,将***分为四层,从上至下分别为应用程序(Applications)层(简称“应用层”),应用程序框架(Application Framework)层(简称“框架层”),***库层(简称“***运行库层”),以及内核层。Referring to Fig. 4, in some embodiments, the system is divided into four layers, from top to bottom are respectively the application (Applications) layer (abbreviated as "application layer"), application framework (Application Framework) layer (abbreviated as "framework layer") "), the system library layer (referred to as the "system runtime layer"), and the kernel layer.
在一些实施例中,应用程序层中运行有至少一个应用程序,这些应用程序可以是操作***自带的窗口(Window)程序、***设置程序或时钟程序等;也可以是第三方开发者所开发的应用程序。在具体实施时,应用程序层中的应用程序包不限于以上举例。In some embodiments, there is at least one application program running in the application program layer, and these application programs can be window (Window) program, system setting program or clock program etc. that come with the operating system; they can also be developed by third-party developers. s application. During specific implementation, the application program packages in the application program layer are not limited to the above examples.
在一些实施例中,内核层是硬件和软件之间的层。如图4所示,内核层至少包含以下驱动中的至少一种:音频驱动、显示驱动、蓝牙驱动、摄像头驱动、WIFI驱动、USB驱动、HDMI驱动、传感器驱动(如指纹传感器,温度传感器,压力传感器等)、以及电源驱动等。In some embodiments, the kernel layer is a layer between hardware and software. As shown in Figure 4, the kernel layer at least includes at least one of the following drivers: audio driver, display driver, bluetooth driver, camera driver, WIFI driver, USB driver, HDMI driver, sensor driver (such as fingerprint sensor, temperature sensor, pressure sensors, etc.), and power drives, etc.
在一些实施例中,以图4中的魔镜应用(拍照应用)为例,当遥控接收装置接收到遥控器输入操作,确认输入操作对应的控件为魔镜应用图标对应的控件,魔镜应用调用应用框架层的接口,启动魔镜应用,进行调用内核层启动摄像头驱动,实现通过摄像头获取图像。In some embodiments, taking the magic mirror application (photographing application) in FIG. Call the interface of the application framework layer, start the magic mirror application, and call the kernel layer to start the camera driver, so as to obtain images through the camera.
当然,显示设备中具有调用摄像头功能的应用不仅仅是拍摄应用,还有可能是视频通话应用,照镜子应用,音视频录制应用,健身应用教学应用等等。以下示例性实施方式中仅以拍摄应用为例进行展开。但须知,本申请中任意具有调用摄像头功能的应用都适用于本申请所要求保护的技术方案。Of course, the applications that have the function of calling the camera in the display device are not only shooting applications, but also video call applications, mirror applications, audio and video recording applications, fitness application teaching applications, and so on. In the following exemplary implementations, only the shooting application is taken as an example for development. However, it should be noted that any application in this application that has the function of calling a camera is applicable to the technical solution claimed in this application.
在一些实施例中,显示设备安装有摄像头后,可拍摄到用户形象如面部和手势动作等,因此,还可具备手势识别功能。例如,在启动拍摄应用中,当用户比出“yeah”相关的手指时候,可触发拍照功能。In some embodiments, after the display device is equipped with a camera, the image of the user, such as the face and gestures, can be captured. Therefore, the gesture recognition function can also be provided. For example, in starting a shooting application, when the user gestures a finger related to "yeah", the camera function may be triggered.
此外,本申请所涉及的显示设备处于运行状态时,在某一物理空间中的位置比较固定,用户所选择的舒适的观看距离较远,或者离得所述显示设备距离较远。以至于当打开摄像头应用拍摄图像时,采集到的人像或者用户面部只占整幅图像的一小部分。加之如果显示设备中有多个窗口在显示器上的同时呈现,导致小窗口中的人像都被进一步缩小,严重影响用户与显示界面上人像互动的乐趣。In addition, when the display device involved in the present application is in a running state, its position in a certain physical space is relatively fixed, and the comfortable viewing distance selected by the user is relatively far away, or the distance from the display device is relatively long. So that when the camera application is opened to take an image, the captured portrait or user's face only accounts for a small part of the entire image. In addition, if multiple windows are presented on the display simultaneously in the display device, the portraits in the small windows will be further reduced, seriously affecting the user's enjoyment of interacting with the portraits on the display interface.
基于此,本申请所涉及的显示设备在调用摄像头功能相关应用时,可启动“人像跟随”功能。具体的,可以通过直接调节摄像头的焦距以使得采集到的人像放大而达到“人像跟随”效果;还可以通过摄像头焦距不调整,仅控制将摄像头采集到图像以识别到的人像或人脸为中心进行放大,在显示器上仅展示摄像头拍摄画面的一部分的形式达到“人像跟随”的效果。Based on this, the display device involved in the present application can activate the "portrait follow" function when invoking the application related to the camera function. Specifically, the "portrait following" effect can be achieved by directly adjusting the focal length of the camera to enlarge the captured portrait; it is also possible to control the image captured by the camera centered on the recognized portrait or face without adjusting the focal length of the camera. Zoom in and display only a part of the camera image on the monitor to achieve the effect of "portrait following".
在一些实施例中,用户利用显示设备200进行拍照时,拍照应用中有摄像头焦距调整功能/或放大拍摄图像功能,以及手势拍照功能,手势拍照功能指的是在摄像头拍摄范围内摆出相应手势能够自动拍照。焦距调整功能指的是能够通过用户执行一定的指令,调整摄像头的焦距,以使得摄像头采集图像的范围改变。放大拍摄图像功能指通过用户执行一定的指令,在摄像头采集图像的范围不变的情况下,放大摄像头拍摄图像在用户界面上的显示范围。In some embodiments, when the user uses the display device 200 to take a picture, the camera application has a function of adjusting the focal length of the camera/or zooming in on the captured image, and a function of taking pictures with gestures. Capable of taking pictures automatically. The focal length adjustment function refers to the ability to adjust the focal length of the camera by executing a certain instruction from the user, so that the range of images collected by the camera can be changed. The function of zooming in on the captured image refers to enlarging the display range of the image captured by the camera on the user interface under the condition that the range of the image captured by the camera remains unchanged by executing a certain instruction from the user.
其中摄像头焦距调整功能能够改变拍摄物在拍摄画面中的大小,同步的,用户界面上拍摄物的显示范围也会随之改变。而放大拍摄图像功能虽然摄像头焦距没有实质改变,但是也可以通过缩放拍摄图像以改变其在用户界面上的显示范围。上述两种方法都可以从用户的直观视觉上使得原来显示范围较小的拍摄物体放大显示。很多情况下,由于相似的用户体验,二者可以相互等同。The focal length adjustment function of the camera can change the size of the subject in the shooting picture, and synchronously, the display range of the subject on the user interface will also change accordingly. Although the function of zooming in on the captured image does not substantially change the focal length of the camera, it can also zoom in on the captured image to change its display range on the user interface. Both of the above two methods can enlarge and display an object whose original display range is small from the user's intuitive vision. In many cases, the two can be equated to each other due to the similar user experience.
还有一些显示设备中虽然相关指令控件记载为焦距调节,但是实质上执行的是放大拍摄图像的动作。这种情况在本申请中,被认为实际触发的就是放大拍摄图像功能。In some display devices, although the relevant instruction controls are recorded as focus adjustment, they actually perform an action of zooming in on the captured image. In this case, in this application, it is considered that what is actually triggered is the function of zooming in on the captured image.
如果用户触发放大拍摄图像功能的话,整个摄像头拍摄区域会分为显示在用户界面上的部分(简称为显示部分)和未在用户界面上显示部分(简称为不显示部分)。用户实际上能观看到的部分为显示部分,这也会被用户直接判定为摄像头拍摄的画面就是显示部分。用户一般也是依据显示部分对摄像进行控制。If the user triggers the function of zooming in and shooting an image, the entire camera shooting area will be divided into a part displayed on the user interface (referred to as the displayed part for short) and a part not displayed on the user interface (referred to as the non-displayed part for short). The part that the user can actually watch is the display part, which will also be directly judged by the user as the image captured by the camera is the display part. The user generally also controls the camera according to the display part.
触发放大拍摄图像功能的方式包括:当开启人像跟随模式时显示部分可以自动地随着用户在摄像头拍摄区间内的移动而动态的移动和缩放。还可以通过用户人为选择缩放图像,则用户界面上的图像也可以根据用户的选择而移动和缩放。The way to trigger the function of zooming in and shooting images includes: when the portrait following mode is turned on, the display part can automatically move and zoom dynamically as the user moves within the camera shooting range. The image can also be zoomed manually through the user's selection, and the image on the user interface can also be moved and zoomed according to the user's selection.
实际上不显示部分也是摄像头采集的画面的一部分,如果进行手势识别时候,控制器一般将整个拍摄的画面全部进行识别,则很有可能出现在显示器上不显示的部分摆出手势也被识别为手势拍照指令而触发拍照。而此时用户没有在用户界面内看到相应的手势,显示设备却给出拍照反馈,影响用户体验。In fact, the part that is not displayed is also a part of the picture captured by the camera. If the controller generally recognizes the entire captured picture when performing gesture recognition, it is very likely that the gesture that appears on the part that is not displayed on the display will also be recognized as The camera is triggered by the gesture camera command. At this time, the user does not see the corresponding gesture in the user interface, but the display device gives feedback on taking pictures, which affects the user experience.
下面对本实施例中显示设备拍照过程进行详细描述。The process of taking pictures by the display device in this embodiment will be described in detail below.
具体地,图5a示出了根据一些实施例的一种显示设备200中显示图像的示例图,如图5a所示,显示器中显示图像,该图像是摄像头拍摄图像经过放大后再呈现在显示器上的图像,需要说明的是,在图像经过放大处理后显示器显示的图像与摄像头拍摄范围对应的图像不同,图5b示出了根据一些实施例的摄像头范围与焦距显示范围对比的示例图,如图5b所示,摄像头拍摄对应图像的拍摄范围大于图5a所示的显示器显示图像对应的拍摄范围,即本申请实施例中所描述的显示器显示范围。Specifically, FIG. 5a shows an example diagram of an image displayed in a display device 200 according to some embodiments. As shown in FIG. 5a, an image is displayed on the display, and the image is an enlarged image captured by a camera and then presented on the display. It should be noted that after the image is enlarged, the image displayed on the display is different from the image corresponding to the shooting range of the camera. FIG. As shown in 5b, the shooting range of the corresponding image captured by the camera is larger than the corresponding shooting range of the display image shown in FIG. 5a, that is, the display display range described in the embodiment of the present application.
在一些情况下焦距显示范围可以等同于实际在显示器上显示的图像范围。In some cases the focal length display range may be equivalent to the actual image range displayed on the monitor.
上述实施例中,用户通过操作控制装置,打开显示设备200安装的拍照应用程序,进而在显示器中显示原始图像,即摄像头拍摄范围对应的图像,接着通过触发放大拍摄图像的指令,进而在显示器中显示被放大的拍摄图像。In the above embodiment, the user opens the camera application program installed on the display device 200 by operating the control device, and then displays the original image on the display, that is, the image corresponding to the shooting range of the camera, and then triggers an instruction to zoom in on the captured image, and then displays the original image on the display. The magnified captured image is displayed.
图6示出了根据一些实施例的一种指令控制方法的流程图。本申请实施例提供的一种显示设备,在根据手势位置和焦距显示范围执行手势指令时,控制器被配置为执行图6所述的指令控制方法,包括:Fig. 6 shows a flowchart of an instruction control method according to some embodiments. In a display device provided in an embodiment of the present application, when executing a gesture instruction according to the gesture position and the focus display range, the controller is configured to execute the instruction control method described in FIG. 6 , including:
响应于放大拍摄图像的指令,将所述图像数据放大,并将至少一部分的所述放大后的图像数据呈现在所述显示器上,如步骤601。In response to an instruction to enlarge a captured image, the image data is enlarged, and at least a part of the enlarged image data is presented on the display, as in step 601 .
在本申请的一些实施例中,触发放大拍摄图像的指令可以是用户通过操作控制装置(比如遥控器)输入,也可以是用户通过显示设备的声音采集器(比如麦克风)等输入,具体根据应用场景选择设置,本申请不作限制。还可以是启动人像跟随功能后,根据人像在拍摄图像中的显示大小触发的自动放大。In some embodiments of the present application, the command to trigger zooming in on an image may be input by the user through an operating control device (such as a remote control), or may be input by the user through a sound collector (such as a microphone) of the display device, depending on the application. The scene selection setting is not limited in this application. It may also be an automatic zoom-in triggered according to the display size of the portrait in the captured image after the portrait following function is activated.
在本申请的一些实施例中,在摄像头启动后,显示器上呈现的初始显示范围为摄像头拍摄范围,响应于放大拍摄图像的指令,将所述拍摄图像进行放大,然后通过裁剪框进行裁剪,所述裁剪框中的图像内容即为显示器上呈现的画面(下文可简称为第二图像)。其中,裁剪框与本申请中显示器中显示窗口的具有一致的宽高比。In some embodiments of the present application, after the camera is started, the initial display range presented on the display is the shooting range of the camera, and in response to an instruction to zoom in on the captured image, the captured image is enlarged and then cropped by a cropping frame, so The image content in the cropping frame is the picture presented on the display (hereinafter referred to as the second image for short). Wherein, the cropping frame has the same aspect ratio as that of the display window in the display in this application.
具体的,根据所述图像数据中目标物的位置,调整所述裁剪框的位置,以使所述裁剪框中包含所述目标物,将调整后的图像画面输出至所述显示器,其中,所述调整后的图像画面为调整后的裁剪框在所述摄像头输入的图像数据对应的图像数据,所述目标物的中心与所述裁剪框的中心距离最小,且所述显示器上不出现黑边。Specifically, according to the position of the target object in the image data, the position of the cropping frame is adjusted so that the target object is included in the cropping frame, and the adjusted image frame is output to the display, wherein the The adjusted image frame is the image data corresponding to the image data of the adjusted cropping frame input by the camera, the distance between the center of the target object and the center of the cropping frame is the smallest, and no black border appears on the display .
在一些实施例中,创建坐标系,获取放大后的图像坐标信息,确定裁剪框在所述放大后的图像中的区域,根据裁剪框的坐标信息更新显示器的显示范围。In some embodiments, a coordinate system is created, the coordinate information of the enlarged image is acquired, the area of the cropping frame in the enlarged image is determined, and the display range of the display is updated according to the coordinate information of the cropping frame.
在一些实施例中,响应于摄像头检测到的手势指令,获取手势矩形区域,步骤602。In some embodiments, a gesture rectangle area is acquired in response to a gesture command detected by a camera, step 602 .
在本申请的一个实施例中,手势指令指的是摄像头检测到特殊手势,比如通过手势识别来完成一系列拍照行为的命令传递方式,手势识别可以理解为用户可以使用简单的手势来控制或与设备交互,摄像头采集图像,检测图像中包括特殊手势,则认为检测到手势指令,其中,用户可以在摄像头拍摄范围内做出不同的手势以向显示设备发送不同指令。In one embodiment of the present application, the gesture command refers to a special gesture detected by the camera, such as a command delivery method for completing a series of photographing actions through gesture recognition. Gesture recognition can be understood as the user can use simple gestures to control or interact with For device interaction, the camera collects images and detects that images include special gestures, and it is considered that a gesture command is detected. Among them, the user can make different gestures within the shooting range of the camera to send different commands to the display device.
具体地,摄像头检测到手势指令,对包括手势指令的图像进行处理,获取手势矩形区域,更具体地,识别到图像区域内手势的所有关键点,根据所有关键点的坐标信息确定最小横纵坐标值和最大横纵坐标值绘制矩形,获取手势矩形区域。Specifically, the camera detects the gesture command, processes the image including the gesture command, obtains the gesture rectangle area, more specifically, recognizes all key points of the gesture in the image area, and determines the minimum horizontal and vertical coordinates according to the coordinate information of all key points value and the maximum horizontal and vertical coordinate values to draw a rectangle to obtain the gesture rectangle area.
在一些实施例中,基于显示器上第二图像的显示和手势矩形区域,如果确定手势矩形区域在用户界面显示范围内,执行手势指令。如步骤603。如果确定手势矩形区域不在用户界面显示范围内,则不执行手势指令,如步骤604。In some embodiments, based on the display of the second image on the display and the gesture rectangle area, if it is determined that the gesture rectangle area is within the display range of the user interface, the gesture command is executed. Such as step 603. If it is determined that the gesture rectangle area is not within the display range of the user interface, the gesture command is not executed, as in step 604 .
在本申请一些实施例中,以用户界面上显示的第二图像显示范围为准,手势矩形区域在第二图像显示范围内执行手势指令,手势矩形区域不在裁剪框显示范围内不执行手势指令,仅执行对于用户可视的拍摄图像的显示范围内的手势指令,提高显示设备控制的有效性和准确性。In some embodiments of the present application, based on the display range of the second image displayed on the user interface, the gesture command is executed in the gesture rectangle area within the display range of the second image, and the gesture command is not executed when the gesture rectangle area is not within the display range of the cropping frame. Only the gesture instructions within the display range of the captured image visible to the user are executed, so as to improve the effectiveness and accuracy of display device control.
在本申请一些实施例中,如图7a所示,手势矩形区域在第二图像显示范围内,执行手势指令,更具体地,手势指令为拍照指令,执行拍照操作,获取图像并存储,进一步地,用户还可以通过控制装置、语音指令或者手势指令等对拍摄图像进行编辑操作,还可以理解的是,用户可以选择重新拍摄,再次通过手势指令执行拍照,满足用户多方面使用需求,提升用户使用体验。In some embodiments of the present application, as shown in FIG. 7a, the gesture rectangle area is within the display range of the second image, and the gesture instruction is executed. More specifically, the gesture instruction is a photographing instruction, and the photographing operation is executed, and the image is acquired and stored, further , the user can also edit the captured image through the control device, voice command or gesture command. experience.
在本申请一些实施例中,如图7b所示,手势矩形区域不在第二图像显示范围内,不 执行手势指令,也就是说,尽管摄像头拍摄范围内获取了手势指令,但是不在用户可视范围内,不执行该手势指令,进一步提升用户体验。In some embodiments of the present application, as shown in Figure 7b, the gesture rectangle area is not within the display range of the second image, and the gesture command is not executed, that is, although the gesture command is acquired within the shooting range of the camera, it is not within the visible range of the user , the gesture command is not executed to further improve the user experience.
由于摄像头拍摄区域具有一定的范围,因此,会有一些采集到的图像不包含完整的目标物的情况。例如图7b中,左上角位置上仅拍摄到一只手,未拍摄到完整的手对应的完成的人物。这种情况下,即便在人脸跟随的情况下也无法捕获到这只手对应的完成的人物。当时当在这种情况下,实际显示范围中包括一个完整的人物,而摄像头拍摄范围包括了这个完整的人物和左上角的手势。因此,由于左上角的手势不在实际显示范围内,不相应对应的手势。Since the shooting area of the camera has a certain range, some collected images may not contain a complete target object. For example, in FIG. 7b, only one hand is photographed at the upper left corner, and the completed character corresponding to the complete hand is not photographed. In this case, the completed character corresponding to this hand cannot be captured even with the face following. In this case, the actual display range includes a complete person, and the camera shooting range includes the complete person and the gesture in the upper left corner. Therefore, since the gesture in the upper left corner is not within the actual display range, the corresponding gesture is not corresponding.
在本申请的一个实施例中,通过第二图像显示范围的坐标信息和手势矩形区域的坐标信息来判断手势矩形区域是否在用户界面上显示范围内,更具体地,以摄像头拍摄范围的左下角为原始坐标点建立坐标轴,摄像头拍摄范围下方边界为横轴,左方边界为纵轴,创建坐标系,获取焦距显示范围的目标坐标信息,获取手势矩形区域的手势坐标信息,基于目标坐标信息和手势坐标信息,确定手势矩形区域在焦距显示范围内执行手势指令、或者是确定手势矩形区域不在用户界面上显示范围内不执行手势指令。In one embodiment of the present application, it is judged whether the gesture rectangle area is within the range displayed on the user interface through the coordinate information of the second image display range and the coordinate information of the gesture rectangle area, more specifically, the lower left corner of the camera shooting range Establish a coordinate axis for the original coordinate point, the lower boundary of the camera shooting range is the horizontal axis, and the left boundary is the vertical axis, create a coordinate system, obtain the target coordinate information of the focal length display range, and obtain the gesture coordinate information of the gesture rectangle area, based on the target coordinate information and the gesture coordinate information, determine that the gesture rectangle area is within the focus display range to execute the gesture command, or determine that the gesture rectangle area is not within the display range on the user interface and do not execute the gesture command.
上述示例中,仅仅执行对于用户可视的用户界面上显示范围内的手势指令,以使用户可以根据看到的手势进行调整和控制操作,进一步提升用户使用体验。In the above example, only the gesture commands within the range displayed on the user interface visible to the user are executed, so that the user can perform adjustment and control operations according to the gestures seen, further improving the user experience.
图8示出了根据一些实施例的另一种指令处理方法的流程图。本实施例在上述实施例的基础上,进一步优化了上述方法。如图8所示,该方法包括:Fig. 8 shows a flowchart of another instruction processing method according to some embodiments. This embodiment further optimizes the foregoing method on the basis of the foregoing embodiments. As shown in Figure 8, the method includes:
在本申请的一些实施例中,判断是否接收到放大所述拍摄图像指令,如步骤801。In some embodiments of the present application, it is determined whether an instruction to enlarge the captured image is received, such as step 801 .
在本申请的一些实施例中,通过控制装置比如遥控器相关按键调整用户界面上对所述拍摄图像的显示范围,在确定放大拍摄图像的情况下,整个摄像头拍摄的图像会分为显示部分和不显示部分,如图7a所示,在用户界面上的显示范围内为显示部分,其他部分为不显示部分。In some embodiments of the present application, the display range of the captured image on the user interface is adjusted through a control device such as a remote control related button. The non-display part, as shown in FIG. 7a, is a display part within the display range on the user interface, and other parts are non-display parts.
在本申请的一些实施例中,摄像头检测到手势指令,执行手势指令,如步骤802。In some embodiments of the present application, the camera detects the gesture instruction, and executes the gesture instruction, such as step 802 .
在本申请一些实施例中,如果摄像头拍摄范围内区域与显示器的第二图像显示范围一致,在摄像头采集到图像,检测图像中包括手势,则认为检测到手势指令,直接执行该手势指令即可,其中,可以根据应用场景需求设置对应的手势指令,比如手势指令A为拍照,手势指令B为存储等,另外,在摄像头范围内未识别到手势指令,则不执行手势指令。In some embodiments of the present application, if the area within the shooting range of the camera is consistent with the display range of the second image of the display, and the image is captured by the camera, and the gesture is detected in the image, it is considered that a gesture command has been detected, and the gesture command can be directly executed , where the corresponding gesture command can be set according to the requirements of the application scenario, for example, gesture command A is to take a picture, gesture command B is to store, etc. In addition, if the gesture command is not recognized within the range of the camera, the gesture command will not be executed.
在本申请的一些实施例中,用户通过操作控制装置(比如遥控器)输入或者用户通过显示设备的声音采集器(比如麦克风)等输入的放大拍摄图像的指令。响应于上述指令,显示设备并不对摄像头进行实质上的焦距调整,也就是说摄像头保持原来的拍摄范围,比如图7a所示,整个摄像头拍摄范围保持不变。而是在获取摄像头采集的图像后,对图像进行放大处理后在显示器中显示,以从用户视觉上,也能够达到放大物体的效果。In some embodiments of the present application, the user inputs an instruction to zoom in on the captured image by operating a control device (such as a remote control) or the user inputs through a sound collector (such as a microphone) of the display device. In response to the above instruction, the display device does not substantially adjust the focal length of the camera, that is to say, the camera maintains the original shooting range. For example, as shown in FIG. 7 a , the entire camera shooting range remains unchanged. Instead, after the image captured by the camera is acquired, the image is enlarged and displayed on the display, so that the effect of zooming in on the object can also be achieved from the user's perspective.
这种情况下,实际摄像头采集的图像范围较大,而显示器仅显示放大后图像的一部分。换句话说,显示器上的实际显示范围显示的是摄像头采集图像经过放大后的一部分,从而显示器的实际显示的图像范围与摄像头拍摄图像范围不一致,比如图7a所示,整个摄像头拍摄区域会分为显示部分和不显示部分,第二图像显示范围内为在用户界面上的实际显示范围,也即为显示部分,其他部分为不显示部分。此时,以第二图像显示范围为准,手势矩形区域在第二图像显示范围内执行手势指令,手势矩形区域不在第二图像显示范围内不执行手势指令。In this case, the actual camera captures a larger image range, while the monitor only displays a part of the enlarged image. In other words, the actual display range on the monitor shows the enlarged part of the image captured by the camera, so the actual range of the displayed image on the monitor is inconsistent with the range of the image captured by the camera. For example, as shown in Figure 7a, the entire camera shooting area will be divided into For the display part and the non-display part, the display range of the second image is the actual display range on the user interface, that is, the display part, and the other parts are non-display parts. At this time, based on the second image display range, the gesture command is executed in the gesture rectangle area within the second image display range, and the gesture command is not executed in the gesture rectangle area not in the second image display range.
这种情况下,虽然有可能用户的手势还在摄像头采集的范围内,但是,如果已经不在显示器显示的范围内,那么,即使摄像头可采集到手势,但是也不识别指令,或者可以识别指令但是不响应手势相对应的指令。In this case, although it is possible that the user's gesture is still within the range captured by the camera, if it is no longer within the range displayed by the display, then even if the gesture can be captured by the camera, the instruction is not recognized, or the instruction can be recognized but Does not respond to commands corresponding to gestures.
因此,用户可以根据直观在显示器上看到的手势进行调整和控制操作,提高显示设备控制的有效性和准确性。Therefore, the user can perform adjustment and control operations according to the gestures intuitively seen on the display, improving the effectiveness and accuracy of display device control.
在本申请的一些实施例中,获取裁剪框坐标信息,基于裁剪框坐标信息更新显示器的第二图像显示范围,如步骤803。In some embodiments of the present application, the coordinate information of the cropping frame is acquired, and the second image display range of the display is updated based on the coordinate information of the cropping frame, such as step 803 .
在本申请的一些实施例中,以摄像头拍摄范围的左下角为原始坐标点(0,0)建立坐标轴,摄像头拍摄范围下方边界为横轴,左方边界为纵轴,创建坐标系,读取放大调整后的裁剪框坐标信息,比如为四个坐标点(x1,y1),(x1,y2),(x2,y1),(x2,y2),从而根据该四个坐标点确定第二图像显示范围。In some embodiments of the present application, the coordinate axis is established with the lower left corner of the shooting range of the camera as the original coordinate point (0, 0), the lower boundary of the shooting range of the camera is the horizontal axis, and the left boundary is the vertical axis, creating a coordinate system, read Get the coordinate information of the cropping frame after zooming in and adjusting, for example, four coordinate points (x1, y1), (x1, y2), (x2, y1), (x2, y2), so as to determine the second Image display range.
在一些实施例中,所述裁剪框仅仅是用于定位被放大的拍摄图像在显示器上的显示范围,以使得所述显示器按照被框选的范围进行显示,并非真实对放大的拍摄图像进行裁剪。In some embodiments, the cropping frame is only used to locate the display range of the enlarged captured image on the display, so that the display will display according to the framed range, rather than actually cropping the enlarged captured image. .
在一些实施例中,控制器直接将裁剪框中的图像呈现在显示器上,而在显示器上,所述裁剪框并不可见。In some embodiments, the controller directly presents the image in the cropping frame on the display, where the cropping frame is not visible.
在本申请的一些实施例中,摄像头检测到手势指令,获取手势指令的多个关键点,获取多个关键点对应的关键点坐标信息确定最小横纵坐标值和最大横纵坐标值,基于最小横纵坐标值和最大横纵坐标值,获取手势矩形区域,如步骤804。In some embodiments of the present application, the camera detects the gesture command, acquires multiple key points of the gesture command, and acquires the key point coordinate information corresponding to the multiple key points to determine the minimum horizontal and vertical coordinate values and the maximum horizontal and vertical coordinate values, based on the minimum The horizontal and vertical coordinate values and the maximum horizontal and vertical coordinate values are used to obtain the gesture rectangle area, as in step 804 .
在本申请实施例中,如图9所示,获取手势指令的多个关键点,并获取多个关键点对应的关键点坐标信息,以摄像头拍摄范围的左下角为原始坐标点(0,0)建立坐标轴,摄像头拍摄范围下方边界为横轴,左方边界为纵轴,创建坐标系,读取关键点坐标信息,确定最小横纵坐标值和最大横纵坐标值,比如为四个坐标点(m1,n1),(m1,n2),(m2,n1),(m2,n2),从而根据该四个坐标点确定手势矩形区域。In the embodiment of the present application, as shown in FIG. 9, a plurality of key points of the gesture command are obtained, and the key point coordinate information corresponding to the plurality of key points is obtained, and the lower left corner of the shooting range of the camera is taken as the original coordinate point (0, 0 ) to establish a coordinate axis, the lower boundary of the camera shooting range is the horizontal axis, and the left boundary is the vertical axis, create a coordinate system, read the coordinate information of key points, and determine the minimum horizontal and vertical coordinate values and maximum horizontal and vertical coordinate values, such as four coordinates Points (m1, n1), (m1, n2), (m2, n1), (m2, n2), so as to determine the gesture rectangle area according to the four coordinate points.
在本申请的一些实施例中,获取第二图像显示范围的目标坐标信息,获取手势矩形区域的手势坐标信息,如步骤805。In some embodiments of the present application, the target coordinate information of the second image display range is acquired, and the gesture coordinate information of the gesture rectangle area is acquired, such as step 805 .
在本申请的一些实施例中,基于目标坐标信息和手势坐标信息,确定手势矩形区域在第二图像显示范围内,执行手势指令,如步骤806。In some embodiments of the present application, based on the target coordinate information and the gesture coordinate information, it is determined that the gesture rectangle area is within the display range of the second image, and the gesture instruction is executed, such as step 806 .
在本申请的一些实施例中,基于目标坐标信息和手势坐标信息,确定手势矩形区域不在第二图像显示范围内,不执行手势指令,如步骤807。In some embodiments of the present application, based on the target coordinate information and the gesture coordinate information, it is determined that the gesture rectangle area is not within the display range of the second image, and the gesture command is not executed, as in step 807 .
在本申请的一些实施例中,获取第二图像显示范围的目标坐标信息比如上述描述的(x1,y1),(x1,y2),(x2,y1),(x2,y2),获取手势矩形区域的手势坐标信息如上述描述的(m1,n1),(m1,n2),(m2,n1),(m2,n2),根据目标坐标信息和手势坐标信息,确定手势矩形区域在第二图像显示范围内,执行手势指令,基于目标坐标信息和手势坐标信息,确定手势矩形区域不在第二图像显示范围内,不执行手势指令。In some embodiments of the present application, the target coordinate information of the second image display range is obtained such as (x1, y1), (x1, y2), (x2, y1), (x2, y2) described above, and the gesture rectangle is obtained The gesture coordinate information of the area is as described above (m1, n1), (m1, n2), (m2, n1), (m2, n2), according to the target coordinate information and gesture coordinate information, determine the gesture rectangle area in the second image Within the display range, the gesture instruction is executed, and based on the target coordinate information and the gesture coordinate information, it is determined that the gesture rectangle area is not within the display range of the second image, and the gesture instruction is not executed.
由此,摄像头拍照界面可通过遥控器等控制设备调整拍照显示范围,当手势在用户界面显示范围内能够拍照,当手势不在焦距显示范围内不可以拍照。Thus, the camera interface can adjust the photographing display range through the remote controller and other control devices. When the gesture is within the display range of the user interface, the photograph can be taken, and when the gesture is not within the focal length display range, the photograph cannot be taken.
在一些实施例中,以图4中的魔镜应用(拍照应用)为例,人工智能试妆是拍照应用上的试妆功能,通过计算出人脸位置,并将妆容描绘到人脸对应位置上,试妆功能内包含有人像跟随效果,或者人脸跟随效果(也可称为人脸跟焦),在摄像头区域内,聚焦人脸位置所在区域,跟焦指的是在拍摄一个镜头时,随着人物或物体趋近或远离摄影机而改变 显示器上对拍摄图像的显示范围,以使其保持在精确的焦点上。In some embodiments, taking the magic mirror application (photographing application) in Figure 4 as an example, the artificial intelligence makeup test is a makeup trial function on the camera application, by calculating the position of the human face, and drawing the makeup to the corresponding position of the human face In terms of makeup, the makeup test function includes a portrait follow effect, or a face follow effect (also known as face follow focus). In the camera area, it focuses on the area where the face is located. Follow focus refers to when shooting a shot, Vary the range of the captured image displayed on the monitor as people or objects move closer to or away from the camera to keep them in precise focus.
由于在用户使用人像跟随功能,当人脸在下方角落时(如图10a所示),使用人体检测位置(0,0,0,0),计算出的矩形缩放区域如图10b所示,调整中心点不准确,导致跟焦效果差。Since the user uses the portrait following function, when the face is in the lower corner (as shown in Figure 10a), the human body detection position (0, 0, 0, 0) is used, and the calculated rectangular zoom area is shown in Figure 10b. Adjust The center point is not accurate, resulting in poor follow-focus effect.
图11示出了根据一些实施例的又一种指令控制方法的流程图。本申请实施例提供的一种显示设备,在进行人像跟随时,控制器被配置为执行图11所述的指令控制方法,包括:Fig. 11 shows a flow chart of yet another instruction control method according to some embodiments. In a display device provided in an embodiment of the present application, when following a portrait, the controller is configured to execute the instruction control method described in FIG. 11 , including:
响应于启动摄像头的控制信号,在显示器中显示摄像头采集的图像,如步骤1101。In response to a control signal for starting the camera, the image captured by the camera is displayed on the display, as in step 1101 .
在本申请一些实施例中,通过遥控器上按键、语音输入、控制面板输入等输入启动摄像头的控制信号,通过摄像头获取图像在显示器中显示。In some embodiments of the present application, a control signal for starting the camera is input through buttons on the remote control, voice input, control panel input, etc., and the images acquired by the camera are displayed on the display.
在本申请一些实施例中,显示器中显示的图像可以包括人脸图像也可以不包括人脸图像,本申请主要针对的是包括人脸图像的处理,人脸图像可以在显示图像的任一位置,比如上述图10所示的图像的右下角落,也可以如图12所示的中心位置,具体根据应用场景中用户所处位置确定。In some embodiments of the present application, the image displayed on the display may or may not include a human face image. This application is mainly aimed at processing including a human face image, and the human face image may be in any position of the displayed image. , for example, the lower right corner of the image shown in FIG. 10 above may also be the center position as shown in FIG. 12 , which is specifically determined according to the location of the user in the application scenario.
响应于指示人脸跟焦的控制信号,基于图像获取人脸位置矩形区域,如步骤1102。In response to the control signal indicating that the face is in focus, acquire a rectangular area of the face position based on the image, as in step 1102 .
在本申请一些实施例中,通过遥控器上按键、语音输入、控制面板输入等输入人脸跟焦的控制信号,人脸区域是指人脸所在的区域,从而对图像进行人脸识别,获取人脸位置矩形区域,识别图像的人脸区域的方式,包括但不限于如下步骤,提取图像中的图像特征,将图像特征输入预先训练的人脸识别模型进行判断,确定人脸识别模型判定为人脸的图像特征对应的位置区域为人脸图像区域,如图12所示的人脸位置矩形区域。In some embodiments of the present application, the face focus control signal is input through buttons on the remote control, voice input, control panel input, etc., and the face area refers to the area where the face is located, so as to perform face recognition on the image and obtain The rectangular area of the face position, the method of recognizing the face area of the image, includes but is not limited to the following steps, extracting the image features in the image, inputting the image features into the pre-trained face recognition model for judgment, and determining that the face recognition model is judged as human The position area corresponding to the image feature of the face is the face image area, such as the rectangular area of the face position as shown in FIG. 12 .
基于人脸位置矩形区域,确定调整中心点,如步骤1103。Determine the adjustment center point based on the rectangular area of the face position, as in step 1103 .
根据调整中心点和预设调整值对图像进行调整,获取目标图像,以及在显示器上显示目标图像,如步骤1104。The image is adjusted according to the adjustment center point and the preset adjustment value, the target image is acquired, and the target image is displayed on the display, such as step 1104 .
在本申请一些实施例中,调整中心点指的是人脸位置矩形区域中一个坐标点,通过人脸位置矩形区域的坐标信息进行计算获取调整中心点,具体地,获取人脸位置矩形区域坐标信息,基于人脸位置矩形区域坐标信息进行计算处理,获取中心点坐标为调整中心点;其中,坐标信息指的是人脸位置矩形区域上的一个或者多个坐标点,以及坐标点对应的坐标值。In some embodiments of the present application, the adjustment center point refers to a coordinate point in the rectangular area of the face position, and the adjustment center point is obtained by calculating the coordinate information of the rectangular area of the face position, specifically, obtaining the coordinates of the rectangular area of the face position The information is calculated and processed based on the coordinate information of the rectangular area of the face position, and the coordinates of the center point are obtained as the adjustment center point; where the coordinate information refers to one or more coordinate points on the rectangular area of the face position, and the coordinates corresponding to the coordinate points value.
在本申请一些实施例中,预设调整值可以根据应用场景需要进行选择设置。In some embodiments of the present application, the preset adjustment value can be selected and set according to the needs of the application scenario.
在本申请一些实施例中,以调整中心点为中心点不动,根据预设调整值对图像进行像素调整,获取目标图像,并在显示器上显示目标图像。继续以图10a为例,计算获取的调整中心点如图10c所示的O点,预设调整数值为一点五倍,对图10c中的图像进行调整,得到的目标图像如图10d所示,保证了跟焦效果。In some embodiments of the present application, the adjustment center point is used as the center point, and the pixels of the image are adjusted according to the preset adjustment value to acquire the target image and display the target image on the display. Continuing to take Figure 10a as an example, the calculated adjustment center point is O as shown in Figure 10c, and the preset adjustment value is 1.5 times, and the image in Figure 10c is adjusted, and the obtained target image is shown in Figure 10d , to ensure the follow-focus effect.
上述示例中,以人脸区域计算调整中心点进行人脸跟焦,保证跟焦效果。In the above example, the face area is used to calculate and adjust the center point to perform face follow focus to ensure the follow focus effect.
图13示出了根据一些实施例的再一种指令控制方法的流程图。本实施例在上述实施例的基础上,进一步优化了上述焦距控制方法。如图11所示,该方法包括:Fig. 13 shows a flow chart of another instruction control method according to some embodiments. In this embodiment, on the basis of the foregoing embodiments, the foregoing focal length control method is further optimized. As shown in Figure 11, the method includes:
在本申请的一些实施例中,在显示器中显示摄像头采集的图像,响应于指示人脸跟焦的控制信号,基于图像获取人脸位置矩形区域,如步骤1301。In some embodiments of the present application, the image captured by the camera is displayed on the display, and in response to a control signal indicating that the face follows focus, a rectangular area of the face position is acquired based on the image, such as step 1301 .
在本申请一些实施例中,显示器中显示的图像可以包括人脸图像也可以不包括人脸图像,本申请主要针对的是包括人脸图像的处理,人脸图像可以在显示图像的任一位置。In some embodiments of the present application, the image displayed on the display may or may not include a human face image. This application is mainly aimed at processing including a human face image, and the human face image may be in any position of the displayed image. .
在本申请一些实施例中,通过遥控器上按键、语音输入、控制面板输入等输入人脸跟 焦的控制信号,从而对图像进行人脸识别,获取人脸位置矩形区域。In some embodiments of the present application, the face focus control signal is input through buttons on the remote control, voice input, control panel input, etc., so as to perform face recognition on the image and obtain the rectangular area of the face position.
在本申请的一些实施例中,在显示器中显示摄像头采集的图像,响应于指示人脸跟焦的控制信号,基于图像获取人脸位置矩形区域之后,可以执行基于图像获取不到人***置矩形区域,基于人脸位置矩形区域,确定调整中心点或基于图像获取人***置矩形区域,基于人脸位置矩形区域和人***置矩形区域确定调整中心点,如步骤1303。In some embodiments of the present application, the image captured by the camera is displayed on the display, and in response to the control signal indicating that the face is in focus, after obtaining the rectangular area of the human face position based on the image, the acquisition of the rectangular area of the human body position based on the image can be performed. Determining the adjustment center point based on the rectangular area of the face position or acquiring the rectangular area of the human body position based on the image, determining the adjustment center point based on the rectangular area of the face position and the rectangular area of the human body position, such as step 1303 .
在本申请的一些实施例中,基于图像获取不到人***置矩形区域,基于人脸位置矩形区域,确定调整中心点,如步骤1302。In some embodiments of the present application, the adjustment center point is determined based on the human face position rectangular area based on the fact that the rectangular area of the human body position cannot be obtained from the image, such as step 1302 .
在本申请一些实施例中,可以理解的是,人***置矩形区域为四肢及躯干部分所在的身体区域,在确定人脸位置矩形区域后,可以采用人体检测算法进行确定,或者建立人体检测模型,进而基于人体检测模型进行检测,获取人***置矩形区域。In some embodiments of the present application, it can be understood that the rectangular area of the human body position is the body area where the limbs and torso are located. After the rectangular area of the human face position is determined, a human body detection algorithm can be used to determine it, or a human body detection model can be established. Furthermore, detection is performed based on the human body detection model, and the rectangular area of the human body position is obtained.
在本申请一些实施例中,基于图像获取不到人***置矩形区域,表示图像仅仅包括人脸位置矩形区域,基于人脸位置矩形区域,确定调整中心点。也就是说,当只能检测到人脸位置矩形区域,未检测到人***置矩形区域时,仅以人脸位置矩形区域作为计算矩形缩放区域的调整中心点。In some embodiments of the present application, based on the fact that the human body position rectangle area cannot be obtained from the image, it means that the image only includes the human face position rectangle area, and the adjustment center point is determined based on the human face position rectangle area. That is to say, when only the rectangular area of human face position can be detected but the rectangular area of human body position is not detected, only the rectangular area of human face position is used as the adjustment center point for calculating the rectangular scaling area.
在本申请一些实施例中,基于人脸位置矩形区域,获取人脸位置矩形区域坐标信息,基于人脸位置矩形区域坐标信息进行计算处理,获取中心点坐标为调整中心点,更具体地,在人脸跟焦功能开启时,检测人脸位置矩形区域和人***置矩形区域,以图像的左下角为原始坐标点(0,0)建立坐标轴,摄像头拍摄范围下方边界为横轴,左方边界为纵轴,创建坐标系,读取人脸位置矩形区域坐标信息,比如为四个坐标点(a1,b1),(a1,b2),(a2,b1),(a2,b2),从而根据该四个坐标点确定人脸位置矩形区域,进一步计算中心点坐标为((a1+a2)/2,(b1+b2)/2)为调整中心点。In some embodiments of the present application, based on the rectangular area of the face position, the coordinate information of the rectangular area of the face position is obtained, and the calculation process is performed based on the coordinate information of the rectangular area of the face position, and the coordinates of the center point are obtained as the adjustment center point. More specifically, in When the face follow focus function is turned on, the rectangular area of the face position and the rectangular area of the human body position are detected, and the coordinate axis is established with the lower left corner of the image as the original coordinate point (0, 0). The lower boundary of the camera shooting range is the horizontal axis, and the left boundary is For the vertical axis, create a coordinate system and read the coordinate information of the rectangular area of the face position, such as four coordinate points (a1, b1), (a1, b2), (a2, b1), (a2, b2), so that according to The four coordinate points determine the rectangular area of the face position, and the coordinates of the center point are further calculated as ((a1+a2)/2, (b1+b2)/2) as the adjustment center point.
在本申请的一些实施例中,基于图像获取人***置矩形区域,基于人脸位置矩形区域和人***置矩形区域确定调整中心点。In some embodiments of the present application, the human body position rectangular area is acquired based on the image, and the adjustment center point is determined based on the human face position rectangular area and the human body position rectangular area.
在本申请一些实施例中,基于人脸位置矩形区域,获取人脸位置矩形区域坐标信息,基于人***置矩形区域,获取人***置矩形区域坐标信息,基于人脸位置矩形区域坐标信息进行计算处理,获取第一中心点坐标,基于人***置矩形区域坐标信息进行计算处理,获取第二中心点坐标,基于第一中心点坐标和第二中心点坐标进行计算处理,获取目标中心点坐标为调整中心点。更具体地,在人脸跟焦功能开启时,检测人脸位置矩形区域和人***置矩形区域,以图像的左下角为原始坐标点(0,0)建立坐标轴,摄像头拍摄范围下方边界为横轴,左方边界为纵轴,创建坐标系,读取人脸位置矩形区域坐标信息和人***置矩形区域坐标信息,比如为四个坐标点(a1,b1),(a1,b2),(a2,b1),(a2,b2),从而根据该四个坐标点确定人脸位置矩形区域,以及四个坐标点(c1,d1),(c1,d2),(c2,d1),(c2,d2),从而根据该四个坐标点确定人***置矩形区域。In some embodiments of the present application, the coordinate information of the rectangular area of the human face position is obtained based on the rectangular area of the human face position, the coordinate information of the rectangular area of the human body position is obtained based on the rectangular area of the human body position, and the calculation process is performed based on the coordinate information of the rectangular area of the human face position, Obtain the coordinates of the first center point, perform calculation processing based on the coordinate information of the rectangular area of the human body position, obtain the coordinates of the second center point, perform calculation processing based on the coordinates of the first center point and the second center point, and obtain the coordinates of the target center point as the adjustment center point . More specifically, when the face focus function is turned on, the rectangular area of the face position and the rectangular area of the human body position are detected, and the coordinate axis is established with the lower left corner of the image as the original coordinate point (0, 0), and the lower boundary of the camera shooting range is horizontal Axis, the left boundary is the vertical axis, create a coordinate system, read the coordinate information of the face position rectangle area and the human body position rectangle area coordinate information, for example, four coordinate points (a1, b1), (a1, b2), (a2 , b1), (a2, b2), so as to determine the rectangular area of the face position according to the four coordinate points, and the four coordinate points (c1, d1), (c1, d2), (c2, d1), (c2, d2), so as to determine the rectangular area of the human body position according to the four coordinate points.
进一步地,根据当(a1,b1,a2,b2),(c1,d1,c2,d2)均不全为零时,判断同时检测到人脸位置矩形区域和人***置矩形区域,二者结合进行双重判断,以人***置矩形区域和人脸位置矩形区域计算调整中心点,即人脸位置矩形区域的第一中心点坐标(h1,h2)为((a1+a2)/2,(b1+b2)/2),人***置矩形区域的第二中心点坐标(k1,k2)为((c1+c2)/2,(d1+d2)/2),根据第一中心点坐标和第二中心点坐标进行计算处理获取目标中心点坐标为((h1+k1)/2,(h2+k2)/2)为调整中心点。Further, according to when (a1, b1, a2, b2), (c1, d1, c2, d2) are not all zero, it is judged that the human face position rectangular area and the human body position rectangular area are detected at the same time, and the two are combined to perform double Judgment, calculate the adjustment center point based on the rectangular area of the human body position and the rectangular area of the face position, that is, the coordinates (h1, h2) of the first center point of the rectangular area of the face position are ((a1+a2)/2, (b1+b2) /2), the second center point coordinates (k1, k2) of the human body position rectangular area are ((c1+c2)/2, (d1+d2)/2), according to the first center point coordinates and the second center point coordinates Perform calculation processing to obtain the coordinates of the target center point as ((h1+k1)/2, (h2+k2)/2) as the adjustment center point.
在本申请的一些实施例中,根据调整中心点和预设调整值对图像进行调整,获取目标 图像,以及在显示器上显示目标图像,如步骤1304。In some embodiments of the present application, the image is adjusted according to the adjustment center point and the preset adjustment value, the target image is acquired, and the target image is displayed on the display, such as step 1304.
在本申请一些实施例中,基于人脸位置矩形区域确定调整中心点,具体地,获取人脸位置矩形区域坐标信息,基于人脸位置矩形区域坐标信息进行计算处理,获取中心点坐标为调整中心点。In some embodiments of the present application, the adjustment center point is determined based on the rectangular area of the face position, specifically, the coordinate information of the rectangular area of the face position is obtained, and the calculation process is performed based on the coordinate information of the rectangular area of the face position, and the coordinates of the center point are obtained as the adjustment center point.
在本申请一些实施例中,预设调整值可以根据应用场景需要进行选择设置。In some embodiments of the present application, the preset adjustment value can be selected and set according to the needs of the application scenario.
在本申请一些实施例中,以调整中心点为中心点不动,根据预设调整值对图像进行像素调整,获取目标图像,并在显示器上显示目标图像。继续以图10a为例,计算获取的调整中心点如图10a所示的O点,预设调整数值为一点五倍,对图10a中的图像进行调整,得到的图像如图10c所示,保证了跟焦效果。继续以图12为例,计算获取的调整中心点如图12所示的O点,预设调整数值为一点五倍,对图12中的图像进行调整,得到的图像如图14所示,保证了跟焦效果。In some embodiments of the present application, the adjustment center point is used as the center point, and the pixels of the image are adjusted according to the preset adjustment value to acquire the target image and display the target image on the display. Continuing to take Figure 10a as an example, the calculated adjustment center point is O as shown in Figure 10a, and the preset adjustment value is 1.5 times, and the image in Figure 10a is adjusted, and the obtained image is shown in Figure 10c, The follow-focus effect is guaranteed. Continuing to take Figure 12 as an example, the calculated adjustment center point is O as shown in Figure 12, and the preset adjustment value is 1.5 times, and the image in Figure 12 is adjusted, and the obtained image is shown in Figure 14. The follow-focus effect is guaranteed.
由此,摄像头在显示器显示界面的人脸跟焦效果,以人脸位置矩形区域和人***置矩形区域为双重检验标准,当检测到人脸位置矩形区域和人***置矩形区域时,以人脸位置矩形区域与人***置矩形区域计算矩形缩放区域,当检测不到人***置矩形区域时,仅以人脸位置矩形区域计算矩形缩放区域。另外,当人***置矩形区域和人脸位置矩形区域都检测不到时,将矩形缩放区域置为空。Therefore, the face focus effect of the camera on the display interface of the monitor is based on the double inspection standard of the face position rectangle area and the body position rectangle area. When the face position rectangle area and the body position rectangle area are detected, the face position The rectangular area and the rectangular area of the human body position calculate the rectangular zoom area. When the rectangular area of the human body position cannot be detected, only the rectangular area of the face position is used to calculate the rectangular zoom area. In addition, when neither the human body position rectangle area nor the face position rectangle area can be detected, the rectangle scaling area is set as empty.
上述示例中,以人脸区域计算调整中心点进行人脸跟焦,保证跟焦效果。In the above example, the face area is used to calculate and adjust the center point to perform face follow focus to ensure the follow focus effect.
为了方便解释,已经结合具体的实施方式进行了上述说明。但是,上述示例性的讨论不是意图穷尽或者将实施方式限定到上述公开的具体形式。根据上述的教导,可以得到多种修改和变形。上述实施方式的选择和描述是为了更好的解释原理以及实际的应用,从而使得本领域技术人员更好的使用所述实施方式以及适于具体使用考虑的各种不同的变形的实施方式。For convenience of explanation, the above description has been made in conjunction with specific implementation manners. However, the above exemplary discussion is not intended to be exhaustive or to limit the implementations to the precise forms disclosed above. Many modifications and variations are possible in light of the above teachings. The selection and description of the above embodiments are to better explain the principles and practical applications, so that those skilled in the art can better use the embodiments and various modified embodiments suitable for specific use considerations.

Claims (10)

  1. 一种显示设备,包括:A display device comprising:
    摄像头,被配置为采集图像数据;a camera configured to collect image data;
    显示器,被配置为呈现用户界面,其中,当打开摄像头相关应用时,呈现所述摄像头采集的图像数据;The display is configured to present a user interface, wherein, when the camera-related application is opened, the image data collected by the camera is presented;
    分别与所述摄像头和所述显示器连接的控制器,所述控制器被配置为:A controller connected to the camera and the display respectively, the controller is configured to:
    响应于打开摄像头相关应用的指令,控制显示器呈现摄像头采集的图像数据;In response to an instruction to open a camera-related application, control the display to present image data collected by the camera;
    响应于用户输入的用于触发放大所述图像数据的指令,将所述图像数据放大,并将至少一部分的所述放大后的图像数据呈现在所述显示器上;Enlarging the image data and presenting at least a portion of the enlarged image data on the display in response to a user-input instruction for triggering enlargement of the image data;
    响应于所述摄像头采集的到预设手势,获取手势矩形区域,以及生成与所述预设手势相对应的手势指令;In response to the preset gesture collected by the camera, acquire a gesture rectangle area, and generate a gesture instruction corresponding to the preset gesture;
    如果所述手势矩形区域在所述显示器呈现的用户界面内,则执行所述手势指令;If the gesture rectangle area is within the user interface presented by the display, execute the gesture instruction;
    如果所述手势矩形区域未呈现在所述用户界面内,则不执行所述手势指令。If the gesture rectangle area is not presented in the user interface, the gesture instruction is not executed.
  2. 根据权利要求1所述的显示设备,所述控制器被进一步配置为:The display device according to claim 1, the controller is further configured to:
    响应于用户输入的用于触发放大所述图像数据的指令,确定所述图像数据需要被放大的倍数;determining a factor by which the image data needs to be enlarged in response to an instruction input by the user for triggering the enlargement of the image data;
    根据所述放大倍数将所述图像数据进行放大。The image data is enlarged according to the magnification factor.
  3. 根据权利要求1所述的显示设备,所述控制器在执行响应于用户输入的用于触发放大所述图像数据的指令,将所述图像数据放大,并将至少一部分的所述放大后的图像数据呈现在所述显示器上,被进一步配置为:The display device according to claim 1, when the controller executes an instruction for triggering the enlargement of the image data in response to a user input, the image data is enlarged, and at least a part of the enlarged image Data is presented on said display, further configured to:
    响应于用户输入的用于触发放大所述图像数据的指令,将所述图像数据放大,zooming in on the image data in response to an instruction input by a user for triggering zooming in on the image data,
    将至少一部分的所述放大后的图像数据呈现在所述显示器上,基于显示器上呈现的图像数据确定焦距显示范围;其中,其余部分的所述放大后的图像数据未显示在所述显示器上。Presenting at least a part of the enlarged image data on the display, and determining a focal length display range based on the image data presented on the display; wherein, the remaining part of the enlarged image data is not displayed on the display.
  4. 根据权利要求1所述的显示设备,所述控制器在执行获取手势矩形区域,被进一步配置为:According to the display device according to claim 1, the controller is further configured to:
    获取所述手势指令的多个关键点;Acquiring multiple key points of the gesture instruction;
    获取所述多个关键点对应的关键点坐标信息;Obtain key point coordinate information corresponding to the plurality of key points;
    基于所述关键点坐标信息,确定最小横纵坐标值和最大横纵坐标值;Based on the key point coordinate information, determine the minimum horizontal and vertical coordinate values and the maximum horizontal and vertical coordinate values;
    基于所述最小横纵坐标值和所述最大横纵坐标值,获取所述手势矩形区域。The gesture rectangle area is acquired based on the minimum horizontal and vertical coordinate values and the maximum horizontal and vertical coordinate values.
  5. 根据权利要求1所述的显示设备,所述控制器在执行基于所述焦距显示范围和所述手势矩形区域,确定所述手势矩形区域在所述焦距显示范围内,执行所述手势指令,被进一步配置为:The display device according to claim 1, when the controller determines that the gesture rectangle area is within the focus display range based on the focus display range and the gesture rectangle area, and executes the gesture command, Further configuration is:
    获取所述焦距显示范围的目标坐标信息;Acquiring target coordinate information of the focal length display range;
    获取所述手势矩形区域的手势坐标信息;Acquiring gesture coordinate information of the gesture rectangle area;
    基于所述目标坐标信息和所述手势坐标信息,确定所述手势矩形区域在所述焦距显示范围内,执行所述手势指令。Based on the target coordinate information and the gesture coordinate information, it is determined that the gesture rectangle area is within the focus display range, and the gesture instruction is executed.
  6. 根据权利要求5所述的显示设备,所述控制器在执行基于所述焦距显示范围和所述手势矩形区域,确定所述手势矩形区域不在所述焦距显示范围内,不执行所述手势指令, 被进一步配置为:The display device according to claim 5, when the controller determines that the gesture rectangle area is not within the focus display range based on the focus display range and the gesture rectangle area, the controller does not execute the gesture instruction, is further configured as:
    基于所述目标坐标信息和所述手势坐标信息,确定所述手势矩形区域不在所述焦距显示范围内,不执行所述手势指令。Based on the target coordinate information and the gesture coordinate information, it is determined that the gesture rectangle area is not within the focus display range, and the gesture instruction is not executed.
  7. 根据权利要求2所述的显示设备,所述控制器在执行响应于用户输入的用于触发放大所述图像数据的指令,确定所述图像数据需要被放大的倍数;根据所述放大倍数将所述图像数据进行放大,具体被配置为:According to the display device according to claim 2, when the controller executes an instruction for triggering magnification of the image data in response to a user input, it determines the factor by which the image data needs to be magnified; according to the magnification factor, the The above image data is enlarged, which is specifically configured as:
    当检测到有人物对象在所述摄像头采集的图像数据中时,When it is detected that there is a human object in the image data collected by the camera,
    响应于用户输入的用于触发放大所述图像数据的指令,确定所述图像数据需要被放大的倍数;determining a factor by which the image data needs to be enlarged in response to an instruction input by the user for triggering the enlargement of the image data;
    根据所述放大倍数将所述图像数据进行放大,以使得所述放大后的图像数据中仍包含所述人物。The image data is enlarged according to the magnification factor, so that the enlarged image data still includes the person.
  8. 根据权利要求2所述的显示设备,所述控制器在执行响应于用户输入的用于触发放大所述图像数据的指令,确定所述图像数据需要被放大的倍数;根据所述放大倍数将所述图像数据进行放大,具体被配置为:According to the display device according to claim 2, when the controller executes an instruction for triggering magnification of the image data in response to a user input, it determines the factor by which the image data needs to be magnified; according to the magnification factor, the The above image data is enlarged, which is specifically configured as:
    当检测到有人物对象在所述摄像头采集的图像数据中时,When it is detected that there is a human object in the image data collected by the camera,
    响应于用户输入的用于触发放大所述图像数据的指令,确定所述图像数据需要被放大的倍数;determining a factor by which the image data needs to be enlarged in response to an instruction input by the user for triggering the enlargement of the image data;
    获取人脸位置矩形区域;Obtain the rectangular area of the face position;
    基于所述人脸位置矩形区域,确定调整中心点;Determine the adjustment center point based on the rectangular area of the human face position;
    根据所述调整中心点和放大倍数对所述图像进行调整,获取目标图像,以及在所述显示器上显示所述目标图像。The image is adjusted according to the adjustment center point and the magnification, a target image is acquired, and the target image is displayed on the display.
  9. 根据权利要求7所述的显示设备,所述控制器在执行所述基于所述人脸位置矩形区域,确定调整中心点,被进一步配置为:According to the display device according to claim 7, the controller is further configured to:
    基于所述人脸位置矩形区域,获取人脸位置矩形区域坐标信息;Based on the rectangular area of the human face position, the coordinate information of the rectangular area of the human face position is obtained;
    基于所述人脸位置矩形区域坐标信息进行计算处理,获取中心点坐标为所述调整中心点。The calculation process is performed based on the coordinate information of the rectangular area of the face position, and the coordinates of the center point are acquired as the adjustment center point.
  10. 一种设备控制方法,包括:A device control method, comprising:
    控制显示器呈现摄像头采集的图像数据;Control the display to present the image data collected by the camera;
    响应于用户输入的用于触发放大所述图像数据的指令,将所述图像数据放大,并将至少一部分的所述放大后的图像数据呈现在所述显示器上,基于显示器上呈现图像数据是获取焦距显示范围;In response to an instruction input by a user for triggering the enlargement of the image data, the image data is enlarged, and at least a part of the enlarged image data is presented on the display, based on the image data presented on the display is acquired focal length display range;
    响应于所述摄像头采集的到预设手势,获取手势矩形区域,以及生成与所述预设手势相对应的手势指令;In response to the preset gesture collected by the camera, acquire a gesture rectangle area, and generate a gesture instruction corresponding to the preset gesture;
    如果所述手势矩形区域在所述显示器呈现的用户界面内,则执行所述手势指令;If the gesture rectangle area is within the user interface presented by the display, execute the gesture instruction;
    如果所述手势矩形区域未呈现在所述用户界面内,则不执行所述手势指令。If the gesture rectangle area is not presented in the user interface, the gesture instruction is not executed.
PCT/CN2022/090560 2021-09-18 2022-04-29 Display device and device control method WO2023040288A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202280058406.8A CN117918057A (en) 2021-09-18 2022-04-29 Display device and device control method

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
CN202111134859.8A CN113747078B (en) 2021-09-18 2021-09-18 Display device and focal length control method
CN202111134859.8 2021-09-18
CN202111104825.4A CN113825002B (en) 2021-09-18 2021-09-18 Display device and focal length control method
CN202111104825.4 2021-09-18

Publications (1)

Publication Number Publication Date
WO2023040288A1 true WO2023040288A1 (en) 2023-03-23

Family

ID=78768107

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/090560 WO2023040288A1 (en) 2021-09-18 2022-04-29 Display device and device control method

Country Status (2)

Country Link
CN (3) CN113747078B (en)
WO (1) WO2023040288A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116204105A (en) * 2023-05-05 2023-06-02 北京睿企信息科技有限公司 Processing system for associated image presentation

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113747078B (en) * 2021-09-18 2023-08-18 海信视像科技股份有限公司 Display device and focal length control method

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110688914A (en) * 2019-09-09 2020-01-14 苏州臻迪智能科技有限公司 Gesture recognition method, intelligent device, storage medium and electronic device
US20200183556A1 (en) * 2017-08-14 2020-06-11 Guohua Liu Interaction position determination method and system, storage medium and smart terminal
CN112565602A (en) * 2020-11-30 2021-03-26 北京地平线信息技术有限公司 Method and apparatus for controlling image photographing apparatus, and computer-readable storage medium
CN112905008A (en) * 2021-01-29 2021-06-04 海信视像科技股份有限公司 Gesture adjustment image display method and display device
CN113747078A (en) * 2021-09-18 2021-12-03 海信视像科技股份有限公司 Display device and focus control method

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10180469B2 (en) * 2015-10-28 2019-01-15 Siemens Healthcare Gmbh Gesture-controlled MR imaging system and method
CN107592458B (en) * 2017-09-18 2020-02-14 维沃移动通信有限公司 Shooting method and mobile terminal

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200183556A1 (en) * 2017-08-14 2020-06-11 Guohua Liu Interaction position determination method and system, storage medium and smart terminal
CN110688914A (en) * 2019-09-09 2020-01-14 苏州臻迪智能科技有限公司 Gesture recognition method, intelligent device, storage medium and electronic device
CN112565602A (en) * 2020-11-30 2021-03-26 北京地平线信息技术有限公司 Method and apparatus for controlling image photographing apparatus, and computer-readable storage medium
CN112905008A (en) * 2021-01-29 2021-06-04 海信视像科技股份有限公司 Gesture adjustment image display method and display device
CN113747078A (en) * 2021-09-18 2021-12-03 海信视像科技股份有限公司 Display device and focus control method
CN113825002A (en) * 2021-09-18 2021-12-21 海信视像科技股份有限公司 Display device and focus control method

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116204105A (en) * 2023-05-05 2023-06-02 北京睿企信息科技有限公司 Processing system for associated image presentation

Also Published As

Publication number Publication date
CN113747078B (en) 2023-08-18
CN113747078A (en) 2021-12-03
CN113825002B (en) 2023-06-06
CN117918057A (en) 2024-04-23
CN113825002A (en) 2021-12-21

Similar Documents

Publication Publication Date Title
US10805522B2 (en) Method of controlling camera of device and device thereof
CN111541845B (en) Image processing method and device and electronic equipment
US11089351B2 (en) Display apparatus and remote operation control apparatus
US20230094025A1 (en) Image processing method and mobile terminal
WO2023040288A1 (en) Display device and device control method
US10516830B2 (en) Guided image composition on mobile devices
WO2022100712A1 (en) Method and system for displaying virtual prop in real environment image, and storage medium
WO2019174628A1 (en) Photographing method and mobile terminal
KR102146858B1 (en) Photographing apparatus and method for making a video
CN111970456B (en) Shooting control method, device, equipment and storage medium
KR20150005270A (en) Method for previewing images captured by electronic device and the electronic device therefor
CN113064684B (en) Virtual reality equipment and VR scene screen capturing method
WO2022037535A1 (en) Display device and camera tracking method
CN112672062B (en) Display device and portrait positioning method
WO2022100262A1 (en) Display device, human body posture detection method, and application
WO2022237839A1 (en) Photographing method and apparatus, and electronic device
WO2022073389A1 (en) Video picture display method and electronic device
WO2022228259A1 (en) Target tracking method and related apparatus
CN112905008B (en) Gesture adjustment image display method and display device
US20230386162A1 (en) Virtual action center based on segmented video feed for a video communication session
WO2022037229A1 (en) Human image positioning methods and display devices
WO2021218473A1 (en) Display method and display device
WO2014117675A1 (en) Information processing method and electronic device
JP2022162409A (en) Electronic apparatus and control method thereof
KR101720607B1 (en) Image photographing apparuatus and operating method thereof

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 202280058406.8

Country of ref document: CN

NENP Non-entry into the national phase

Ref country code: DE