CN112565844A - Video communication method and device and electronic equipment - Google Patents

Video communication method and device and electronic equipment Download PDF

Info

Publication number
CN112565844A
CN112565844A CN202011412842.XA CN202011412842A CN112565844A CN 112565844 A CN112565844 A CN 112565844A CN 202011412842 A CN202011412842 A CN 202011412842A CN 112565844 A CN112565844 A CN 112565844A
Authority
CN
China
Prior art keywords
video communication
video
frame
communication device
input
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011412842.XA
Other languages
Chinese (zh)
Other versions
CN112565844B (en
Inventor
陈喆
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vivo Mobile Communication Co Ltd
Original Assignee
Vivo Mobile Communication Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vivo Mobile Communication Co Ltd filed Critical Vivo Mobile Communication Co Ltd
Priority to CN202011412842.XA priority Critical patent/CN112565844B/en
Publication of CN112565844A publication Critical patent/CN112565844A/en
Application granted granted Critical
Publication of CN112565844B publication Critical patent/CN112565844B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/4104Peripherals receiving signals from specially adapted client devices
    • H04N21/4122Peripherals receiving signals from specially adapted client devices additional display device, e.g. video projector
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B5/00Electrically-operated educational appliances
    • G09B5/08Electrically-operated educational appliances providing for individual presentation of information to a plurality of student stations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • H04N21/44213Monitoring of end-user related data
    • H04N21/44218Detecting physical presence or behaviour of the user, e.g. using sensors to detect if the user is leaving the room or changes his face expression during a TV program
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47205End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for manipulating displayed content, e.g. interacting with MPEG-4 objects, editing locally
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Social Psychology (AREA)
  • Educational Technology (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Business, Economics & Management (AREA)
  • Educational Administration (AREA)
  • Manipulator (AREA)
  • Telephone Function (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The application discloses a video communication method, a video communication device and electronic equipment, belongs to the technical field of communication and aims to solve the problem that a teaching mode is too single in a video teaching process. The method comprises the following steps: receiving a first clear input of an operator during video communication with a second video communication device; responding to the first space input, and displaying a moving track corresponding to the first space input; transmitting the target video frame to the target video communication device; the target video frame is obtained by synthesizing a first track image frame and a second video frame acquired by a second video communication device; the first track image frame comprises the moving track. The method and the device are applied to enriching video teaching mode scenes.

Description

Video communication method and device and electronic equipment
Technical Field
The application belongs to the technical field of communication, and particularly relates to a video communication method, a video communication device and electronic equipment.
Background
As the quality of video communication improves, the application range of video communication is wider and wider, for example, more and more users can use video communication for video teaching.
At present, in the process of video teaching, the video content of a teaching party is generally taken as a main part, and the teaching party and a lecture listening party interact with each other mainly through voice, so that the video teaching is realized.
However, there is a need for a teaching party to instruct a lecturer in real time in a scene where the lecturer wants to read a test paper of the lecturer, for example. At the moment, the teaching party can only inform the teaching party of the reading and amending condition of the test paper through voice, so that the teaching mode in the video teaching process is too single.
Disclosure of Invention
The embodiment of the application aims to provide a video communication method, a video communication device and electronic equipment, and the problem that a teaching mode is too single in a video teaching process can be solved.
In order to solve the technical problem, the present application is implemented as follows:
in a first aspect, an embodiment of the present application provides a video communication method, where the method includes: receiving a first clear input of an operator during video communication with a second video communication device; responding to the first space input, and displaying a moving track corresponding to the first space input; transmitting the target video frame to the target video communication device; the target video frame is obtained by synthesizing a first track image frame and a second video frame acquired by a second video communication device; the first track image frame comprises the moving track.
In a second aspect, an embodiment of the present application provides a video communication apparatus, including: the device comprises a receiving module, a display module and a sending module; the receiving module is used for receiving a first spaced input of the operation body in the process of video communication with the second video communication device; the display module is used for responding to the first space input received by the receiving module and displaying a moving track corresponding to the first space input; a transmitting module for transmitting the target video frame to the target video communication device; the target video frame is obtained by synthesizing a first track image frame and a second video frame acquired by a second video communication device; the first track image frame comprises the moving track.
In a third aspect, an embodiment of the present application provides an electronic device, which includes a processor, a memory, and a program or instructions stored on the memory and executable on the processor, and when executed by the processor, the program or instructions implement the steps of the method according to the first aspect.
In a fourth aspect, embodiments of the present application provide a readable storage medium, on which a program or instructions are stored, which when executed by a processor implement the steps of the method according to the first aspect.
In a fifth aspect, an embodiment of the present application provides a chip, where the chip includes a processor and a communication interface, where the communication interface is coupled to the processor, and the processor is configured to execute a program or instructions to implement the method according to the first aspect.
In the embodiment of the present application, first, in the process of video communication between the first video communication apparatus and the second video communication apparatus, after the first video communication apparatus receives a first gap input of the operation body, the first video communication apparatus may display a movement trajectory corresponding to the first gap input. The first video communication device may then transmit the target video frame to the target video communication device. The target video frame is obtained by synthesizing a first track image frame and a second video frame acquired by a second video communication device; the first track image frame comprises the moving track. Through the scheme, compared with a scheme that interaction can be carried out only through voice in the related technology, the video teaching method can be enriched through an image interaction mode. Taking the example that the lecturer who uses the electronic device a is the lecturer who uses the electronic device B to perform the examination paper reviewing, the first video communication device can acquire the examination paper image acquired by the electronic device B and acquire the reviewing trace image acquired by the electronic device a. Then, the first video communication means may transmit the test paper image with the reading trace obtained by combining the image of the reading trace and the test paper image to the electronic device B. Finally, the first video communication device can send the test paper image with the reading and amending traces to the electronic equipment B, so that the lecturer using the electronic equipment B can visually see the reading and amending conditions of the test paper after receiving the test paper image with the reading and amending traces, thereby enriching the teaching mode of video teaching.
Drawings
Fig. 1 is a schematic flowchart of a video communication method according to an embodiment of the present application;
fig. 2 is a schematic interface diagram of an application of a video communication method according to an embodiment of the present disclosure;
fig. 3 is a second schematic interface diagram of an application of a video communication method according to an embodiment of the present application;
fig. 4 is a third schematic interface diagram of an application of a video communication method according to an embodiment of the present application;
fig. 5 is a schematic diagram of an application of a video communication method according to an embodiment of the present application;
fig. 6 is a second schematic diagram illustrating an application of a video communication method according to an embodiment of the present application;
fig. 7 is a fourth schematic interface diagram of an application of a video communication method according to an embodiment of the present application;
fig. 8 is a fifth schematic interface diagram of an application of a video communication method according to an embodiment of the present application;
fig. 9 is a third schematic diagram illustrating an application of a video communication method according to an embodiment of the present application;
fig. 10 is a fourth schematic diagram illustrating an application of a video communication method according to an embodiment of the present application;
fig. 11 is a fifth schematic diagram illustrating an application of a video communication method according to an embodiment of the present application;
fig. 12 is a schematic structural diagram of a video communication apparatus according to an embodiment of the present application;
fig. 13 is a schematic structural diagram of an electronic device according to an embodiment of the present disclosure;
fig. 14 is a second schematic structural diagram of an electronic device according to an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some, but not all, embodiments of the present application. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
The terms first, second and the like in the description and in the claims of the present application are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the application are capable of operation in sequences other than those illustrated or described herein. The objects distinguished by "first", "second", and the like are usually a class, and the number of the objects is not limited, and for example, the first object may be one or a plurality of objects. In addition, "and/or" in the specification and claims means at least one of connected objects, a character "/" generally means that a preceding and succeeding related objects are in an "or" relationship.
The video communication method provided by the embodiment of the present application is described in detail below with reference to the accompanying drawings through specific embodiments and application scenarios thereof.
In the following embodiments of the present application, a video communication method will be exemplarily described by taking an execution subject as a first video communication apparatus as an example. Fig. 1 is a schematic flow chart of a video communication method according to an embodiment of the present application, including steps 201 to 203:
step 201: the first video communication device receives a first clear input of an operator during video communication between the first video communication device and a second video communication device.
In the embodiment of the present application, the operation body may be a user (for example, a finger of the user), an article (for example, a pencil, a pen, an eraser, a stylus, or the like), or any operable object, which is not limited in the embodiment of the present application.
In the embodiment of the present application, the first spaced input is an input in which the operating body is in the air and is not in contact with the first video communication device.
It should be noted that the camera of the first video communication device acquires the moving process of the operation body, that is, the first video communication device receives the first spaced input of the operation body.
In this application embodiment, the camera in this application can be leading camera, also can be rearmounted camera, still can be leading camera and rearmounted camera, and concrete can be set for according to actual need, and this application embodiment does not limit this.
Illustratively, the first spaced-apart input described above may be a specific spaced-apart gesture of the operator input.
The specific space-isolated gesture in the embodiment of the application may be any one of a space-isolated single-click gesture, a space-isolated sliding gesture, a space-isolated dragging gesture, a space-isolated long-press gesture, a space-isolated area change gesture, a space-isolated double-press gesture, a space-isolated double-click gesture and a space-isolated click gesture of any number of times.
Step 202: and responding to the first space input, and displaying a moving track corresponding to the first space input by the first video communication device.
In the embodiment of the present application, the movement track may be a movement track of the operation body. It is to be understood that the movement trajectory may be a virtual movement trajectory, that is, the movement trajectory is the same trajectory as the operation trajectory generated by identifying the operation trajectory of the operation body in the air.
For example, the movement trajectory may be determined by the first video communication device according to the captured image of the operation body. For example, when the user draws a √ shape in the air with a finger, the first video communication device can capture a multi-frame image in the process that the user draws the √ shape, so that the first video communication device can generate a virtual trajectory having a trajectory of the √ shape from the trajectory of the user's finger in the multi-frame image.
In an example, the specific generation process of the movement trajectory may be as follows: the first video communication device may generate the moving trajectory according to a position of the operating body in the acquired image. For example, taking the operating body as a stylus as an example, the previous and next 2 frames of images collected by the first video communication device are respectively the previous frame of image 1 and the next frame of image 2, wherein the pen point of the stylus is displayed at the a coordinate point in the image 1, and the pen point of the stylus is displayed at the B coordinate point in the image 2. Then, the first video communication apparatus may superimpose the image 1 and the image 2, and connect the a-coordinate point and the B-coordinate point to generate the trajectory 1, resulting in an image 2a including the trajectory 1. After the next frame of image 3 of the image 1 and the image 2 collected by the first video communication device, where the pen point of the stylus is displayed at the C coordinate point in the image 3, the first video communication device may superimpose the image 2a and the image 3, and connect the B coordinate point and the C coordinate point to generate the trajectory 2, so as to obtain the image 3a including the trajectory 1 and the trajectory 2, and so on, the first video communication device may generate the movement trajectory of the pen point of the stylus.
Step 203: the first video communication device transmits the target video frame to the target video communication device.
The target video frame is obtained by synthesizing a first track image frame and a second video frame acquired by a second video communication device; the first track image frame comprises the moving track.
In this embodiment, the target video communication device may include the second video communication device, or may not include the second video communication device, which is not limited in this embodiment.
Accordingly, in the case where the target video communication apparatus includes the second video communication apparatus, the second video communication apparatus may receive and display the above-described target video frame from the first video communication apparatus.
In this embodiment, the first track image frame may include one or more first sub-image frames, which is not limited in this embodiment. The second video frame may include one or more second sub-video frames, which is not limited in this embodiment.
In this embodiment of the present application, the first track image frame may be captured by a camera of the first video communication device. The first track image frame may include a real image (for example, an image of a finger of a user) or a virtual image (i.e., the movement track described above), which is not limited in this embodiment of the present application.
The real image is a real picture captured by a camera of the first video communication device, for example, both hands of the user, a touch pen used by the user, and the like.
In this embodiment, the second video frame may be captured by a camera of the second video communication device. For example, the second video frame may be a test paper image, a drawing image, a calligraphy image, or the like, which is not limited in this embodiment of the application.
In this embodiment of the present application, the second video frame may be received by the first video communication device or downloaded by the first video communication device, which is not limited in this embodiment of the present application.
It should be noted that the first track image frame may include the operation body. For example, if the operating body is a user finger, the first track image frame may include the user finger and a movement track traced by the user finger.
It should be noted that the first video communication device and the second video communication device may be Augmented Reality (AR) devices, and the first track image frame and the second video frame may be AR images.
Alternatively, in this embodiment of the present application, before step 201, the first video communication apparatus may extract a partial image containing the operation body. For example, taking the operating body as a stylus pen as an example, as shown in (a) of fig. 2, an image 31 captured by a camera of the mobile phone 1 is displayed on the screen of the mobile phone 1, and the image 31 includes two hands of the user 1 and a stylus pen 33 held by a right hand 32 of the user 1. At this time, the mobile phone 1 may extract only the image 34 of the stylus 33 and the right hand 32 of the user 1 from the image 31, and as shown in (b) of fig. 2, the mobile phone 1 may display the image 34 of the stylus 33 and the right hand 32 of the user 1.
For example, the first video communication device may perform image synthesis with the first track image frame as a foreground image and the second video frame as a background image to obtain a target video frame.
It is to be understood that the above-mentioned composition of the first track image frame and the second video frame means that the first video communication device may display the first track image frame on the second video frame in an overlapping manner, or perform image fusion on the first track image frame and the second video frame.
For example, when the user 1 using the mobile phone 1 wants to review the test paper image of the user 2 using the mobile phone 2 in real time, as shown in (a) of fig. 3, the mobile phone 1 may display the test paper image 41 sent by the mobile phone 2 to the mobile phone 1 in the screen, which includes 5-channel choice questions. Then, in conjunction with (b) in fig. 2, as shown in (b) in fig. 3, the mobile phone 1 may display an image of the stylus pen 42 and the right hand 43 of the user 1 on the test paper image 41, and when the user 1 determines that the answer to the choice question 1 is correct, the user 1 may draw a "√" on the choice question 1 with the stylus pen 42. Next, the mobile phone 1 can obtain the reading and review image with the movement locus "check mark". Finally, the mobile phone 1 may superimpose and display the reading and amending image on the test paper image 41 to obtain the test paper image (i.e. the above-mentioned target video frame) with the reading and amending image as the foreground layer and the test paper image 41 as the background layer, and as shown in fig. 4, the mobile phone 1 may display the test paper image 44 with the reading and amending trace "√" in real time.
Optionally, in this embodiment of the application, the first video communication device may specifically perform image synthesis on the first track image frame and the second video frame in a manner as possibly implemented as follows.
Illustratively, the target video frame may include M target sub-video frames, the first track image frame may include M first sub-image frames, and the second video frame may include M second sub-video frames. Each frame of target sub-video frame is obtained by superposing and displaying a frame of first sub-video frame on a frame of second sub-video frame, wherein the frame of first sub-video frame corresponds to the frame of second sub-video frame. The first sub-image frame of each frame is synthesized by a current video frame acquired by the first video communication device and a synthesized frame, wherein the synthesized frame is synthesized by an X frame video frame acquired by the first video communication device before the acquired current video frame.
Wherein M is a positive integer, and X is an integer less than M.
For example, referring to (b) in fig. 3, firstly, in the process of drawing a "√" on the first choice question by the user 1 using the stylus 43, the mobile phone 1 may obtain 3 frames of sub reading images, which are the sub reading image 1, the sub reading image 2 and the sub reading image 3 respectively. Meanwhile, the mobile phone 1 can obtain 3 frames of sub-test paper images sent to the mobile phone 1 by the mobile phone 2, namely the sub-test paper image 1, the sub-test paper image 2 and the sub-test paper image 3. Secondly, the mobile phone 1 can display the sub reading image 1 as a foreground layer in the sub test paper image 1 in an overlapping manner. Furthermore, the mobile phone 1 may generate the track image 1 according to the sub reading image 1 and the sub reading image 2, and superimpose and display the track image 1 as a foreground layer in the sub test paper image 2. Then, the mobile phone 1 can generate a track image 2 according to the track image 1 and the sub review image 3, superimpose and display the track image 2 in the sub test paper image 3 as a foreground layer, and send the track image 2 to the mobile phone 2. Finally, as shown in fig. 4, the mobile phone 1 can display the test paper image 44 having the marking of reading and approving "√" in real time, and the mobile phone 2 can also display the test paper image 44 having the marking of reading and approving "√" in real time.
Optionally, in this embodiment of the application, the user may trigger the video communication device to be in a mode of acquiring the track image according to a requirement.
Illustratively, before the first video communication device receives the first spaced input of the operation body in the above step 201, the method may further include the following steps 201a and 201 b:
step 201 a: the first video communication device receives a seventh input from the user.
Illustratively, the seventh input may specifically include: the specific gesture input by the user may be determined according to the actual use requirement, and this is not limited in the embodiment of the present application.
The specific gesture in the embodiment of the application may be any one of a single-click gesture, a sliding gesture, a dragging gesture, a pressure recognition gesture, a long-press gesture, an area change gesture, a double-press gesture, a double-click gesture and a click gesture of any number of times; the click input in the embodiment of the application can be click input, double-click input, click input of any number of times and the like, and can also be long-time press input or short-time press input.
In an example, if the first video communication device is AR glasses, the seventh input may be an input that a user touches a lens of the AR glasses. That is, an input by the user touching the AR glasses lens is associated with triggering the first video communication device to be in the first mode described below.
In another example, if the first video communication device is a mobile phone, the seventh input may be a blank input of double-clicking a finger of the user in the air, or may be an input of long-pressing a screen of the mobile phone by the user.
Step 201 b: in response to the seventh input, the first video communication device controls the first video communication device to be in the first mode.
The first mode is a mode for acquiring a track image.
For example, in the case of a video call between the user 1 using the AR glasses 1 and the user 2 using the AR glasses 2, when the user 1 wants to book the test paper image of the user 2 in real time, as shown in fig. 5, the user 1 may touch the lenses of the AR glasses 1 with fingers from the outside, and at this time, the AR glasses 1 controls the AR glasses 1 to be in the book mode (i.e., the first mode described above).
It is to be appreciated that with the first video communication device in the first mode, the first video communication device may capture an image and generate a track image frame from the captured image. And in the case that the first video communication device is not in the first mode, the first video communication device may capture an image but may not generate a track image frame from the captured image.
It should be noted that, when the first video communication apparatus is in the first mode, the first video communication apparatus may automatically switch the camera to the rear camera to capture an image. It will be appreciated that the first video communication device may remain deployed with a rear facing camera if the first video communication device is the rear facing camera that is deployed.
In an example, the step 201b may specifically include the following steps: in response to the seventh input, the first video communication device controls the first video communication device to be in the first mode and transmits the first indication information to the second video communication device. Wherein the first indication information is used for indicating that the second video communication device is in the third mode.
Accordingly, the second video communication apparatus may receive the above-described first indication information.
It should be noted that, during the process of video communication between the first video communication device and the second video communication device, when the second video communication device is in the third mode, the second video communication device may capture an image and send the image to the first video communication device in real time, but when displaying the image, the second video communication device does not directly display the captured image, but displays the image transmitted by the first video communication device to the second video communication device.
In the video communication method provided by the embodiment of the application, first, in a process of video communication between a first video communication device and a second video communication device, after the first video communication device receives a first idle input of an operation body, the first video communication device may display a movement track corresponding to the first idle input. The first video communication device may then transmit the target video frame to the target video communication device. The target video frame is obtained by synthesizing a first track image frame and a second video frame acquired by a second video communication device; the first track image frame comprises the moving track. Through the scheme, compared with a scheme that interaction can be carried out only through voice in the related technology, the video teaching method can be enriched through an image interaction mode. Taking the example that the lecturer who uses the electronic device a is the lecturer who uses the electronic device B to perform the examination paper reviewing, the first video communication device can acquire the examination paper image acquired by the electronic device B and acquire the reviewing trace image acquired by the electronic device a. Then, the first video communication means may transmit the test paper image with the reading trace obtained by combining the image of the reading trace and the test paper image to the electronic device B. Finally, the first video communication device can send the test paper image with the reading and amending traces to the electronic equipment B, so that the lecturer using the electronic equipment B can visually see the reading and amending conditions of the test paper after receiving the test paper image with the reading and amending traces, thereby enriching the teaching mode of video teaching.
Optionally, in this embodiment of the application, the first video communication device may adjust a display position of the operation body, so as to ensure accuracy of a movement trajectory of the display.
In an example, in the case that the movement trajectory is a movement trajectory of an operation body, before the receiving of the first spaced input of the operation body in the step 201, the method may further include the following steps 201c to 201 e:
step 201 c: in the case where the first video communication apparatus displays a third video frame, the first video communication apparatus displays the first flag in the third video frame.
The third video frame is a video frame collected by the first video communication device; the display position of the first mark is used for indicating the real-time position of the operation body; the third video frame includes the operation body.
Illustratively, the first video communication device may recognize the operation body in the third video frame through the image recognition model, and display the first identifier at the position of the operation body in the third video frame.
In one example, the first video communication device may display a preset handwriting and prompt a user to control the stylus to move the operation body along with the preset handwriting, and then the first video communication device may recognize a feature of the operation body according to a movement condition of the operation body and locate the operation body. For example, taking the first video communication device as AR glasses as an example, with reference to fig. 5, after the user 1 wears the AR glasses 1 and the user 1 triggers the AR glasses 1 to be in the reading and closing mode, when the user 1 wants to position the stylus, as shown in (a) in fig. 6, the user may hold the stylus with the right hand and contact the pen tip of the stylus with the desktop 51. After the AR glasses 1 (i.e., 52 in (a) in fig. 6) capture the right hand of the user and the stylus held by the right hand of the user, as shown in (b) in fig. 6, the AR glasses 1 may display one curved trace 54 on the virtual screen 53 and display the text "move the pen tip following the trace". Then, the user may control the pen tip of the stylus 55 to move on the desktop along with the curved writing 54, so that the AR glasses may recognize the stylus 55 and locate the pen tip position of the stylus 55.
For example, the marks in the embodiments of the present application are used to indicate words, symbols, images, and the like of information, and a control or other container may be used as a carrier for displaying information, including but not limited to a word mark, a symbol mark, and an image mark.
For example, the first mark may be in any shape, for example, the first mark may be in a circle, or in a rectangle, or in a square, which may be set according to actual requirements, and this is not limited in this embodiment of the application.
In one example, the first identifier may be a recognition box.
Step 201 d: the first video communication device receives a second clear input to the first identification by the first video communication device user.
For example, the first spaced-apart input may be a specific spaced-apart gesture of the first video communication device user to the first identification input.
Step 201 e: in response to the second clear input, the first video communication device determines a display position of the first identifier.
It should be noted that, after the video communication device determines the display position of the first identifier, that is, after the video communication device completes the recognition and positioning of the stylus, the video communication device may display a dot at the recognized pen tip position of the stylus, where the dot may move along with the movement of the pen tip.
For example, taking the operating body as a stylus as an example, and referring to fig. 2 (b), when the image 34 (i.e., the third video frame) is displayed on the mobile phone 1, the mobile phone 1 can automatically recognize the pen point of the stylus 33 in the image 34. At this time, as shown in fig. 7 (a), the cellular phone 1 may display a circle recognition frame 61 in the image 34. When the user wants to adjust the tip position of the stylus pen 33, the user can drag the circular recognition box 61 with a finger to the position that wants to be adjusted (i.e., the second space input described above). At this time, as shown in fig. 7 (b), the mobile phone 1 displays the pen tip of the stylus pen 33 at the center position of the circular recognition frame 61, and specifies the display position of the circular recognition frame 61.
The video communication method provided by the embodiment of the application can be applied to a scene of positioning the operation body, a user can visually check the real-time display position of the operation body through the first identification displayed by the first video communication device, meanwhile, the user can quickly adjust the real-time display position of the operation body through the input of the first identification, the user can quickly position the operation body, and the accuracy of a follow-up generated moving track can be further improved.
Further optionally, in this embodiment of the application, the video communication device may prompt the user to confirm the position of the stylus.
For example, the displaying the first identifier in the third video frame in step 201c may specifically include the following step a 1:
step A1: the first video communication device outputs the first information and displays the first mark in the third image.
The first information is used for prompting a user to confirm whether the display position of the first identifier is correct or not.
Illustratively, the first information may include at least one of: text, picture, voice and key, which are not limited in the embodiment of the present application.
Based on the step a1, the step 201e may specifically include the following steps B1 to B3:
step B1: and responding to the second space input, and adjusting the display position of the first identifier by the first video communication device.
Step B2: the first video communication device receives an eighth clear input of the first information from the user.
For example, the above-mentioned eighth space-between input may be a specific space-between gesture of the operator on the first information input.
Step B3: in response to the eighth space input, the first video communication apparatus determines a display position of the first mark.
For example, referring to fig. 7 (a), the mobile phone 1 may display a circular recognition box 61 in the image 34, and also display the text "please select the pen tip position and identify and position the pen tip" and the "confirm" button 62 (i.e. the first information). When the user wants to adjust the pen tip position of the stylus pen 33, the user may drag the circular recognition box 61 to the position that wants to be adjusted (i.e., the second space input described above) with space. At this time, as shown in fig. 7 (b), the mobile phone 1 displays the pen tip of the stylus pen 33 at the center of the circle of the circular recognition frame 61. If the user wants to confirm that the circular recognition box 61 is the position of the pen tip of the stylus pen 33, the user can click the "confirm" button 62 (i.e., the eighth space input described above), and at this time, the mobile phone 1 can specify the display position of the circular recognition box 61.
The video communication method provided by the embodiment of the application can be applied to a scene for prompting a user to determine the position of the operation body, the user can determine the real-time display position of the operation body through inputting the first information displayed by the first video communication device, and the situation that the user triggers the first video communication device to determine the display position of the operation body after adjusting the first identifier is avoided, so that the process of adjusting the display position of the operation body is more flexible.
Optionally, in this embodiment of the present application, the first video communication device may generate the first track image frame according to different track parameters.
For example, in the case that the movement trajectory is a movement trajectory of the operation body, before the step 203, the method may further include the following steps 203a to 203 c:
step 203 a: the first video communication device identifies the movement track of the operation body.
For example, the first video communication device may recognize the movement trajectory according to a position of the operator in the captured image.
For example, taking the operating body as a stylus as an example, the first video communication device captures front and back 2 frames of images, namely an image 1 and an image 2, wherein the position a in the image 1 shows the pen point of the stylus, and the position B in the image 2 shows the pen point of the stylus, and the first video communication device can recognize the movement track of the pen point of the stylus according to the position a and the position B.
Step 203 b: the first video communication device obtains the trajectory parameters.
Wherein the trajectory parameters include at least one of: track color, track thickness. It should be noted that the trajectory parameters include, but are not limited to, the two aforementioned parameters, which can be specifically set according to actual requirements, and this is not limited in this embodiment of the application.
For example, the trajectory parameter may be set by default in the system or may be set by a user, which is not limited in this embodiment of the application.
The track color may be any color, such as red, black, gray, or color, for example, and the embodiment of the present application is not limited thereto.
For example, the video communication apparatus may set different trajectory parameters according to different scenes. For example, in a scenario where user 1 reviews user 2's test paper, the video communication device may set the track color to red; alternatively, in a scene where the user 1 instructs the user 2 on the calligraphy, the video communication apparatus may set the track color to black.
Step 203 c: the first video communication device generates the first trajectory image frame based on the recognized movement trajectory of the operation body in accordance with the trajectory parameter.
The video communication method provided by the embodiment of the application can be applied to a scene for generating the first track image frame, and the first video communication device can generate different track images based on the identified moving track according to different track parameters, so that the flexibility of generating the track images can be improved.
Further optionally, in this embodiment of the application, the user may manually adjust the target parameter.
Illustratively, before the step 203b, the method may further include the following steps C1 to C3:
step C1: and under the condition that the first video communication device displays the target adjusting control and the target adjusting control indicates the first parameter, the first video communication device receives a third spaced input of the operating body.
The third space input is a movement input of the first video communication device user control operation body on the target adjustment control.
For example, the target adjustment control may include a plurality of sub adjustment controls, one sub adjustment control for adjusting one trajectory parameter.
For example, the target adjustment control may be an adjustment progress bar or a selection area, which is not limited in this embodiment of the application. The adjusting progress bar comprises a sliding rail and a sliding block, the sliding block can slide on the sliding rail along with the movement of the operating body, and the parameter to be adjusted can change along with the sliding of the sliding block.
For example, the first parameter may be set by default in the system, or may be set by a user, which is not limited in this embodiment of the application.
Illustratively, the first parameter described above may include at least one of: track color, track thickness. It should be noted that the first parameter includes, but is not limited to, the foregoing two parameters, which can be specifically set according to actual requirements, and this is not limited in this embodiment of the application.
Step C2: in response to the third clear input, the first video communication device updates the first parameter to the trajectory parameter.
In an example, the trajectory parameter may be greater than the first parameter, or less than the first parameter, or equal to the first parameter, which is not limited in this application.
Step C3: the first video communication device receives a fourth spaced-apart input of the operator.
The fourth space input is input when the first video communication device user controls the operation body to stay on the target adjustment control for a target duration which is greater than or equal to a first preset threshold.
For example, the first preset threshold may be set by default in the system, or may be set by a user, which is not limited in this embodiment of the application.
Based on the step C3, the step 203b may specifically include the following step C4:
step C4: in response to the fourth spaced-apart input, the first video communication device obtains the trajectory parameter.
For example, referring to (b) in fig. 2, as shown in fig. 8, a color adjustment control 72 and a thickness adjustment control 73 are displayed in a screen 71 of the mobile phone 1, where the color adjustment control 72 includes 4 color options, which are a red option 74, a black option, a yellow option, and a blue option. When the user wants to set the color of the trace to red, the user can control the tip of the stylus 33 to stay on the red option 74 for 1 second; when the user wants to set the track thickness, the user may control the pen tip of the stylus 33 to drag the slider in the thickness adjusting control 73 (i.e., the third space input), and at this time, the mobile phone 1 may display the corresponding track thickness preview image in real time. After the user determines the thickness of the trajectory line, the user may control the pen point of the stylus 33 to stay on the thickness adjustment control 73 for 1 second, and at this time, the mobile phone 1 may obtain the trajectory color and the trajectory thickness set by the user.
The video communication method provided by the embodiment of the application can be applied to a parameter scene for setting the track image, the first video communication device can assist a user to quickly and flexibly adjust and set the display parameters of the track in a mode of displaying the target adjusting control, the track size and the color of the user-defined moving track are realized, and the efficiency and the flexibility for the user to set the track parameters are improved.
Optionally, in this embodiment of the application, the displaying of the movement track corresponding to the first space input in step 202 may specifically include the following step 202 a:
step 202 a: and under the condition that the distance between the operating body and the first video communication device is greater than or equal to a second preset threshold value, the first video communication device displays the moving track.
For example, the distance between the operating body and the first video communication device may be determined according to the depth camera.
For example, the second preset threshold may be set by default in the system, or may be set by a user, which is not limited in this embodiment of the application.
For example, referring to fig. 8, a depth adjustment control 75 is further displayed in the screen of the mobile phone 1, when the user wants to set a preset distance (i.e., the second preset threshold), the user may control the pen tip of the stylus 33 to drag a slider in the depth adjustment control 75, and after the user determines the preset distance, the user may control the pen tip of the stylus 33 to stay on the depth adjustment control 75 for 1 second, so that the mobile phone 1 determines the preset distance.
In an example, taking the operating body as a stylus pen as an example, in a case that a distance between a pen tip of the stylus pen and the first video communication device is greater than or equal to a second preset threshold, the first video communication device may display handwriting on the screen, otherwise, the user moves the stylus pen, and only the hand and the pen of the user are displayed on the screen, and no handwriting occurs. For example, the user only wants to point out a certain position in the test paper and does not want to leave the reading trace, the user only needs to bring the pen tip of the stylus close to the mobile phone, at this time, the mobile phone determines that the distance between the pen tip of the stylus and the mobile phone is smaller than the preset distance (namely, the second preset threshold), and the mobile phone only displays the hand and the pen of the user and does not display the handwriting.
The first video communication device may display the movement trajectory when the distance between the operation body and the first video communication device is greater than or equal to a second preset threshold, or may display the movement trajectory when the distance between the operation body and the first video communication device is within a preset interval. In one example, the preset interval may include the second preset threshold.
The video communication method provided by the embodiment of the application can be applied to a scene that the first video communication device is triggered to display the moving track, and the first video communication device can display the moving track only under the condition that the distance between the operation body and the first video communication device meets the second preset threshold, so that the first video communication device can be triggered to display or not display the track image more flexibly.
Optionally, in this embodiment of the application, a user may trigger the first video communication device and the second video communication device to exchange corresponding modes according to a requirement.
Illustratively, the method may further include the steps 204a and 204b of:
step 204 a: the first video communication device receives a fifth clear input of the operator while the first video communication device is in the first mode.
Illustratively, the first spaced-apart input described above may be a specific spaced-apart gesture of the operator input.
Step 204 b: in response to the fifth blanking input, the first video communication device updates the first mode to the second mode.
The first mode is a mode for acquiring a track image; the second mode is a mode for displaying a fourth video frame, and the fourth video frame is a video frame obtained by synthesizing a fifth track image frame and a sixth video frame; the fifth track image frame is an image frame collected by the second video communication device; the sixth video frame is a video frame captured by the first video communication device.
For example, when the mobile phone 1 is in the first mode, the user 1 using the mobile phone 1 may press the switch key (i.e. the fifth blank input) with a blank length when the user 1 wants the user 2 to make a set of test paper for testing in real time. At this time, the mobile phone 1 may update the first mode to the second mode, and send an indication message to the mobile phone 2 to indicate that the mobile phone 2 is switched to the first mode. Then, the mobile phone 1 can collect the test paper image and send the test paper image to the mobile phone 2, the mobile phone 2 can collect the hand and pen image after receiving the test paper image sent by the mobile phone 1, and the handwriting and the test paper image transmitted by the mobile phone 1 are generated and synthesized and then transmitted to the mobile phone 1, so that remote virtual answering can be realized.
The video communication method provided by the real-time example can be applied to the scene of switching the mode of the electronic equipment, and a user can quickly switch the mode of the electronic equipment, so that the functions of the electronic equipment are quickly switched, and the flexibility of video communication is improved.
Optionally, in this embodiment of the application, before receiving the first spaced input of the operation body in step 201 described above, the method may further include step 205 of:
step 205: in the case where the first video communication device is in video communication with at least two third video communication devices, the first video communication device determines the second video communication device.
Wherein the target video communication device is at least one of the at least two third video communication devices.
Illustratively, the above-mentioned second video communication apparatus is one of the above-mentioned at least two third video communication apparatuses.
In one example, the second video communication device described above may be system default.
In another example, the second video communication apparatus described above may be set for a user.
Illustratively, the step 205 may specifically include the following steps 205a to 205 c:
step 205 a: in the case where the first video communication device is in video communication with at least two third video communication devices, the first video communication device displays N identifiers.
Wherein one identifier corresponds to one third video communication device, and N is a positive integer.
Step 205 b: the first video communication device receives a sixth spaced input of the operator to a target identifier of the N identifiers.
For example, the sixth spaced-apart input described above may be a specific spaced-apart gesture of the operator on the target identification input.
Step 205 c: in response to the sixth blanking input, the first video communication device determines the third video communication device corresponding to the target identifier as the second video communication device.
For example, as shown in (a) of fig. 9, when the first video communication apparatus is the AR device 1, and the AR device 1, the AR device 2, and the AR device 3 perform video teaching, an icon may be displayed in the upper right corner of the virtual screen 81 of the AR device 1, and the image is the lecturer icon 82. If the user using the AR device 1 wants to perform image teaching with the image captured by the AR device 2 as a background image, the user of the AR device 1 can click on the lecturer icon 82 with the stylus 83. At this time, as shown in (b) of fig. 9, the virtual screen 81 of the AR device 1 may spread to display 2 lecturer identifiers, which are a lecturer identifier 84 of the user of the AR device 2 and a lecturer identifier 85 of the user of the AR device 3, respectively. Then, the user using the AR device 1 may click on the lecturer identifier 84 of the user of the AR device 2 using the stylus pen 83 (i.e., the above-described sixth spaced input), at which time the AR device 1 determines that the AR device 2 is the second video communication apparatus.
In an example, the number of the lecturers can be displayed in real time in the above-mentioned lecturer icon 82, for example, if there are 2 lecturers, the number 2 can be displayed on the lecturer icon 82; if there are 3 participants, the number 3 can be displayed on the participant icon 82.
The video communication method provided by the embodiment of the application can be applied to a scene that the second video communication device is determined from a plurality of electronic devices, the first video communication device can assist a user in rapidly determining the second video communication device in a display identification mode, and the video communication efficiency is improved.
Optionally, in this embodiment of the present application, after the step 205 described above, the method may further include the step 206:
step 206: the video communication apparatus transmits the request information to the second video communication apparatus.
The request information is used for requesting the second video communication device to send the second video frame to the first video communication device.
Accordingly, the second video communication apparatus may receive the above-described request message.
For example, the second electronic device may transmit the second video frame to the first video communication apparatus after receiving the blanking input for the request message.
Illustratively, the above-mentioned blank input may specifically include: the operation body identifies the specific blank gesture input to the target.
For example, referring to (b) in fig. 9, after the AR device 1 determines that the AR device 2 is the second video communication apparatus, the AR device 1 may send a request message to the AR device 2, as shown in fig. 10, after the AR device 2 receives the request message, 2 options and a central point 92 of the AR device 2, which are a "start transmission" option 93 and a "reject transmission" option 94, respectively, may be displayed on a virtual screen 91 of the AR device 2. If the user using the AR device 2 turns his head, moves the central point 92 of the AR device 2 onto the "start transmission" option 93, i.e., selects the "start transmission" option 93, the AR device 2 starts to transmit the image captured by the AR device 2 (i.e., the second video frame described above) to the AR device 1; if the user using AR device 2 turns his head, moves the center point 92 of AR device 2 over the "reject transmission" option 94, i.e., selects the "reject transmission" option 94, then AR device 2 will not transmit an image to AR device 1.
It should be noted that the first video communication device and the second video communication device may be fixed to a fixing device, for example, as shown in fig. 11, a mobile phone or a tablet may be fixed to a work table with a bracket.
For example, the above-mentioned working table may be a table, a white paper or a blackboard, etc., which is not limited in this embodiment of the application.
In one example, in the case where the first video communication device is fixed on a table as shown in fig. 11, the second preset threshold may be a distance from the camera when the pen tip of the stylus pen contacts the table.
It should be noted that, in the video communication method provided in the embodiment of the present application, the execution subject may be a video communication device, or a control module in the video communication device for executing the video communication method. In the embodiment of the present application, a video communication apparatus executing a video communication method is taken as an example, and the video communication apparatus provided in the embodiment of the present application is described.
Fig. 12 is a schematic diagram of a possible structure of a video communication apparatus for implementing the embodiment of the present application, and as shown in fig. 12, the video communication apparatus 900 includes: a receiving module 901, a displaying module 902 and a sending module 903, wherein: a receiving module 901, configured to receive a first spaced input of an operation body during video communication with a second video communication apparatus; a display module 902, configured to display, in response to the first space-time input received by the receiving module 901, a movement trajectory corresponding to the first space-time input; a sending module 903, configured to send a target video frame to a target video communication device; the target video frame is obtained by synthesizing a first track image frame and a second video frame acquired by a second video communication device; the first track image frame comprises the moving track.
Optionally, the target video frame includes M target sub-video frames, the first track image frame includes M first sub-image frames, the second video frame includes M second sub-video frames, and M is a positive integer; each frame of target sub-video frame is obtained by superposing and displaying a frame of first sub-video frame on a frame of second sub-video frame, wherein the frame of first sub-video frame corresponds to the frame of second sub-video frame; the first sub-image frame of each frame is synthesized by a current video frame acquired by the first video communication device and a synthesized frame, the synthesized frame is synthesized by an X frame video frame acquired by the first video communication device before the acquired current video frame, and X is an integer less than M.
Optionally, the video communication apparatus 900 further comprises: a determination module 904; the moving track is the moving track of the operation body; a display module 902, further configured to, in a case that a third video frame is displayed, display the first identifier in the third video frame; a receiving module 901, further configured to receive a second space-between input of the first identifier by the first video communication device user; a determining module 904, configured to determine a display position of the first identifier in response to the second space input received by the receiving module 901; the third video frame is a video frame collected by the first video communication device; the display position of the first mark is used for indicating the real-time position of the operation body; the third video frame includes an operator.
Optionally, the video communication apparatus 900 further comprises: an identification module 905, an acquisition module 906 and a generation module 907; the moving track is the moving track of the operation body; an identifying module 905, configured to identify a moving trajectory of the operation body; an obtaining module 906, configured to obtain a trajectory parameter; a generating module 907, configured to generate a first track image frame based on the moving track of the operation body identified by the identifying module according to the track parameter acquired by the acquiring module; wherein the trajectory parameter comprises at least one of: track color, track thickness.
Optionally, the video communication apparatus 900 further comprises: an update module 908; the receiving module 901 is further configured to receive a third blank input of the operating body when the target adjustment control is displayed and indicates the first parameter, where the third blank input is a movement input of the first video communication apparatus user controlling the operating body on the target adjustment control; an updating module 908, configured to update the first parameter to a trajectory parameter in response to the third spaced input received by the receiving module 901; the receiving module 901 is further configured to receive a fourth blank input of the operation body, where the fourth blank input is an input that a user of the first video communication device controls the operation body to stay on the target adjustment control for a target duration that is greater than or equal to a first preset threshold; the obtaining module 906 is specifically configured to obtain the trajectory parameter in response to the fourth space input received by the receiving module 901.
Optionally, the display module 902 is specifically configured to display the moving track when a distance between the operation body and the first video communication device is greater than or equal to a second preset threshold.
Optionally, the video communication apparatus 900 further comprises: an update module 908; a receiving module 901, further configured to receive a fifth spaced input of an operating body when the first video communication apparatus is in the first mode; an updating module 908, configured to update the first mode to the second mode in response to the fifth spaced input received by the receiving module 901; the first mode is a mode for acquiring a track image; the second mode is a mode for displaying a fourth video frame, and the fourth video frame is a video frame obtained by synthesizing a fifth track image frame and a sixth video frame; the fifth track image frame is an image frame collected by the second video communication device; the sixth video frame is a video frame captured by the first video communication device.
Optionally, the video communication apparatus 900 further comprises: a determination module 904; a display module 902, further configured to display N identifiers in case of video communication with at least two third video communication apparatuses, where one identifier corresponds to one third video communication apparatus, and N is a positive integer; a receiving module 901, further configured to receive a sixth spaced input of an operator to a target identifier in the N identifiers; a determining module 904, configured to determine, in response to the sixth spacing input received by the receiving module 901, a third video communication device corresponding to the target identifier as the second video communication device.
In the video communication apparatus provided in the embodiment of the present application, first, in a process of video communication between the first video communication apparatus and the second video communication apparatus, after the first video communication apparatus receives a first idle input of the operation body, the first video communication apparatus may display a movement trajectory corresponding to the first idle input. The first video communication device may then transmit the target video frame to the target video communication device. The target video frame is obtained by synthesizing a first track image frame and a second video frame acquired by a second video communication device; the first track image frame comprises the moving track. Through the scheme, compared with a scheme that interaction can be carried out only through voice in the related technology, the video teaching method can be enriched through an image interaction mode. Taking the example that the lecturer who uses the electronic device a is the lecturer who uses the electronic device B to perform the examination paper reviewing, the first video communication device can acquire the examination paper image acquired by the electronic device B and acquire the reviewing trace image acquired by the electronic device a. Then, the first video communication means may transmit the test paper image with the reading trace obtained by combining the image of the reading trace and the test paper image to the electronic device B. Finally, the first video communication device can send the test paper image with the reading and amending traces to the electronic equipment B, so that the lecturer using the electronic equipment B can visually see the reading and amending conditions of the test paper after receiving the test paper image with the reading and amending traces, thereby enriching the teaching mode of video teaching.
The video communication apparatus in the embodiment of the present application may be an apparatus, or may be a component, an integrated circuit, or a chip in a terminal. The device can be mobile electronic equipment or non-mobile electronic equipment. Illustratively, the mobile electronic device may be a mobile phone, a tablet computer, a notebook computer, a palm top computer, a vehicle-mounted electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook or a Personal Digital Assistant (PDA), and the like, and the non-mobile electronic device may be a server, a Network Attached Storage (NAS), a Personal Computer (PC), a Television (TV), a teller machine, a self-service machine, and the like, and the embodiments of the present application are not particularly limited.
The video communication apparatus in the embodiment of the present application may be an apparatus having an operating system. The operating system may be an Android (Android) operating system, an ios operating system, or other possible operating systems, and embodiments of the present application are not limited specifically.
The video communication apparatus provided in the embodiment of the present application can implement each process implemented by the method embodiments in fig. 1 to fig. 10, and is not described herein again to avoid repetition.
Optionally, as shown in fig. 13, an electronic device 1100 is further provided in an embodiment of the present application, and includes a processor 1101, a memory 1102, and a program or an instruction stored in the memory 1102 and executable on the processor 1101, where the program or the instruction is executed by the processor 1101 to implement each process of the video communication method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
It should be noted that the electronic devices in the embodiments of the present application include the mobile electronic devices and the non-mobile electronic devices described above.
Fig. 14 is a schematic hardware structure diagram of an electronic device implementing an embodiment of the present application.
The electronic device 100 includes, but is not limited to: a radio frequency unit 101, a network module 102, an audio output unit 103, an input unit 104, a sensor 105, a display unit 106, a user input unit 107, an interface unit 108, a memory 109, and a processor 110.
Those skilled in the art will appreciate that the electronic device 100 may further comprise a power source (e.g., a battery) for supplying power to various components, and the power source may be logically connected to the processor 110 through a power management system, so as to implement functions of managing charging, discharging, and power consumption through the power management system. The electronic device structure shown in fig. 14 does not constitute a limitation of the electronic device, and the electronic device may include more or less components than those shown, or combine some components, or arrange different components, and thus, the description is not repeated here.
The input unit 104 is used for receiving a first spacing input of an operation body in the process of video communication with a second video communication device; the display unit 106 is configured to display a movement track corresponding to the first space-time input in response to the first space-time input received by the input unit 104; a radio frequency unit 101, configured to send a target video frame to a target video communication device; the target video frame is obtained by synthesizing a first track image frame and a second video frame acquired by a second video communication device; the first track image frame comprises the moving track.
Optionally, the target video frame includes M target sub-video frames, the first track image frame includes M first sub-image frames, the second video frame includes M second sub-video frames, and M is a positive integer; each frame of target sub-video frame is obtained by superposing and displaying a frame of first sub-video frame on a frame of second sub-video frame, wherein the frame of first sub-video frame corresponds to the frame of second sub-video frame; the first sub-image frame of each frame is synthesized by a current video frame acquired by the first video communication device and a synthesized frame, the synthesized frame is synthesized by an X frame video frame acquired by the first video communication device before the acquired current video frame, and X is an integer less than M.
Optionally, the movement track is a movement track of the operation body; the display unit 106 is further configured to display the first identifier in a third video frame if the third video frame is displayed; an input unit 104, further configured to receive a second clear input of the first identification by the first video communication device user; a processor 110 for determining a display position of the first identifier in response to the second spaced input received by the input unit 104; the third video frame is a video frame collected by the first video communication device; the display position of the first mark is used for indicating the real-time position of the operation body; the third video frame includes an operator.
Optionally, the movement track is a movement track of the operation body; a processor 110 for recognizing a movement trajectory of an operation body; and is used for obtaining the track parameters; the first track image frame is generated based on the identified moving track of the operation body according to the track parameters; wherein the trajectory parameter comprises at least one of: track color, track thickness.
Optionally, the input unit 104 is further configured to receive a third blank input of the operation body when the target adjustment control is displayed and indicates the first parameter, where the third blank input is a movement input of the first video communication apparatus user controlling the operation body on the target adjustment control; a processor 110, configured to update the first parameter to a trajectory parameter in response to a third spaced input received by the input unit 104; the input unit 104 is further configured to receive a fourth blank input of the operation body, where the fourth blank input is an input that a target duration for the first video communication device user to control the operation body to stay on the target adjustment control is greater than or equal to a first preset threshold; the processor 110 is specifically configured to obtain the trajectory parameter in response to the fourth space input received by the input unit 104.
Optionally, the display unit 106 is specifically configured to display the movement track when a distance between the operation body and the first video communication device is greater than or equal to a second preset threshold.
Optionally, the input unit 104 is further configured to receive a fifth spaced-apart input of the operation body if the first video communication apparatus is in the first mode; a processor 110, configured to update the first mode to the second mode in response to a fifth blanking input received by the input unit 104; the first mode is a mode for acquiring a track image; the second mode is a mode for displaying a fourth video frame, and the fourth video frame is a video frame obtained by synthesizing a fifth track image frame and a sixth video frame; the fifth track image frame is an image frame collected by the second video communication device; the sixth video frame is a video frame captured by the first video communication device.
Optionally, the display unit 106 is further configured to display N identifiers in the case of video communication with at least two third video communication apparatuses, where one identifier corresponds to one third video communication apparatus, and N is a positive integer; the input unit 104 is further configured to receive a sixth space input of the operation body to a target identifier of the N identifiers; and a processor 110, configured to determine, in response to the sixth blanking input received by the input unit 104, a third video communication device corresponding to the target identifier as the second video communication device.
In the electronic device provided by the embodiment of the application, first, in a process of video communication between the electronic device and the second video communication device, after the electronic device receives a first space input of an operation body, the electronic device may display a movement track corresponding to the first space input. The electronic device may then transmit the target video frame to the target video communication device. The target video frame is obtained by synthesizing a first track image frame and a second video frame acquired by a second video communication device; the first track image frame comprises the moving track. Through the scheme, compared with a scheme that interaction can be carried out only through voice in the related technology, the video teaching method can be enriched through an image interaction mode. Taking the example that the lecturer who uses the electronic device a is the lecturer who uses the electronic device B to perform the examination paper reviewing, the electronic device a may acquire the examination paper image acquired by the electronic device B, and acquire the reviewing trace image acquired by the electronic device a. Then, the electronic apparatus a can synthesize the image of the reading trace and the image of the test paper to obtain the image of the test paper with the reading trace, and send the image of the test paper with the reading trace to the electronic apparatus B. Finally, the electronic equipment A can send the test paper image with the reading trace to the electronic equipment B, so that a lecturer using the electronic equipment B can visually see the reading condition of the test paper after receiving the test paper image with the reading trace, and the teaching mode of video teaching is enriched.
It should be understood that, in the embodiment of the present application, the input Unit 104 may include a Graphics Processing Unit (GPU) 1041 and a microphone 1042, and the Graphics Processing Unit 1041 processes image data of a still picture or a video obtained by an image capturing device (such as a camera) in a video capturing mode or an image capturing mode. The display unit 106 may include a display panel 1061, and the display panel 1061 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like. The user input unit 107 includes a touch panel 1071 and other input devices 1072. The touch panel 1071 is also referred to as a touch screen. The touch panel 1071 may include two parts of a touch detection device and a touch controller. Other input devices 1072 may include, but are not limited to, a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, a mouse, and a joystick, which are not described in detail herein. The memory 109 may be used to store software programs as well as various data including, but not limited to, application programs and an operating system. The processor 110 may integrate an application processor, which primarily handles operating systems, user interfaces, applications, etc., and a modem processor, which primarily handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into the processor 110.
The embodiment of the present application further provides a readable storage medium, where a program or an instruction is stored on the readable storage medium, and when the program or the instruction is executed by a processor, the program or the instruction implements each process of the video communication method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
The processor is the processor in the electronic device described in the above embodiment. The readable storage medium includes a computer readable storage medium, such as a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and so on.
The embodiment of the present application further provides a chip, where the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is configured to run a program or an instruction to implement each process of the above video communication method embodiment, and can achieve the same technical effect, and the details are not repeated here to avoid repetition.
It should be understood that the chips mentioned in the embodiments of the present application may also be referred to as system-on-chip, system-on-chip or system-on-chip, etc.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element. Further, it should be noted that the scope of the methods and apparatus of the embodiments of the present application is not limited to performing the functions in the order illustrated or discussed, but may include performing the functions in a substantially simultaneous manner or in a reverse order based on the functions involved, e.g., the methods described may be performed in an order different than that described, and various steps may be added, omitted, or combined. In addition, features described with reference to certain examples may be combined in other examples.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solutions of the present application may be embodied in the form of a software product, which is stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal (such as a mobile phone, a computer, a server, an air conditioner, or a network device) to execute the method according to the embodiments of the present application.
While the present embodiments have been described with reference to the accompanying drawings, it is to be understood that the invention is not limited to the precise embodiments described above, which are meant to be illustrative and not restrictive, and that various changes may be made therein by those skilled in the art without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (10)

1. A video communication method applied to a first video communication apparatus, the method comprising:
receiving a first clear input of an operator during video communication with a second video communication device;
responding to the first space input, and displaying a movement track corresponding to the first space input;
transmitting the target video frame to the target video communication device;
the target video frame is synthesized by a first track image frame and a second video frame collected by the second video communication device; the first track image frame includes the movement track.
2. The method of claim 1, wherein the target video frame comprises M target sub-video frames, the first track image frame comprises M first sub-image frames, the second video frame comprises M second sub-video frames, M being a positive integer;
each frame of target sub-video frame is obtained by superposing and displaying a frame of first sub-video frame on a frame of second sub-video frame, wherein the frame of first sub-video frame corresponds to the frame of second sub-video frame;
each frame of the first sub-image frame is synthesized by a current video frame acquired by the first video communication device and a synthesized frame, wherein the synthesized frame is synthesized by an X frame of video frames acquired by the first video communication device before the acquired current video frame, and X is an integer less than M.
3. The method according to claim 1, wherein the movement trajectory is a movement trajectory of an operation body;
prior to receiving the first spaced-apart input of the operational body, the method further comprises:
if a third video frame is displayed, displaying a first identifier in the third video frame;
receiving a second clear input to the first identification by a first video communication device user;
determining a display position of the first marker in response to the second clear input;
the third video frame is a video frame collected by the first video communication device; the display position of the first mark is used for indicating the real-time position of the operation body; the third video frame comprises the operation body.
4. The method according to claim 1, wherein the movement trajectory is a movement trajectory of an operation body;
prior to transmitting the target video frame to a target video communication device, the method further comprises:
identifying a moving track of the operation body;
acquiring a track parameter;
generating the first track image frame based on the recognized moving track of the operation body according to the track parameters;
wherein the trajectory parameters include at least one of: track color, track thickness.
5. The method of claim 4, wherein prior to obtaining trajectory parameters, the method further comprises:
receiving a third blank input of an operation body under the condition that a target adjusting control is displayed and indicates a first parameter, wherein the third blank input is a movement input of a first video communication device user for controlling the operation body on the target adjusting control;
updating the first parameter to the trajectory parameter in response to the third spaced input;
receiving a fourth air-space input of an operation body, wherein the fourth air-space input is an input of a first video communication device user controlling the operation body to stay on the target adjustment control for a target time length which is greater than or equal to a first preset threshold value;
the acquiring of the track parameters comprises:
in response to the fourth spaced-apart input, obtaining the trajectory parameter.
6. The method of claim 1, wherein displaying a movement trajectory corresponding to the first spaced input comprises:
and displaying the movement track when the distance between the operation body and the first video communication device is larger than or equal to a second preset threshold value.
7. The method of claim 1, further comprising:
receiving a fifth clear input of an operator while the first video communication device is in the first mode;
updating the first mode to a second mode in response to the fifth blanking input;
the first mode is a mode for acquiring a track image; the second mode is a mode for displaying a fourth video frame, and the fourth video frame is a video frame obtained by synthesizing a fifth track image frame and a sixth video frame; the fifth track image frame is an image frame collected by the second video communication device; the sixth video frame is a video frame acquired by the first video communication device.
8. The method of claim 1, wherein prior to receiving the first spaced-apart input of the operational body, the method further comprises:
displaying N identifiers under the condition of video communication with the at least two third video communication devices, wherein one identifier corresponds to one third video communication device, and N is a positive integer;
receiving a sixth space input of an operator to a target identifier in the N identifiers;
in response to the sixth blanking input, determining a third video communication device corresponding to the target identification as the second video communication device.
9. A video communication apparatus, characterized in that the video communication apparatus comprises: the device comprises a receiving module, a display module and a sending module;
the receiving module is used for receiving a first spaced input of an operation body in the process of video communication with a second video communication device;
the display module is used for responding to the first space-above input received by the receiving module and displaying a moving track corresponding to the first space-above input;
the sending module is used for sending the target video frame to the target video communication device;
the target video frame is synthesized by a first track image frame and a second video frame collected by the second video communication device; the first track image frame includes the movement track.
10. An electronic device comprising a processor, a memory, and a program or instructions stored on the memory and executable on the processor, the program or instructions when executed by the processor implementing the steps of the video communication method according to any one of claims 1 to 8.
CN202011412842.XA 2020-12-04 2020-12-04 Video communication method and device and electronic equipment Active CN112565844B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011412842.XA CN112565844B (en) 2020-12-04 2020-12-04 Video communication method and device and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011412842.XA CN112565844B (en) 2020-12-04 2020-12-04 Video communication method and device and electronic equipment

Publications (2)

Publication Number Publication Date
CN112565844A true CN112565844A (en) 2021-03-26
CN112565844B CN112565844B (en) 2023-05-12

Family

ID=75048963

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011412842.XA Active CN112565844B (en) 2020-12-04 2020-12-04 Video communication method and device and electronic equipment

Country Status (1)

Country Link
CN (1) CN112565844B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114025237A (en) * 2021-12-02 2022-02-08 维沃移动通信有限公司 Video generation method and device and electronic equipment

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102075727A (en) * 2010-12-30 2011-05-25 中兴通讯股份有限公司 Method and device for processing images in videophone
CN104092958A (en) * 2014-07-01 2014-10-08 广东威创视讯科技股份有限公司 Video signal labeling method, system and device
CN106339094A (en) * 2016-09-05 2017-01-18 山东万腾电子科技有限公司 Interactive remote expert cooperation maintenance system and method based on augmented reality technology
CN106354266A (en) * 2016-09-22 2017-01-25 北京小米移动软件有限公司 Control method and device of terminal as well as terminal
CN106454199A (en) * 2016-10-31 2017-02-22 维沃移动通信有限公司 Video communication method and mobile terminal
CN206712945U (en) * 2017-04-26 2017-12-05 联想新视界(天津)科技有限公司 Video communications system
WO2018119632A1 (en) * 2016-12-27 2018-07-05 深圳市大疆创新科技有限公司 Image processing method, device and equipment
CN108932053A (en) * 2018-05-21 2018-12-04 腾讯科技(深圳)有限公司 Drawing practice, device, storage medium and computer equipment based on gesture
CN110233841A (en) * 2019-06-11 2019-09-13 上海文景信息科技有限公司 A kind of long-distance education data interaction system and method based on AR holographic glasses
CN111614922A (en) * 2019-02-22 2020-09-01 ***通信有限公司研究院 Information interaction method, network terminal and terminal

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102075727A (en) * 2010-12-30 2011-05-25 中兴通讯股份有限公司 Method and device for processing images in videophone
CN104092958A (en) * 2014-07-01 2014-10-08 广东威创视讯科技股份有限公司 Video signal labeling method, system and device
CN106339094A (en) * 2016-09-05 2017-01-18 山东万腾电子科技有限公司 Interactive remote expert cooperation maintenance system and method based on augmented reality technology
CN106354266A (en) * 2016-09-22 2017-01-25 北京小米移动软件有限公司 Control method and device of terminal as well as terminal
CN106454199A (en) * 2016-10-31 2017-02-22 维沃移动通信有限公司 Video communication method and mobile terminal
WO2018119632A1 (en) * 2016-12-27 2018-07-05 深圳市大疆创新科技有限公司 Image processing method, device and equipment
CN206712945U (en) * 2017-04-26 2017-12-05 联想新视界(天津)科技有限公司 Video communications system
CN108932053A (en) * 2018-05-21 2018-12-04 腾讯科技(深圳)有限公司 Drawing practice, device, storage medium and computer equipment based on gesture
CN111614922A (en) * 2019-02-22 2020-09-01 ***通信有限公司研究院 Information interaction method, network terminal and terminal
CN110233841A (en) * 2019-06-11 2019-09-13 上海文景信息科技有限公司 A kind of long-distance education data interaction system and method based on AR holographic glasses

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114025237A (en) * 2021-12-02 2022-02-08 维沃移动通信有限公司 Video generation method and device and electronic equipment

Also Published As

Publication number Publication date
CN112565844B (en) 2023-05-12

Similar Documents

Publication Publication Date Title
CN108958615B (en) Display control method, terminal and computer readable storage medium
US8244233B2 (en) Systems and methods for operating a virtual whiteboard using a mobile phone device
US9640144B2 (en) Projector, figure input/display apparatus, portable terminal, and program
CN107122113B (en) Method and device for generating picture
CN107977083B (en) Operation execution method and device based on VR system
CN105045504B (en) Picture material extracting method and device
CN110096209A (en) Handwriting trace display methods and device
CN105739857A (en) Mobile terminal control method and apparatus
US9116757B2 (en) Data processing apparatus including plurality of applications and method
CN112312217A (en) Image editing method and device, computer equipment and storage medium
JP2022020703A (en) Handwriting device and speech and handwriting communication system
CN105335198A (en) Font addition method and device
CN113794923A (en) Video processing method and device, electronic equipment and readable storage medium
WO2021197260A1 (en) Note creating method and electronic device
CN112565844B (en) Video communication method and device and electronic equipment
JP2008117083A (en) Coordinate indicating device, electronic equipment, coordinate indicating method, coordinate indicating program, and recording medium with the program recorded thereon
CN110519517B (en) Copy guiding method, electronic device and computer readable storage medium
CN112416199A (en) Control method and device and electronic equipment
CN113986080A (en) Multimedia file editing method and device and electronic equipment
CN113923392A (en) Video recording method, video recording device and electronic equipment
CN114296587A (en) Cursor control method and device, electronic equipment and storage medium
CN112506393B (en) Icon display method and device and storage medium
CN117311526A (en) Method, device and system for controlling terminal equipment based on handwriting pen
CN117555414A (en) Interaction method, interaction device, electronic equipment and storage medium
CN115840506A (en) Data interaction method, device, equipment and medium based on wearable equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant