CN114915852A - Video call interaction method and device, computer equipment and storage medium - Google Patents

Video call interaction method and device, computer equipment and storage medium Download PDF

Info

Publication number
CN114915852A
CN114915852A CN202110178087.1A CN202110178087A CN114915852A CN 114915852 A CN114915852 A CN 114915852A CN 202110178087 A CN202110178087 A CN 202110178087A CN 114915852 A CN114915852 A CN 114915852A
Authority
CN
China
Prior art keywords
camera
video
picture
real time
area
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110178087.1A
Other languages
Chinese (zh)
Other versions
CN114915852B (en
Inventor
许显杨
肖欢
贺坤
刘天成
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Technology Shenzhen Co Ltd
Original Assignee
Tencent Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Technology Shenzhen Co Ltd filed Critical Tencent Technology Shenzhen Co Ltd
Priority to CN202110178087.1A priority Critical patent/CN114915852B/en
Publication of CN114915852A publication Critical patent/CN114915852A/en
Application granted granted Critical
Publication of CN114915852B publication Critical patent/CN114915852B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/141Systems for two-way working between two video terminals, e.g. videophone
    • H04N7/142Constructional details of the terminal equipment, e.g. arrangements of the camera and the display

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • General Engineering & Computer Science (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The application relates to a video call interaction method and device, computer equipment and a storage medium. The method comprises the following steps: responding to the triggering operation of the video call between a first terminal and a second terminal, and displaying an interaction area of the video call, wherein the first terminal is provided with a first camera and a second camera which work simultaneously when the video call is carried out, and the visual fields of the first camera and the second camera are different; and in the interaction area, displaying at least one part of a first video picture acquired by the first camera in real time, displaying at least one part of a second video picture acquired by the second camera in real time, and displaying at least one part of a video picture acquired by at least one camera of the second terminal in real time. By adopting the method, the pictures shot by different cameras can be displayed simultaneously under the condition that the user does not need to switch the cameras repeatedly, and the video call interaction effect of the user is improved.

Description

Video call interaction method and device, computer equipment and storage medium
Technical Field
The present application relates to the field of computer technologies, and in particular, to a video call interaction method and apparatus, a computer device, and a storage medium.
Background
With the rapid development of internet technology and intelligent devices, the communication modes commonly used by people are more and more diversified, and the video call can transmit the voice and the image of both parties of the call in real time, so that the video call gradually becomes one of the favorite communication modes of the public.
In a current video call mode, in a video call process, a user can display a picture shot by one camera of the intelligent device through a screen, for example, a picture shot by a front camera or a picture shot by a rear camera.
However, when the pictures shot by different cameras need to be displayed, the user needs to repeatedly switch the different cameras, so that the operation is very complicated, and the video pictures have a certain delay due to the fact that more time and equipment operation resources need to be consumed for frequently switching the different cameras back and forth, so that the video call interaction effect of the user is poor.
Disclosure of Invention
In view of the foregoing, it is necessary to provide a video call interaction method, an apparatus, a computer device and a storage medium capable of improving a video call interaction effect.
A video call interaction method, the method comprising:
responding to the triggering operation of the video call between a first terminal and a second terminal, and displaying the interactive area of the video call, wherein the first terminal is provided with a first camera and a second camera which work simultaneously when the video call is carried out, and the visual fields of the first camera and the second camera are different;
and in the interaction area, displaying at least one part of a first video picture acquired by the first camera in real time, displaying at least one part of a second video picture acquired by the second camera in real time, and displaying at least one part of a video picture acquired by at least one camera of the second terminal in real time.
In one embodiment, the displaying, in the interaction area, at least a portion of a first video picture captured by the first camera in real time and at least a portion of a second video picture captured by the second camera in real time includes: and displaying a first video picture acquired by the first camera in real time in a main picture area of the interactive area, and displaying a second video picture acquired by the second camera in real time in a secondary picture area of the interactive area, wherein the size of the main picture area is larger than that of the secondary picture area.
In one embodiment, the method further comprises: and responding to the switching operation of the main picture and the secondary picture triggered in the interactive area, displaying a second video picture acquired by the second camera in real time in the main picture area of the interactive area, and displaying a first video picture acquired by the first camera in real time in the secondary picture area of the interactive area.
In one embodiment, the displaying at least a part of a video picture acquired by at least one camera of the second terminal in real time includes: displaying a video picture acquired by at least one camera of the second terminal in real time in the main picture area of the interactive area; or displaying the video pictures acquired by at least one camera of the second terminal in real time in the secondary picture area of the interactive area.
In one embodiment, the second terminal is provided with a third camera and a fourth camera which work simultaneously when the video call is carried out; the third camera and the fourth camera have different fields of view; the displaying of at least a part of the video picture acquired by the at least one camera of the second terminal in real time comprises: and displaying a third video picture acquired by the third camera in real time in a main picture area of the interactive area, and displaying a fourth video picture acquired by the fourth camera in real time in a secondary picture area of the interactive area.
In one embodiment, the method further comprises: and responding to the switching operation of the main picture and the secondary picture triggered in the interactive area, displaying a fourth video picture acquired by the fourth camera in real time in the main picture area of the interactive area, and displaying a third video picture acquired by the third camera in real time in the secondary picture area of the interactive area.
In one embodiment, the displaying, in the interaction area, at least a part of a first video picture acquired by the first camera in real time, at least a part of a second video picture acquired by the second camera in real time, and at least a part of a video picture acquired by at least one camera of the second terminal in real time includes: and in a main picture area of the interactive area, displaying a first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time in a split screen mode, and displaying a video picture acquired by at least one camera in real time in a secondary picture area of the interactive area, wherein the size of the main picture area is larger than that of the secondary picture area.
In one embodiment, the method further comprises: and responding to the picture fusion operation triggered in the interactive area, and displaying a fusion picture of the first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time in a main picture area of the interactive area.
In one embodiment, the displaying, in the interaction area, at least a part of a first video picture captured by the first camera in real time, at least a part of a second video picture captured by the second camera in real time, and at least a part of a video picture captured by at least one camera of the second terminal in real time includes: and displaying a fusion picture of a first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time in a main picture area of the interaction area, and displaying a video picture acquired by at least one camera in real time in the second terminal in a secondary picture area of the interaction area, wherein the size of the main picture area is larger than that of the secondary picture area.
In one embodiment, the first video picture is a picture containing a portrait; the main picture area in the interaction area displays a fusion picture of a first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time, and the method comprises the following steps: and displaying a fusion picture for fusing the portrait in the first video picture acquired by the first camera in real time into the second video picture acquired by the second camera in real time in the main picture area of the interaction area.
In one embodiment, the first video picture and the second video picture are both pictures containing human figures; the main picture area in the interaction area displays a fusion picture of a first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time, and the method comprises the following steps: and displaying a fusion picture comprising the portrait in the first video picture acquired by the first camera in real time and the portrait in the second video picture acquired by the second camera in real time in the main picture area of the interaction area.
In one embodiment, the displaying, in the interaction area, at least a part of a first video picture acquired by the first camera in real time, at least a part of a second video picture acquired by the second camera in real time, and at least a part of a video picture acquired by at least one camera of the second terminal in real time includes: and displaying a first video picture acquired by the first camera in real time and a video picture acquired by the second terminal in real time in the interactive area, and merging the first video picture and the video picture into a merged picture of a second video picture acquired by the second camera in real time.
In one embodiment, the first video picture is a picture containing a portrait, and the video picture collected by the camera of the second terminal in real time contains the portrait; the interactive region displays a first video picture acquired by the first camera in real time and a video picture acquired by the second terminal in real time, and blends the first video picture and the video picture into a blended picture of a second video picture acquired by the second camera in real time, and the method comprises the following steps: and displaying a portrait in a first video picture acquired by the first camera in real time and a portrait in a video picture acquired by the second terminal in real time in the interaction area, and merging the portrait in the first video picture acquired by the first camera in real time and the portrait in the video picture acquired by the second terminal in real time into a merged picture of a second video picture acquired by the second camera in real time.
In one embodiment, the displaying the interactive area of the video call in response to the triggering operation of the video call between the first terminal and the second terminal comprises: responding to a triggering operation of a multi-person video call between a first terminal and a plurality of second terminals, and displaying an interaction area of the multi-person video call;
the interactive region displays a first video picture acquired by the first camera in real time and a video picture acquired by the second terminal in real time, and blends the first video picture and the video picture into a blended picture of a second video picture acquired by the second camera in real time, and the method comprises the following steps: and displaying a first video picture acquired by the first camera in real time and a video picture acquired by each second terminal in real time in the interaction area, and fusing a fused picture of a second video picture acquired by the second camera in real time.
In one embodiment, the second terminal is provided with a third camera and a fourth camera which work simultaneously when the video call is carried out; the third camera and the fourth camera have different fields of view;
the displaying, in the interaction area, at least a part of a first video picture acquired by the first camera in real time, displaying at least a part of a second video picture acquired by the second camera in real time, and displaying at least a part of a video picture acquired by at least one camera in real time at the second terminal includes:
displaying a fusion picture of a third video picture acquired by the third camera in real time and a fourth video picture acquired by the fourth camera in real time in a main picture area of the interaction area;
and displaying a first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time in a secondary picture area of the interactive area.
In one embodiment, the displaying the interaction area of the video call in response to the triggering operation of the video call between the first terminal and the second terminal includes:
displaying a session page for carrying out instant session with the session object;
and responding to the triggering operation of the video call between the session page and the session object, and entering a video call interaction interface for performing video call interaction between the first terminal and a second terminal corresponding to the session object from the session page.
In one embodiment, the first camera and the second camera are respectively one of a rear camera and a front camera of the first terminal.
A video-call interaction device, the device comprising:
the response module is used for responding to the triggering operation of the video call between the first terminal and the second terminal and displaying the interaction area of the video call, the first terminal is provided with a first camera and a second camera which work simultaneously when the video call is carried out, and the visual fields of the first camera and the second camera are different;
and the display module is used for displaying at least one part of a first video picture acquired by the first camera in real time, displaying at least one part of a second video picture acquired by the second camera in real time and displaying at least one part of a video picture acquired by at least one camera of the second terminal in real time in the interaction area.
A computer device comprising a memory and a processor, the memory storing a computer program, the processor implementing the following steps when executing the computer program:
responding to the triggering operation of a video call between a first terminal and a second terminal, and displaying an interaction area of the video call, wherein the first terminal is provided with a first camera and a second camera which work simultaneously when the video call is carried out, and the visual fields of the first camera and the second camera are different;
and in the interaction area, displaying at least one part of a first video picture acquired by the first camera in real time, displaying at least one part of a second video picture acquired by the second camera in real time, and displaying at least one part of a video picture acquired by at least one camera of the second terminal in real time.
A computer-readable storage medium, on which a computer program is stored which, when executed by a processor, carries out the steps of:
responding to the triggering operation of a video call between a first terminal and a second terminal, and displaying an interaction area of the video call, wherein the first terminal is provided with a first camera and a second camera which work simultaneously when the video call is carried out, and the visual fields of the first camera and the second camera are different;
and in the interaction area, displaying at least one part of a first video picture acquired by the first camera in real time, displaying at least one part of a second video picture acquired by the second camera in real time, and displaying at least one part of a video picture acquired by at least one camera of the second terminal in real time.
A computer program comprising computer instructions stored in a computer readable storage medium, the computer instructions being read from the computer readable storage medium by a processor of a computer device, the processor executing the computer instructions, causing the computer device to perform the steps of the above-mentioned video call interaction method.
According to the video call interaction method, the device, the computer equipment and the storage medium, when the first terminal and the second terminal carry out video call, the first camera and the second camera on the first terminal can be started simultaneously, the first camera collects a first video picture in real time, and the second camera collects a second video picture in real time, so that at least one part of the first video picture and at least one part of the second video picture which are collected by the first camera and the second camera respectively in real time and have different visual fields can be displayed in the interaction area of the video call, the pictures shot by different cameras can be displayed simultaneously under the condition that a user does not need to switch the cameras repeatedly, the operation of repeatedly switching the cameras is omitted, and the video call interaction effect of the user is improved.
Drawings
FIG. 1 is a diagram of an exemplary video call interaction method;
FIG. 2 is a flow diagram illustrating a method for video call interaction in one embodiment;
FIG. 3 is a schematic interface diagram of an interaction area for a video call in one embodiment;
FIG. 4 is a schematic interface diagram of an interaction area for a video call in another embodiment;
FIG. 5 is a diagram illustrating an interface for displaying video frames in a primary frame and a secondary frame in an interactive area, according to an embodiment;
FIG. 6 is a diagram illustrating an interface for switching primary and secondary screens in an interactive area, according to an embodiment;
FIG. 7 is a diagram illustrating an interface for displaying video frames in primary and secondary frames in an interactive area according to another embodiment;
FIG. 8 is a diagram illustrating an interface for switching primary and secondary screens in an interactive area according to another embodiment;
FIG. 9 is a diagram illustrating an interface for displaying a fusion screen in an interactive area, according to an embodiment;
FIG. 10 is a diagram illustrating an interface for displaying a merged image in an interactive area, in accordance with an illustrative embodiment;
FIG. 11 is a schematic interface diagram illustrating interaction areas for a multi-person video call in one embodiment;
FIG. 12 is a block diagram that illustrates the collection and processing of multiple video frames in one embodiment;
FIG. 13 is a block diagram that illustrates the multi-channel video capture and processing for a video call scenario in one embodiment;
FIG. 14 is a timing diagram illustrating two-way video data processing according to one embodiment;
FIG. 15 is a block diagram of a video call interaction device in one embodiment;
FIG. 16 is a diagram of an internal structure of a computer device in one embodiment.
Detailed Description
In order to make the objects, technical solutions and advantages of the present application more clearly understood, the present application is further described in detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of and not restrictive on the broad application.
The video call interaction method provided by the application can be applied to the application environment shown in fig. 1. Referring to fig. 1, the video call interaction method is applied to a video call interaction system including a first terminal 102, a server 104, and a second terminal 106. Wherein the first terminal 102 communicates with the server 104 via a network. The second terminal 106 communicates with the server 104 through a network. The first terminal 102 may communicate with the second terminal 106 through the server 104. The first terminal 102 is provided with a first camera and a second camera which work simultaneously when a video call is made, and the shooting views of the first camera and the second camera are different. The first terminal 102 may respond to the triggering operation of the video call with the second terminal 106, and display an interaction area of the video call, in which at least a part of the first video picture acquired by the first camera in real time is displayed, and at least a part of the second video picture acquired by the second camera in real time is displayed at the same time. The first terminal 102 receives the video pictures collected by the at least one camera of the second terminal 106, and displays at least a part of the video pictures collected by the at least one camera of the second terminal 106 in real time in the interaction area. Therefore, the first terminal can simultaneously display the pictures shot by different cameras under the condition that the user does not need to repeatedly switch the cameras, the operation of repeatedly switching the cameras is omitted, and the video call interaction effect of the user is improved.
In one embodiment, the first terminal is a terminal used by a first user and the second terminal is a terminal used by a second user. The first terminal 102 may send an image formed by at least a part of a first video image acquired by the first camera in real time and at least a part of a second video image acquired by the second camera in real time, which are displayed in the interaction area, to the second terminal 106, and the second terminal 106 may display the image in the interaction area after receiving the image. Therefore, the first user using the first terminal can display the video pictures acquired by the two cameras to the second user at the same time without repeatedly switching the pictures shot by the cameras, and the interaction effect of video call is improved.
In one embodiment, a social application is running on the first terminal 102, a social application is also running on the second terminal 106, and a video call between the first terminal 102 and the second terminal 106 can be established through the social application. The Social application may be an instant messaging application, an SNS (Social Network Service) application, or a live application, among others. The second terminal is a terminal used by a call object currently performing a video call with the first terminal, and the number of the second terminals may be one or more, for example, in a scene of a multi-person video call, there may be a plurality of call objects performing a video call with the first terminal at the same time. The video call between the first terminal and the second terminal may be initiated by the first terminal or initiated by the second terminal. Video calls are a communication mode based on networks that transmit voice and pictures between terminals in real time.
The first terminal 102 and the second terminal 106 may be, but not limited to, various personal computers, notebook computers, smart phones, tablet computers, and portable wearable devices, and the server 104 may be implemented by an independent server or a server cluster composed of a plurality of servers.
In one embodiment, as shown in fig. 2, a video call interaction method is provided, which is described by taking the method as an example applied to the first terminal 102 in fig. 1, and includes the following steps:
step 202, responding to a triggering operation of a video call between a first terminal and a second terminal, and displaying an interaction area of the video call, wherein the first terminal is provided with a first camera and a second camera which work simultaneously when the video call is carried out, and the first camera and the second camera have different visual fields.
When the first user needs to have a video call with the second user, the video call can be established with the second terminal through the first terminal. In this embodiment, the first terminal is provided with at least two cameras, and to distinguish the two cameras, the two cameras are respectively referred to as a first camera and a second camera, it should be noted that terms "first" and "second" in the description, claims and drawings of the present application are used to distinguish different objects, rather than to refer to a certain object specifically, for example, the first camera may be any one of the cameras provided on the first terminal and having a shooting function, and for example, the first video picture may be a video picture shot by any one of the cameras provided on the first terminal and having a shooting function.
In the embodiment of the present application, the first camera and the second camera are two cameras with different shooting views, which are disposed on the first terminal. For example, the first camera may be a front camera and the second camera is a rear camera, or the first camera is a rear camera and the second camera is a front camera. For another example, the first camera and the second camera may be cameras respectively disposed at the left and right sides of the first terminal, or may be cameras respectively disposed at the upper and lower sides of the first terminal.
The triggering operation of the video call may be a triggering operation of initiating the video call, or a triggering operation of receiving the video call. For example, the first terminal may initiate a video call to the second terminal, and after the video call is initiated, the interactive area of the video call is displayed, and the first terminal may also display the interactive area of the video call after the second terminal receives the video call. For another example, the second terminal may initiate a video call to the first terminal, and after the first terminal receives the message of the video call, the interactive area of the video call is automatically displayed, and the first terminal may further display the interactive area of the video call after receiving the video call. The triggering operation of the video call can be a single-click operation, a double-click operation or a sliding operation triggered in the video interaction area.
The interactive area is an area in the display interface for the user to interact with the video call. In some embodiments, only the interaction zone may be displayed on the screen of the first terminal, i.e. the interaction zone fills up the screen of the first terminal. Fig. 3 is a schematic diagram of an interaction area of a video call in one embodiment. Optionally, the interaction area may include at least one icon for turning on a camera, may further include at least one icon for turning off a camera, may further include at least one of an icon for ending a video call and an icon for inviting more users to participate in the video call, and may further include attribute information of the users participating in the video call, for example, may be a user avatar or a user nickname, and the like. The interactive zone may also include attribute information related to the video call, such as information on the length of time the video call was established. The interactive area may further include an icon to turn on or off the mute, and may further include an icon to turn on or off the handsfree function. In some embodiments, the interactive area may also be a small window suspended on other interfaces displayed on the first terminal, for example, the interactive area may be reduced to a small window suspended in the instant messaging session interface, as shown in fig. 4.
Specifically, in implementing the step 202, the first terminal may display the interactive area of the video call when detecting a trigger operation of initiating the video call, and may also display the interactive area of the video call when detecting a trigger operation of accepting a video call initiated by another person.
In one embodiment, the video call interaction method is applicable to instant messaging applications. Specifically, the first terminal can display a session page for performing an instant session with the session object through the instant messaging application, and enter a video call interaction interface for performing video call interaction between the first terminal and a second terminal corresponding to the session object from the session page in response to a triggering operation of a video call between the session page and the session object.
The session page may be a session page of a single chat session, that is, a session page of an instant session with one session object. The session page may also be a session page of a group session, i.e. a session page of an instant session with at least one group member of the group session.
Specifically, after the session page is displayed, the first terminal may enter the video call interaction interface from the session page according to the triggering operation of the video call between the first user and the session object in the session page. In the video call interaction interface, a first user may perform video call interaction with a session object using a second terminal through a first terminal, for example, the first user may display, to the session object, video pictures acquired by a first camera and a second camera simultaneously through a trigger operation, and may also view video pictures acquired by at least one camera of the second terminal.
And 204, in the interaction area, displaying at least one part of a first video picture acquired by the first camera in real time, displaying at least one part of a second video picture acquired by the second camera in real time, and displaying at least one part of a video picture acquired by at least one camera of the second terminal in real time.
Specifically, during a video call, the first terminal may display at least a portion of the first video picture and at least a portion of the second video picture, and simultaneously display at least a portion of the video picture captured by the camera of the second terminal. For example, the first terminal may display the entire content of the first video picture, the entire content of the second video picture, and the entire content of the video picture captured by the camera of the second terminal in real time in the interactive area. The first terminal can also display a fusion picture of the first video picture and the second video picture in the interaction area, and can also display the fusion picture formed by fusing the first video picture, the second video picture and the video picture acquired by the second terminal in real time in the interaction area. The detailed description of step 204 will be described in detail later.
In one embodiment, in order to enable the first terminal to display a video picture acquired by a camera of the second terminal, the second terminal may directly transmit the video picture acquired by at least one camera to the first terminal, and after receiving the video picture, the first terminal displays at least a part of the video picture in the interaction area. In other embodiments, the second terminal may further send a processed picture obtained by processing a video picture acquired by at least one camera to the first terminal, and the first terminal displays at least a part of the video picture in the interactive region according to the processed picture, for example, the processed picture may be a fusion picture obtained by fusing video pictures acquired by two cameras of the second terminal.
In one embodiment, the first terminal may default to automatically turn on the first camera and the second camera after the video call is successfully established, where the first camera acquires the first video picture in real time, the second camera acquires the second video picture in real time, and displays at least a portion of the first video picture and at least a portion of the second video picture in the interaction area.
In an embodiment, the first terminal may further automatically turn on the first camera by default after the video call is established, and according to an operation of turning on the second camera triggered by the first user in the interaction area, the first camera and the second camera work simultaneously, that is, both the first camera and the second camera are in an on state, the first camera acquires the first video picture in real time, the second camera acquires the second video picture in real time, and displays at least a part of the first video picture and at least a part of the second video picture in the interaction area.
According to the video call interaction method, when a first terminal and a second terminal carry out video call, a first camera and a second camera on the first terminal can be started simultaneously, the first camera collects a first video picture in real time, and the second camera collects a second video picture in real time, so that at least one part of the first video picture and at least one part of the second video picture which are collected by the first camera and the second camera respectively in real time and have different visual fields can be displayed in the interaction area of the video call.
In one embodiment, the step 204 of displaying at least a part of a first video picture acquired by a first camera in real time, displaying at least a part of a second video picture acquired by a second camera in real time, and displaying at least a part of a video picture acquired by at least one camera of a second terminal in real time in the interactive area includes: and displaying a first video picture acquired by the first camera in real time in a main picture area of the interactive area, displaying a second video picture acquired by the second camera in real time in a secondary picture area of the interactive area, and displaying a video picture acquired by at least one camera in real time in the interactive area, wherein the size of the main picture area is larger than that of the secondary picture area.
In this case, the main screen area is an area having a larger area than the sub screen area, and therefore, the content displayed in the main screen area is more easily seen by the user than the content displayed in the sub screen area. The main picture area and the secondary picture area divide the whole interactive area into two parts, so that the video pictures acquired by the two cameras in real time can be displayed separately.
Optionally, the position distribution and size of the main picture area and the sub picture area are both fixedly set. For example, the main picture area occupies two thirds of the screen size, and the sub-picture area occupies one third of the screen size, and for example, the main picture area is disposed above the screen and the sub-picture area is disposed below the screen. Alternatively, the position of the main screen area may be fixedly set, and the position of the sub-screen area may be moved, for example, the sub-screen area may be displayed superimposed on the main screen area, and the position of the sub-screen area may be moved to an arbitrary position above the main screen area along with a sliding operation by the user. Optionally, the position distribution and size of the main picture area and the sub-picture area may be adjusted according to user personalized operation, and the ratio of the occupied screen of the main picture area and the sub-picture area may be dynamically adjusted according to a real-time acquisition state of the camera in the call process. It can be understood that when the main picture area is used for displaying the portrait acquired by the rear camera, or when the sub-picture area is used for displaying the portrait acquired by the front camera, or when the sub-picture area is used for displaying the portrait acquired by the rear camera, the proportion of the picture area occupying the screen can also be adjusted in this way.
Specifically, the first terminal may display, in the main picture area, a video picture acquired by the rear camera of the first terminal in real time, and display, in the sub-picture area, a video picture acquired by the front camera of the first terminal in real time. Or the first terminal can display the video pictures acquired by the front camera of the first terminal in real time in the main picture area and display the video pictures acquired by the rear camera of the first terminal in real time in the secondary picture area.
In addition, the first terminal also displays a video picture acquired by at least one camera of the second terminal in real time in the interaction area, and displays a mode of the video picture acquired by the camera of the second terminal in the interaction area.
In some embodiments, the first terminal may display a video picture captured in real time by a camera of the second terminal in the main picture area. Specifically, as shown in part (c) of fig. 5, displaying a video picture captured by at least one camera of the second terminal in real time in the interaction area includes: and displaying the video pictures acquired by at least one camera of the second terminal in real time in the main picture area of the interactive area.
In some embodiments, the first terminal may display a video picture captured in real time by a camera of the second terminal in the secondary picture area. Specifically, as shown in part (a) or part (b) of fig. 5, displaying a video picture captured by at least one camera of the second terminal in real time in an interaction area includes: and displaying the video pictures acquired by at least one camera of the second terminal in real time in a secondary picture area of the interactive area.
Fig. 5 is a schematic interface diagram of an interaction area of a video call in one embodiment. Referring to fig. 5, in the interface diagram, an interactive area is divided into a main screen area 502 and a sub screen area 504. Referring to part (a) of fig. 5, the first terminal displays a video picture captured by a rear camera in a main picture area 502, displays a video picture captured by a front camera in a sub picture area 504, and simultaneously displays a video picture captured by a camera of the second terminal in the sub picture area 504. Referring to part (b) of fig. 5, the first terminal displays a video picture captured by a front camera in a main picture area 502, displays a video picture captured by a rear camera in a sub-picture area 504, and simultaneously displays a video picture captured by a camera of the second terminal in the sub-picture area 504. Referring to part (c) of fig. 5, the first terminal displays a video picture captured by the rear camera in the main picture area 502, displays a video picture captured by the front camera in the sub picture area 504, and simultaneously displays a video picture captured by the camera of the second terminal in the main picture area 502. In fig. 5, the portions (a), (b), and (c) are different embodiments.
In the embodiment, the video pictures acquired by the two cameras of the first terminal are separately displayed through the main picture area and the secondary picture area which are different in size, so that a first user can simultaneously view the video pictures acquired by the two cameras in different visual fields in the video call process without repeatedly switching.
In some embodiments, the first user may change the area in which the first video frame and the second video frame are displayed. Specifically, the method may further include: and responding to the switching operation of the primary and secondary pictures triggered in the interactive area, displaying a second video picture acquired by the second camera in real time in the primary picture area of the interactive area, and displaying a first video picture acquired by the first camera in real time in the secondary picture area of the interactive area.
The main and sub-picture switching operation is an operation for switching a video picture displayed in the main picture area and the sub-picture area. Optionally, the first terminal may display a primary and secondary screen switching icon in the interactive area, and switch the video screen displayed in the primary screen area and the secondary screen area when detecting a trigger operation of the user for the primary and secondary screen switching icon. Optionally, the first terminal may detect a dragging operation of a user for a video picture in the main picture area, and may switch the video picture displayed in the main picture area and the video picture displayed in the secondary picture area when the video picture in the main picture area is dragged to the secondary picture area, or when the video picture in the secondary picture area is dragged to the main picture area by the user.
For example, the first terminal displays a video picture acquired by the rear camera of the first terminal in real time in the main picture area and displays a video picture acquired by the front camera of the first terminal in real time in the secondary picture area according to the triggering operation of the video call, and the first terminal can display the video picture acquired by the front camera of the first terminal in real time in the main picture area and display the video picture acquired by the rear camera of the first terminal in real time in the secondary picture area according to the one-key switching operation of the user.
Fig. 6 is a schematic interface diagram of an interaction area of a video call in one embodiment. Referring to fig. 6, in the interface diagram, an interactive area is divided into a main screen area 602 and a sub screen area 604. The first terminal displays the video pictures collected by the rear camera in the main picture area 602, displays the video pictures collected by the front camera in the secondary picture area 604, and simultaneously displays the video pictures collected by the camera of the second terminal in the secondary picture area 604. The first terminal may display a video picture captured by a front camera in the main picture area 602, a video picture captured by a rear camera in the secondary picture area 604, and simultaneously display a video picture captured by a camera of the second terminal in the secondary picture area 604, in response to a main-secondary picture switching operation 606 triggered by the first user.
In the above embodiment, different video pictures acquired by the two cameras of the first terminal can be displayed in a large picture or a small picture according to the needs or preferences of the user, so that the preferences of different users can be met, and the video call interaction effect is improved.
In some embodiments, the second terminal is provided with a third camera and a fourth camera which work simultaneously when the video call is carried out; the third camera and the fourth camera have different visual fields; displaying at least one part of a video picture acquired by at least one camera of the second terminal in real time in an interaction area, wherein the method comprises the following steps: and displaying a third video picture acquired by the third camera in real time in a main picture area of the interactive area, and displaying a fourth video picture acquired by the fourth camera in real time in a secondary picture area of the interactive area.
In this embodiment, the second terminal can open two cameras when carrying out the video conversation, and the second user just can look over the video picture that two cameras were gathered, can also convey the video picture that two cameras gathered separately to first terminal simultaneously. The second terminal is provided with a third camera and a fourth camera, and the third camera and the fourth camera are two different cameras with different shooting visual fields arranged on the second terminal. For example, the third camera is a front camera and the fourth camera is a rear camera, or the third camera is a rear camera and the fourth camera is a front camera. For example, the third camera and the fourth camera may be cameras respectively disposed on the left and right sides of the second terminal, or may be cameras respectively disposed on the upper and lower sides of the second terminal.
The second terminal can start the third camera and the fourth camera according to the triggering operation of the video call between the first terminals, and sends the video pictures acquired by the third camera and the video pictures acquired by the fourth camera in real time to the first terminal, and the first terminal can display the third video pictures acquired by the third camera in real time and the fourth video pictures acquired by the fourth camera in real time in the main picture area of the interaction area and the secondary picture area of the interaction area while displaying the video pictures acquired by the first camera in real time in the interaction area. For example, the first terminal displays a video picture acquired by a rear camera of the second terminal in real time in a main picture area, and displays a video picture acquired by a front camera of the second terminal in real time in a secondary picture area. For another example, the first terminal displays a video picture acquired by a front camera of the second terminal in real time in the main picture area, and displays a video picture acquired by a rear camera of the second terminal in real time in the secondary picture area. The video pictures collected by different cameras displayed in the main picture area can be displayed in a vertical split screen mode or in a left-right split screen mode. The video pictures collected by different cameras displayed in the secondary picture area can be displayed in a vertical split screen mode or in a left-right split screen mode. In one embodiment, when the main screen area displays the video pictures acquired by different cameras in a split-screen manner, the position distribution and the size of the different video pictures in the main screen area may be fixedly set. For example, when the first terminal displays the front video picture collected by the front camera and the rear video picture collected by the rear camera in a split screen manner, the front video picture and the rear video picture respectively occupy half of the main picture area. For another example, the front video frame occupies two thirds of the main screen area, and the rear video frame occupies one third of the main screen area. For another example, the front video screen is displayed in the upper half of the main screen area and the rear video screen is displayed in the lower half of the main screen area, or vice versa, the front video screen is displayed in the lower half of the main screen area and the rear video screen is displayed in the upper half of the main screen area.
In one embodiment, when the main screen area displays the video pictures acquired by different cameras in a split-screen manner, the position distribution and the size of the different video pictures in the main screen area can be adjusted according to the triggering operation of the user. Specifically, the first terminal may obtain a trigger operation of a user for a screen splitting line in the main picture area, and adjust a ratio occupied by different video pictures according to the trigger operation. For example, when the user slides the split screen line up and down, the proportion occupied by the front video picture and the rear video picture in the main picture area is dynamically adjusted in real time, so that the personalized requirements of the user can be met, the user can be helped to adjust the video pictures in real time in the video call process, and the call video pictures are clearer.
In an embodiment, when the main screen area displays the video pictures acquired by different cameras in a split-screen manner, the position distribution and the size of the different video pictures in the main screen area may be dynamically adjusted by the first terminal in real time according to the real-time video acquisition state. Specifically, the closer the distance to the acquired target is, the larger the target in the acquired video picture is, and then, under the condition that the display area is smaller, a target may be locally missing and a complete video picture cannot be presented, and the first terminal may dynamically adjust the ratio of the front video picture and the rear video picture occupied in the main picture area according to the distance from the target. For example, the first terminal may obtain a distance between the front camera and the portrait of the user in real time, and dynamically adjust a proportion of the front video picture in the main picture area according to the distance, where the larger the distance is, the smaller the collected portrait of the user is, the smaller the proportion of the front video picture in the main picture area is, the smaller the distance is, the larger the collected portrait of the user is, and the larger the proportion of the front video picture in the main picture area is.
Fig. 7 is a schematic interface diagram of an interaction area of a video call in one embodiment. Referring to fig. 7, in the interface diagram, an interactive area is divided into a main screen area 702 and a sub screen area 704. Referring to part (a) of fig. 7, the first terminal displays a video picture captured by a rear camera of the first terminal in a main picture area 702, displays a video picture captured by a front camera of the first terminal in a sub-picture area 704, and simultaneously displays a video picture captured by a rear camera of the second terminal in the main picture area 702 and displays a video picture captured by a front camera of the second terminal in the sub-picture area 704. Referring to part (b) of fig. 7, the first terminal displays a video picture captured by a rear camera of the first terminal in the main picture area 702, displays a video picture captured by a front camera of the first terminal in the sub-picture area 704, and simultaneously displays a video picture captured by a front camera of the second terminal in the main picture area 702 and displays a video picture captured by a rear camera of the second terminal in the sub-picture area 704. Fig. 7 shows an embodiment in which the portion (a) and the portion (b) are different.
In the embodiment, the video pictures acquired by the two cameras of the second terminal are separately displayed through the main picture area and the secondary picture area which are different in size, so that the first user can simultaneously view the video pictures with different visual fields acquired by the two cameras of the second terminal in the video call process without repeatedly switching. In addition, both video call parties can simultaneously view the video pictures of the two cameras of the video call party and the video pictures of the two cameras of the other party, the cameras do not need to be switched repeatedly, and the two cameras can be simultaneously viewed.
In some embodiments, the first user may change the area in which the third video picture and the fourth video picture are displayed. Specifically, the method may further include: and responding to the switching operation of the primary and secondary pictures triggered in the interactive area, displaying a fourth video picture acquired by the fourth camera in real time in the primary picture area of the interactive area, and displaying a third video picture acquired by the third camera in real time in the secondary picture area of the interactive area.
The main and sub picture switching operation is an operation for switching a video picture displayed in the main picture area and the sub picture area. Optionally, the first terminal may display a primary and secondary screen switching icon in the interaction area, and switch the video screen displayed in the primary screen area and the secondary screen area when detecting a trigger operation of the user for the primary and secondary screen switching icon. Optionally, the first terminal may detect a dragging operation of a user for a video picture in the main picture area or the secondary picture area, and when the user drags the video picture in the main picture area to the secondary picture area, or when the user drags the video picture in the secondary picture area to the main picture area, the first terminal may switch the video pictures displayed in the main picture area and the secondary picture area.
For example, the first terminal defaults to display a video picture acquired by the rear camera of the second terminal in real time in the main picture area and display a video picture acquired by the front camera of the second terminal in real time in the secondary picture area according to the triggering operation of the video call, and a user can display the video picture acquired by the front camera of the second terminal in real time in the main picture area and display the video picture acquired by the rear camera of the second terminal in real time in the secondary picture area by one-key switching operation.
Fig. 8 is a schematic interface diagram of an interaction area of a video call in one embodiment. Referring to fig. 6, in the interface diagram, an interactive area is divided into a main screen area, 802, and a sub screen area 804. Referring to fig. 8, the first terminal displays a video frame captured by a rear camera of the first terminal in a main frame area 802, displays a video frame captured by a front camera of the first terminal in a sub-frame area 804, displays a video frame captured by a rear camera of the second terminal in the main frame area 802, and displays a video frame captured by a front camera of the second terminal in the sub-frame area 804. The first terminal can respond to the primary and secondary picture switching operation triggered by the first user, and display the video picture collected by the front camera of the second terminal in the primary picture area 802 and display the video picture collected by the rear camera of the second terminal in the secondary picture area 804.
In the embodiment, the user can freely switch the video pictures displayed in the main picture area and the secondary picture area, compared with the mode of continuously switching the video pictures between different cameras, the seamless switching of the video pictures of different cameras can be realized, the user can always see all the video pictures, and the video picture delay does not exist.
In some embodiments, the first terminal may perform fusion display on the separately displayed video frames according to a trigger operation of the first user. Specifically, the method further comprises: and responding to the picture fusion operation triggered in the interactive area, and displaying a fusion picture of a first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time in a main picture area of the interactive area.
The image fusion operation is an operation for fusing and displaying video images acquired by different cameras. The fusion picture comprises at least one part of the first video picture and at least one part of the second video picture, the fusion picture can be obtained by fusing the first video picture into the second video picture, can be obtained by fusing the second video picture into the first video picture, and can also be obtained by extracting part of the first video picture and part of the second video picture and splicing the extracted part of the first video picture and the part of the second video picture with other images.
Optionally, the first terminal may display the fusion icon in the interactive region, and when a trigger operation of the user for the fusion icon is detected, the video pictures displayed in the main picture region and the secondary picture region are fused to obtain a fusion picture, and then the fusion picture is displayed in the interactive region. Optionally, the first terminal may detect a zoom-in operation of the user with respect to the video pictures in the primary picture area and the secondary picture area, for example, the zoom-in operation is an operation in which at least one contact point detected in the primary picture area and at least one contact point detected in the secondary picture area slide toward each other, fuse the video pictures displayed in the primary picture area and the secondary picture area in response to the zoom-in operation, and display the fused picture in the primary picture area of the interactive area.
For example, the first terminal displays a video picture acquired by a rear camera of the first terminal in real time in a main picture area according to a triggering operation of a video call, and displays a video picture acquired by a front camera of the first terminal in real time in a secondary picture area, and the first terminal fuses a second video picture into a first video picture to obtain a fused picture in response to a picture fusing operation triggered by a user, and then displays the fused picture in an interactive area.
Fig. 9 is a schematic interface diagram of an interaction area of a video call in one embodiment. Referring to fig. 9, the interactive area is divided into a main picture area 902 and a sub picture area 904, the first terminal displays a video picture captured by a rear camera of the first terminal in the main picture area 902, displays a video picture captured by a front camera of the first terminal in the sub picture area 904, and displays a video picture captured by a camera of the second terminal in the sub picture area. The first terminal responds to a picture fusion operation 906 triggered by a first user, a fusion picture of a video picture collected by a front camera of the first terminal and a video picture collected by a rear camera of the second terminal is displayed in a main picture area 902 of an interaction area, and a video picture collected by a camera of the second terminal is displayed in a secondary picture area.
In some embodiments, the first terminal may further switch from the above-described mode of displaying the video picture in the merged state back to the mode of displaying the video picture separately in the primary and secondary picture areas in response to a triggering operation of the user in the interactive area.
In the above embodiment, the user can switch the first terminal from the separate display mode to the fusion display mode of the video pictures acquired by different cameras according to the own needs or preferences, thereby improving the video call interaction effect in the video call process.
The above describes an embodiment in which a first video picture acquired by a first camera of a first terminal in real time and a second video picture acquired by a second camera in real time are displayed in a main picture area and a secondary picture area, respectively. In some embodiments, the first terminal may further display video pictures acquired by the first camera and the second camera in real time in the main picture area or in the secondary picture area at the same time.
In one embodiment, in the interaction area, displaying at least a part of a first video picture acquired by a first camera in real time, displaying at least a part of a second video picture acquired by a second camera in real time, and displaying at least a part of a video picture acquired by at least one camera of a second terminal in real time comprises: and displaying a first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time in a main picture area of the interactive area in a split screen manner, and displaying a video picture acquired by at least one camera in real time in a secondary picture area of the interactive area, wherein the size of the main picture area is larger than that of the secondary picture area.
For example, after the front and rear cameras are simultaneously turned on by the first terminal, the video pictures respectively acquired by the front and rear cameras can be displayed in a split screen manner in the main picture area of the interaction area, and the video pictures acquired by the camera of the second terminal can be displayed in the secondary picture area. The split screen display can be an up-down split screen display or a left-right split screen display.
In the embodiment, the video pictures acquired by the two cameras of the first terminal are separately displayed in the main picture area, and in the video call process, the first user can simultaneously view the video pictures with different visual fields acquired by the two cameras of the first terminal without repeatedly switching and can simultaneously view the video pictures.
In the above embodiment, the first terminal responds to the triggering operation of the video call between the first terminal and the second terminal, and when the interactive region is displayed, the video pictures acquired by the two cameras of the first terminal are separately displayed in a primary and secondary picture manner by default, and the fused picture is displayed in response to the triggering operation of the user. In other embodiments, the first terminal responds to a triggering operation of a video call between the first terminal and the second terminal, and defaults to display video pictures acquired by two cameras of the first terminal in a form of a fusion picture in a main picture area.
Specifically, step 204, in the interaction area, displaying at least a part of a first video picture acquired by the first camera in real time, displaying at least a part of a second video picture acquired by the second camera in real time, and displaying at least a part of a video picture acquired by at least one camera of the second terminal in real time, includes: and displaying a fusion picture of a first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time in a main picture area of the interactive area, and displaying a video picture acquired by at least one camera in real time in a secondary picture area of the interactive area, wherein the size of the main picture area is larger than that of the secondary picture area.
Optionally, the first terminal may further switch from the mode of displaying the video pictures in the merged mode back to the mode of displaying the video pictures separately in the primary and secondary picture areas in response to a trigger operation of the user in the interactive area.
In the above embodiment, after the first terminal starts the first camera and the second camera, a default may be made to display a fusion picture of a first video picture acquired by the first camera and a second video picture acquired by the second camera in the main picture area.
In some embodiments, the first video frame is a frame containing a portrait; in the main picture area of interactive area, show the first video picture of first camera real time collection and the second video picture's of second camera real time collection fuse the picture, include: and displaying a fusion picture for fusing the portrait in the first video picture acquired by the first camera in real time into the second video picture acquired by the second camera in real time in the main picture area of the interaction area.
Specifically, when the first video picture acquired by the first camera in real time contains a portrait, the second video picture acquired by the second camera can be used as a background picture, and after the portrait is divided from the first video picture, the first terminal merges the divided portrait into the second video picture.
For example, the front camera and the rear camera of the first terminal simultaneously capture video pictures, the first terminal uses the video pictures captured by the rear camera as background pictures, and the portrait and the body part in the video pictures captured by the front camera are merged into the video pictures captured by the rear camera. Therefore, the background picture is a real, dynamic and real-time picture, and compared with a mode of integrating the portrait into a preset background picture, the user can carry out video call in a real, dynamic and real-time collected scene, and particularly in journey and outdoor activities, the user can experience the aesthetic feeling of people in the midstream of drawing, the interactive feeling is stronger, and the interactive effect of video call is better.
In one embodiment, in order to make the effect of the fusion picture more real and make the interaction feel stronger, the first terminal may further perform scaling processing on the portrait extracted from the front-facing camera, so that the portrait fused into the background picture is more real, and the situation of fusion picture distortion caused by too large portrait and too small background picture or too small portrait and too much background picture is avoided. Specifically, the smaller the distance between the front camera and the user portrait, the larger the collected user portrait, and vice versa, so that the first terminal can obtain the distance between the front camera and the user in real time, determine the scaling of the user portrait and the body part extracted from the front video image according to the distance, perform scaling processing on the user portrait and the body part according to the scaling, and blend the user portrait and the body part obtained after scaling processing into the video image collected by the rear camera to obtain a blended image.
In one embodiment, the first terminal can also send the fusion picture to the second terminal, and the second terminal can show the fusion picture to the other party of the call, so that both parties of the call can see the fusion picture, and when the first user makes a video call with the other party, the first user does not need to repeatedly switch the camera for showing the landscape picture again in order to show the first user, thereby improving the video call effect.
In one embodiment, the first terminal may display a video picture acquired by a camera of the second terminal in real time in the secondary picture area of the interaction area. Optionally, when the second terminal is provided with only one camera or only one camera is turned on, the first terminal displays a video picture acquired by the camera in real time in the secondary picture area of the interactive area. Optionally, when the second terminal turns on two cameras, the first terminal displays at least one of the video pictures acquired by the two cameras in real time in a secondary picture area of the interaction area, for example, the video picture acquired by the rear camera of the second terminal is only displayed in the secondary picture area, or the video pictures acquired by the front camera and the rear camera of the second terminal are simultaneously displayed in the secondary picture area.
Fig. 10 is a schematic interface diagram of an interaction area of a video call in one embodiment. Referring to fig. 10, in the interface diagram, an interactive area is divided into a main screen area 1002 and a sub screen area 1004. Referring to part (a) of fig. 10, the first terminal displays a video frame captured by the camera of the second terminal in the secondary screen area 1004, in which a portrait in a video frame captured by the camera of the front terminal is merged with a video frame captured by the camera of the rear terminal in the main screen area 1002. Referring to part (b) of fig. 10, in the case that the first terminal displays a video frame acquired by the rear camera in the main frame region 1002, and the second terminal simultaneously turns on the front camera and the rear camera, the first terminal may also display a video frame acquired by the rear camera of the second terminal in the sub frame region 1004. Fig. 10 shows an embodiment in which the portion (a) and the portion (b) are different.
In the embodiment, the portrait is fused into a real scene picture, so that a user can carry out video call in a real, dynamic and real-time acquired scene, and the video call interaction effect is better.
In some embodiments, the first video frame and the second video frame are both frames containing human images; in the main picture area of interactive area, show the first video picture of first camera real time collection and the second video picture's of second camera real time collection fuse the picture, include: and displaying a fusion picture comprising the portrait in the first video picture acquired by the first camera in real time and the portrait in the second video picture acquired by the second camera in real time in the main picture area of the interaction area.
Optionally, the first terminal may extract a portrait from the first video image, and then display a fusion image obtained by fusing the portrait into the second video image acquired by the second camera. Optionally, the first terminal may further extract a portrait from the second video image, and then display a fusion image obtained by fusing the portrait into the first video image acquired by the first camera. Optionally, the first terminal may further blend the portrait extracted from the first video picture and the portrait extracted from the second video picture into a preset virtual background.
It should be noted that, when the fusion picture is generated according to the portrait extracted from the first video picture and the portrait extracted from the second video picture, in order to ensure the effect of the fusion picture and the actual situation of the user's photo-integration, the portraits in the generated fusion picture should not be overlapped or overlapped, so that the portraits appearing in the first video picture and the second video picture can be displayed through the fusion picture.
For example, when a family member or an outdoor event occurs, the video pictures acquired by the front camera and the rear camera both include the portrait, the video pictures acquired by the front camera include the portrait of a photographer, the video pictures acquired by the rear camera include the portrait of other people, and at this time, the first terminal can blend the portrait extracted from the video pictures acquired by the front camera into the video pictures acquired by the rear camera and display the blended picture in the main picture area. Therefore, all the people participating in the party can display the fusion picture, when the first terminal sends the fusion picture to the second terminal, the video call opposite side can see the picture including all the people, and the first user does not need to repeatedly switch the front camera and the rear camera to transmit different live pictures to the video call opposite side.
In one embodiment, when the first terminal blends the portrait in the video image acquired by the front camera and the portrait in the video image acquired by the rear camera into the same image to obtain the fused image, in order to ensure that the proportions of the obtained portrait in the fused image are uniform and the fused image is more real and better, the first terminal can obtain a first distance and a second distance respectively between the front camera and the rear camera and different objects to be shot, determine a first scaling ratio of the portrait and the body part of the user extracted from the front video image according to the first distance, determine a second scaling ratio of the portrait and the body part of the user extracted from the rear video image according to the second distance, and scale the portrait and the body part of the user extracted from the front video image and the face and the body part of the user extracted from the rear video image according to the first scaling ratio and the second scaling ratio, so that the face and the body part extracted from the front video image and the face and the body part of the user extracted from the rear video image are scaled according to the first scaling ratio and the second scaling ratio The proportion of the points is uniform, and the user portrait and the body part obtained after the zooming processing are blended into the same video picture to obtain a blended picture.
Referring to part (c) of fig. 10, if the video frame captured by the front camera of the first terminal includes a portrait and the video frame captured by the rear camera also includes a plurality of portraits, the first terminal may display the portrait in the video frame captured by the front camera in the main frame area 1002, and the first terminal may display the video frame captured by the camera of the second terminal in the sub-frame area 1004. Fig. 10 shows different embodiments in parts (a), (b), and (c).
In the above embodiment, when both the first video picture and the second video picture include the portrait, a group photo mode is formed by displaying a fusion picture including the portrait in the two video pictures in the main picture area, so that the opposite party can conveniently and quickly view the pictures of all the people.
In some embodiments, displaying at least a portion of a first video picture captured in real time by a first camera, displaying at least a portion of a second video picture captured in real time by a second camera, and displaying at least a portion of a video picture captured in real time by at least one camera of a second terminal in the interaction area comprises: and displaying a first video picture acquired by the first camera in real time and a video picture acquired by the second terminal in real time in the interactive area, and fusing a fused picture of a second video picture acquired by the second camera in real time.
Optionally, when the first video picture acquired by the first terminal includes a portrait and the video picture acquired by the second terminal includes a portrait, the first terminal may extract the portrait from the first video picture and extract the portrait from the video picture acquired by the second terminal, and then blend the extracted portrait into the second video picture acquired by the second camera to obtain a blended picture, and directly display the blended picture in the interactive area. For example, the first terminal can fuse the portrait in the video picture collected by the front camera and the portrait in the video picture collected by the camera of the second terminal into the video picture collected by the rear camera of the first terminal to form a group photo mode, so that the interestingness of both video call parties in the video call process is improved.
In one embodiment, in response to a triggering operation of a video call between a first terminal and a second terminal, displaying an interaction area of the video call, comprises: responding to the triggering operation of the multi-person video call between the first terminal and a plurality of second terminals, and displaying an interaction area of the multi-person video call;
in the interactive area, show the first video picture of gathering in real time with first camera and the video picture of gathering in real time at second terminal, merge into the integration picture of the second video picture of gathering in real time with the second camera, include: and displaying a first video picture acquired by the first camera in real time and a video picture acquired by each second terminal in real time in the interactive area, and fusing a fusion picture of a second video picture acquired by the second camera in real time.
In a multi-person video call scene, when a first terminal carries out video call with a plurality of second terminals, in the scene, the first terminal opens two cameras, a video picture collected by the second camera is used as a background, the video picture collected by the first camera and the video pictures collected by the cameras of a plurality of call opposite sides are merged into the background, and pictures of all users participating in the call can appear in the merged picture.
For example, in a multi-person meeting scene or a multi-person online meeting scene, a multi-person picture can be displayed in a fusion picture, and people participating in the multi-person video call can view the pictures of all people. Optionally, the terminal of all the people participating in the multi-person video call may display the merged picture.
Optionally, the first terminal may use a video picture acquired by a second camera of the originator who initiates the multi-person video call as a background picture, and merge the person images in the video pictures acquired by the other cameras into the background picture to obtain a merged picture and display the merged picture, where the originator who initiates the multi-person video call may be a first user who uses the first terminal. Optionally, the first terminal may further use a video picture acquired by any one terminal specified by the initiator as the background picture, or default to use a video picture acquired by the second camera of the initiator as the background picture when the initiator starts two cameras, and use a video picture acquired by any one terminal specified by the initiator as the background picture when the initiator only starts one camera.
Fig. 11 is a schematic interface diagram of an interaction area of a video call in one embodiment. Referring to fig. 11, in the interface schematic diagram, when the first terminal is a terminal used by a user initiating a multi-user video call, the first terminal uses a video picture acquired by a rear camera in real time as a background picture, and merges the portrait in the video picture acquired by each camera of other terminals participating in the multi-user video call in real time into the background picture, so as to obtain a multi-user merged picture. The first terminal can display the multi-person fusion picture in the interaction area.
In one embodiment, the second terminal is provided with a third camera and a fourth camera which work simultaneously when the video call is carried out; the third camera and the fourth camera have different visual fields; the first terminal can also display a fusion picture of video pictures collected by the two cameras of the second terminal. Specifically, step 204, in the interaction area, displaying at least a part of a first video picture acquired by the first camera in real time, displaying at least a part of a second video picture acquired by the second camera in real time, and displaying at least a part of a video picture acquired by at least one camera of the second terminal in real time, includes: displaying a fusion picture of a third video picture acquired by the third camera in real time and a fourth video picture acquired by the fourth camera in real time in a main picture area of the interaction area; and displaying a first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time in a secondary picture area of the interactive area.
In one embodiment, in order to enable the first terminal to display a fusion picture of video pictures acquired by a camera of the second terminal, the second terminal may directly send the third video picture and the fourth video picture to the first terminal, and after receiving the video pictures, the first terminal displays the fusion picture of the third video picture and the fourth video picture in the interaction area. In other embodiments, the second terminal may further fuse the third video picture and the fourth video picture to obtain a fused picture, and send the fused picture to the first terminal, and the first terminal displays the fused picture in the interactive area.
The application also provides an application scene, and the application scene applies the video call interaction method. Specifically, the application of the video call interaction method in the application scene is as follows:
in an application scene, a first user opens a single chat session page corresponding to a session, in the single chat session, a video call between second terminals corresponding to a session object is initiated, the first terminal simultaneously starts a front camera and a rear camera of the first terminal, the front camera of the first terminal collects a front picture in real time, the rear camera of the first terminal collects a rear picture in real time, the first terminal can display the rear picture in a main picture area of an interaction area by default, the front picture is displayed in a secondary picture area of the interaction area, and simultaneously the video picture collected by the camera of the second terminal in real time is displayed in the secondary picture area. The first terminal can also display a front picture in a main picture area of the interactive area and display a rear picture in a secondary picture area of the interactive area according to the triggering operation of the first user in the interactive area. When the front camera and the rear camera are both simultaneously opened at the first terminal and the second terminal, the first terminal can split-screen display the respective rear pictures of the first terminal and the second terminal in the main picture area, and split-screen display the respective front pictures of the first terminal and the second terminal in the secondary picture area. The first terminal can also switch the mode of displaying the video pictures in the primary picture and the secondary picture in the interactive area to the mode of displaying the video pictures in the fusion picture in the interactive area according to the triggering operation of the first user in the interactive area. The first terminal takes the rear picture as a background picture, a portrait in the front picture is fused into the background picture to obtain a fusion picture, the fusion picture is displayed in the main picture area, and a video picture collected by a camera of the second terminal is displayed in the secondary picture area. The rear screen may be a landscape screen, and may be a screen including a plurality of persons. When the front camera and the rear camera are both simultaneously opened by the first terminal and the second terminal, the first terminal can display a front-end picture or a rear-end picture of the second terminal in the secondary picture area.
In another application scene, a first user opens a group session page, a video call between second terminals corresponding to a plurality of session objects is initiated in the group session page, the first terminal simultaneously starts a front camera and a rear camera of the first terminal, the front camera of the first terminal collects a front picture in real time, the rear camera of the first terminal collects a rear picture in real time, the first terminal can fuse a portrait in the front picture and other portraits of participants participating in the multi-person video session into the rear picture collected by the first terminal to obtain a fused picture, and the fused picture is displayed in an interactive area. The first terminal can also send the fusion picture to a plurality of second terminals participating in the multi-person video call, so that all the persons participating in the multi-person video call can view the fusion picture, and all the persons can be viewed.
In this embodiment of the application, in order to satisfy the requirement that a dual-open camera carries out a video call, the first terminal may adopt a multi-channel video acquisition and processing framework, and the acquisition and processing of multi-channel video pictures are realized based on the framework. Referring to fig. 12, which is a schematic diagram of a framework for acquiring and processing multiple video frames in an embodiment, the framework includes a front camera, a rear camera, a first video data input source, a second video data input source, a camera session management, a first video data output source, a second video data output source, and a camera data rendering.
The acquisition and processing flow of the multi-channel video pictures based on the frame comprises the following steps: firstly, initializing a front camera and a rear camera according to video acquisition parameters, wherein the video acquisition parameters can comprise resolution, frame rate, data stream mode and the like, then binding a first video data input source with the front camera, binding a second video data input source with the rear camera, binding the first video data input source with a second video data output source through camera session management, and binding the second video data input source with the second video data output source. Therefore, after the front camera and the rear camera respectively collect video data, the video data are transmitted to the respective bound video data input sources, the video data flow from the video data input sources to the bound video data output sources, the video data are called back to the service layer through the proxy function after being output by the video data output sources, the call-back parameters comprise video frame data, and finally the service layer takes out original video data through the call-back parameters to perform subsequent video processing, including preprocessing, rendering display, network transmission and the like.
In one embodiment, the multi-path video acquisition and processing framework may be an AVFoundation, and the first may perform dual-camera video data acquisition using the AVFoundation, including a front camera and a rear camera, and render the acquired video frames onto a screen.
In an embodiment, the above multi-channel video capturing and processing framework is applied to an audio/video call scene, and a schematic diagram of the multi-channel video capturing and processing framework in the video call interaction process is shown in fig. 13. Referring to fig. 13, the device comprises a video data acquisition module, a video data preprocessing module, a rendering module, a display module, a coding module and a network module, wherein the audio and video data acquisition module is used for transmitting a triggering operation of starting a video call or an operation of turning on/off a camera, which is triggered by a user, to a multi-path video acquisition and processing framework so as to trigger acquisition of video data; the video data preprocessing module is used for preprocessing video data acquired by the two cameras in the video call process, and the preprocessing includes but is not limited to filter adding processing, splicing processing, fusion processing and the like; the rendering module is used for rendering the preprocessed video data, then displaying the rendered data through the display module, and the encoding module is used for encoding the acquired original video data and then sending the encoded data to a terminal used by the other party of the call through the network module.
In one embodiment, as shown in fig. 14, a timing diagram of two-way video data processing during a video call in one embodiment is shown. Referring to fig. 14, since the camera is started and the camera collects video data more time-consuming, in order to ensure user experience of a user in a video call process, a main thread is used for responding to a trigger operation of the user and processing the video data, and a collection thread is used as a sub-thread and is used for interacting with the camera, for example, the camera collects video pictures before and after the camera is started or the camera collects video surfaces before and after the camera is closed, and sends the collected video data to the main thread for processing. The life cycles of the main thread and the sub-thread are not affected.
Referring to fig. 14, after a user initiates a video call, a main thread sends initial video acquisition parameters to an acquisition thread in response to the operation of the user, where the initial video acquisition parameters may include a resolution, a frame rate, a data stream mode, and an on or off state of two cameras, where the resolution, the frame rate, and the data stream mode may be determined according to the device performance of the current terminal. After the acquisition thread returns that the initialization is successful, the main thread informs the acquisition thread to start acquiring video data by using the data acquisition module, the acquisition thread starts the front camera and the rear camera to acquire the video data and returns the acquired video data to the data acquisition module in the main thread, and after the main thread preprocesses the video data by using the data preprocessing module, the main thread displays a video picture by using the display module and transmits the video data by using the network module. And after the user finishes the video call, the main thread responds to the triggering operation of the user to inform the acquisition thread of finishing the acquisition of the video data, and the acquisition thread feeds back to the main thread to close the video data acquisition.
It should be understood that, although the steps in the flowchart of fig. 2 are shown in order as indicated by the arrows, the steps are not necessarily performed in order as indicated by the arrows. The steps are not performed in the exact order shown and described, and may be performed in other orders, unless explicitly stated otherwise. Moreover, at least a part of the steps in fig. 2 may include multiple steps or multiple stages, which are not necessarily performed at the same time, but may be performed at different times, and the order of performing the steps or stages is not necessarily performed in sequence, but may be performed alternately or alternately with other steps or at least a part of the steps or stages in other steps.
In one embodiment, as shown in fig. 15, there is provided a video call interaction apparatus 1500, which may be a part of a computer device using a software module or a hardware module, or a combination of the two, and specifically includes: a response module 1502 and a display module 1504, wherein:
the response module 1502 is configured to display an interaction area of a video call in response to a trigger operation of the video call between a first terminal and a second terminal, where the first terminal is provided with a first camera and a second camera that simultaneously work when the video call is performed, and the first camera and the second camera have different fields of view;
the display module 1504 is configured to display at least a part of a first video picture acquired by the first camera in real time, display at least a part of a second video picture acquired by the second camera in real time, and display at least a part of a video picture acquired by at least one camera of the second terminal in real time in the interaction area.
In one embodiment, the display module 1504 is further configured to display a first video picture captured by the first camera in real time in a main picture area of the interactive area, and display a second video picture captured by the second camera in real time in a sub-picture area of the interactive area, where a size of the main picture area is larger than a size of the sub-picture area. In one embodiment, the display module 1504 is further configured to display the second video frame captured by the second camera in real time in the main frame area of the interactive area and display the first video frame captured by the first camera in real time in the sub-frame area of the interactive area in response to the main-sub frame switching operation triggered in the interactive area.
In one embodiment, the display module 1504 is further configured to display, in a main picture area of the interaction area, a video picture acquired by at least one camera of the second terminal in real time; or displaying the video pictures acquired by at least one camera of the second terminal in real time in a secondary picture area of the interactive area.
In one embodiment, the second terminal is provided with a third camera and a fourth camera which work simultaneously when the video call is carried out; the third camera and the fourth camera have different visual fields; the display module 1504 is further configured to display a third video picture acquired by the third camera in real time in the main picture area of the interactive area, and display a fourth video picture acquired by the fourth camera in real time in the sub-picture area of the interactive area.
In one embodiment, the display module 1504 is further configured to display a fourth video picture captured by the fourth camera in real time in the main picture area of the interactive area and a third video picture captured by the third camera in real time in the sub-picture area of the interactive area in response to the main-sub picture switching operation triggered in the interactive area.
In one embodiment, the display module 1504 is further configured to display, in a main picture area of the interaction area, the first video picture collected by the first camera in real time and the second video picture collected by the second camera in real time in a split-screen manner, and display the video picture collected by at least one camera of the second terminal in a sub-picture area of the interaction area, where a size of the main picture area is larger than a size of the sub-picture area.
In one embodiment, the display module 1504 is further configured to display a fused picture of a first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time in a main picture area of the interactive area in response to a picture fusion operation triggered in the interactive area.
In one embodiment, the display module 1504 is further configured to display, in a main picture area of the interaction area, a fusion picture of a first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time, and display, in a sub-picture area of the interaction area, a video picture acquired by at least one camera of the second terminal in real time, where a size of the main picture area is larger than a size of the sub-picture area.
In one embodiment, the first video picture is a picture containing a portrait; the display module 1504 is further configured to display, in the main picture area of the interaction area, a fusion picture in which a portrait in a first video picture acquired by the first camera in real time is fused into a second video picture acquired by the second camera in real time.
In one embodiment, the first video frame and the second video frame are both frames containing human images; the display module 1504 is further configured to display a fusion picture including a portrait in a first video picture acquired by the first camera in real time and a portrait in a second video picture acquired by the second camera in real time in the main picture area of the interaction area.
In an embodiment, the display module 1504 is further configured to display, in the interaction area, a fusion picture in which a first video picture acquired by the first camera in real time and a video picture acquired by the second terminal in real time are fused into a second video picture acquired by the second camera in real time.
In one embodiment, the first video picture is a picture containing a portrait, and the video picture collected by the camera of the second terminal in real time contains the portrait; the display module 1504 is further configured to display a fusion picture in which a portrait in a first video picture acquired by the first camera in real time and a portrait in a video picture acquired by the second terminal in real time are fused into a second video picture acquired by the second camera in real time in the interaction area.
In one embodiment, the response module 1502 is further configured to display an interaction area of the multi-person video call in response to a triggering operation of the multi-person video call between the first terminal and the plurality of second terminals, where the triggering operation is initiated by the first terminal; the display module 1504 is further configured to display a fusion picture in which a first video picture acquired by the first camera in real time and video pictures acquired by each second terminal in real time are fused into a second video picture acquired by the second camera in real time in the interaction area.
In one embodiment, the second terminal is provided with a third camera and a fourth camera which work simultaneously when the video call is carried out; the third camera and the fourth camera have different visual fields; the display module 1504 is further configured to display, in the main picture area of the interaction area, a fusion picture of a third video picture acquired by the third camera in real time and a fourth video picture acquired by the fourth camera in real time; and displaying a first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time in a secondary picture area of the interactive area.
In one embodiment, response module 1502 is further configured to present a session page for an instant session with a session object; and responding to the triggering operation of the video call between the session page and the session object, and entering a video call interaction interface for performing video call interaction between the first terminal and a second terminal corresponding to the session object from the session page.
In one embodiment, the first camera and the second camera are each one of a rear camera and a front camera of the first terminal.
According to the video call interaction device 1500, when a first terminal and a second terminal perform a video call, a first camera and a second camera on the first terminal can be started simultaneously, the first camera collects a first video picture in real time, and the second camera collects a second video picture in real time, so that at least one part of the first video picture and at least one part of the second video picture which are collected by the first camera and the second camera respectively in real time and have different visual fields can be displayed in the interaction area of the video call.
For specific limitations of the video call interaction apparatus 1500, reference may be made to the above limitations on the video call interaction method, which is not described herein again. All or part of the modules in the video call interaction apparatus 1500 may be implemented by software, hardware, and a combination thereof. The modules can be embedded in a hardware form or independent of a processor in the computer device, and can also be stored in a memory in the computer device in a software form, so that the processor can call and execute operations corresponding to the modules.
In one embodiment, a computer device is provided, which may be a terminal, and its internal structure diagram may be as shown in fig. 16. The computer equipment comprises a processor, a memory, a communication interface, a display screen, an input device, a first camera and a second camera which are connected through a system bus. Wherein the processor of the computer device is configured to provide computing and control capabilities. The memory of the computer device comprises a nonvolatile storage medium and an internal memory. The non-volatile storage medium stores an operating system and a computer program. The internal memory provides an environment for the operation of an operating system and computer programs in the non-volatile storage medium. The communication interface of the computer device is used for carrying out wired or wireless communication with an external terminal, and the wireless communication can be realized through WIFI, an operator network, NFC (near field communication) or other technologies. The computer program is executed by a processor to implement a video call interaction method. The display screen of the computer equipment can be a liquid crystal display screen or an electronic ink display screen, and the input device of the computer equipment can be a touch layer covered on the display screen, a key, a track ball or a touch pad arranged on a shell of the computer equipment, an external keyboard, a touch pad or a mouse and the like.
It will be appreciated by those skilled in the art that the configuration shown in fig. 16 is a block diagram of only a portion of the configuration associated with the present application, and is not intended to limit the computing device to which the present application may be applied, and that a particular computing device may include more or less components than those shown, or may combine certain components, or have a different arrangement of components.
In one embodiment, a computer device is further provided, which includes a memory and a processor, the memory stores a computer program, and the processor implements the steps of the above method embodiments when executing the computer program.
In an embodiment, a computer-readable storage medium is provided, in which a computer program is stored which, when being executed by a processor, carries out the steps of the above-mentioned method embodiments.
In one embodiment, a computer program product or computer program is provided that includes computer instructions stored in a computer readable storage medium. The computer instructions are read by a processor of a computer device from a computer-readable storage medium, and the computer instructions are executed by the processor to cause the computer device to perform the steps in the above-mentioned method embodiments.
It will be understood by those skilled in the art that all or part of the processes of the methods of the embodiments described above can be implemented by hardware instructions of a computer program, which can be stored in a non-volatile computer-readable storage medium, and when executed, can include the processes of the embodiments of the methods described above. Any reference to memory, storage, database or other medium used in the embodiments provided herein can include at least one of non-volatile and volatile memory. Non-volatile Memory may include Read-Only Memory (ROM), magnetic tape, floppy disk, flash Memory, optical Memory, or the like. Volatile Memory can include Random Access Memory (RAM) or external cache Memory. By way of illustration and not limitation, RAM can take many forms, such as Static Random Access Memory (SRAM) or Dynamic Random Access Memory (DRAM), for example.
The technical features of the above embodiments can be arbitrarily combined, and for the sake of brevity, all possible combinations of the technical features in the above embodiments are not described, but should be considered as the scope of the present specification as long as there is no contradiction between the combinations of the technical features.
The above-mentioned embodiments only express several embodiments of the present application, and the description thereof is more specific and detailed, but not construed as limiting the scope of the invention. It should be noted that, for a person skilled in the art, several variations and modifications can be made without departing from the concept of the present application, which falls within the scope of protection of the present application. Therefore, the protection scope of the present patent shall be subject to the appended claims.

Claims (20)

1. A video call interaction method, the method comprising:
responding to the triggering operation of a video call between a first terminal and a second terminal, and displaying an interaction area of the video call, wherein the first terminal is provided with a first camera and a second camera which work simultaneously when the video call is carried out, and the visual fields of the first camera and the second camera are different;
and in the interaction area, displaying at least one part of a first video picture acquired by the first camera in real time, displaying at least one part of a second video picture acquired by the second camera in real time, and displaying at least one part of a video picture acquired by at least one camera of the second terminal in real time.
2. The method of claim 1, wherein displaying at least a portion of a first video frame captured by the first camera in real time and at least a portion of a second video frame captured by the second camera in real time in the interactive region comprises:
and displaying a first video picture acquired by the first camera in real time in a main picture area of the interactive area, and displaying a second video picture acquired by the second camera in real time in a secondary picture area of the interactive area, wherein the size of the main picture area is larger than that of the secondary picture area.
3. The method of claim 2, further comprising:
and responding to the switching operation of the main picture and the secondary picture triggered in the interactive area, displaying a second video picture acquired by the second camera in real time in the main picture area of the interactive area, and displaying a first video picture acquired by the first camera in real time in the secondary picture area of the interactive area.
4. The method according to claim 2, wherein the displaying at least a portion of the video pictures captured by the at least one camera of the second terminal in real time comprises:
displaying a video picture acquired by at least one camera of the second terminal in real time in the main picture area of the interactive area; or displaying the video pictures acquired by at least one camera of the second terminal in real time in the secondary picture area of the interactive area.
5. The method according to claim 2, wherein the second terminal is provided with a third camera and a fourth camera which operate simultaneously when the video call is made; the third camera and the fourth camera have different fields of view; the displaying of at least a part of the video pictures acquired by the at least one camera of the second terminal in real time comprises:
and displaying a third video picture acquired by the third camera in real time in a main picture area of the interactive area, and displaying a fourth video picture acquired by the fourth camera in real time in a secondary picture area of the interactive area.
6. The method of claim 5, further comprising:
and responding to the switching operation of the main picture and the secondary picture triggered in the interactive area, displaying a fourth video picture acquired by the fourth camera in real time in the main picture area of the interactive area, and displaying a third video picture acquired by the third camera in real time in the secondary picture area of the interactive area.
7. The method of claim 1, wherein the displaying, in the interaction area, at least a portion of a first video picture captured in real time by the first camera, at least a portion of a second video picture captured in real time by the second camera, and at least a portion of a video picture captured in real time by at least one camera of the second terminal comprises:
and in a main picture area of the interactive area, displaying a first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time in a split screen mode, and displaying a video picture acquired by at least one camera in real time in a secondary picture area of the interactive area, wherein the size of the main picture area is larger than that of the secondary picture area.
8. The method of claim 2, further comprising:
and responding to the picture fusion operation triggered in the interactive area, and displaying a fusion picture of the first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time in a main picture area of the interactive area.
9. The method of claim 1, wherein the displaying, in the interaction area, at least a portion of a first video picture captured in real time by the first camera, at least a portion of a second video picture captured in real time by the second camera, and at least a portion of a video picture captured in real time by at least one camera of the second terminal comprises:
and displaying a fusion picture of a first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time in a main picture area of the interaction area, and displaying a video picture acquired by at least one camera in real time in the second terminal in a secondary picture area of the interaction area, wherein the size of the main picture area is larger than that of the secondary picture area.
10. The method of claim 9, wherein the first video frame is a frame containing a portrait; the main picture area in the interaction area displays a fusion picture of a first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time, and the method comprises the following steps:
and displaying a fusion picture for fusing the portrait in the first video picture acquired by the first camera in real time into the second video picture acquired by the second camera in real time in the main picture area of the interaction area.
11. The method of claim 9, wherein the first video frame and the second video frame are both frames containing human images; the main picture area in the interaction area displays a fusion picture of a first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time, and the method comprises the following steps:
and displaying a fusion picture comprising the portrait in the first video picture acquired by the first camera in real time and the portrait in the second video picture acquired by the second camera in real time in the main picture area of the interaction area.
12. The method of claim 1, wherein the displaying, in the interaction area, at least a portion of a first video picture captured in real time by the first camera, at least a portion of a second video picture captured in real time by the second camera, and at least a portion of a video picture captured in real time by at least one camera of the second terminal comprises:
and displaying a first video picture acquired by the first camera in real time and a video picture acquired by the second terminal in real time in the interactive area, and fusing a fused picture of a second video picture acquired by the second camera in real time.
13. The method according to claim 12, wherein the first video frame is a frame containing a portrait, and the video frame collected by the camera of the second terminal in real time contains a portrait; the interactive area displays a first video picture acquired by the first camera in real time and a video picture acquired by the second terminal in real time, and blends the first video picture and the video picture into a fusion picture of a second video picture acquired by the second camera in real time, and the method comprises the following steps:
and displaying a human image in a first video picture acquired by the first camera in real time and a human image in a video picture acquired by the second terminal in real time in the interaction area, and merging the human image in the first video picture and the human image in the video picture acquired by the second terminal in real time into a merged picture of a second video picture acquired by the second camera in real time.
14. The method according to claim 12, wherein the displaying the interaction area of the video call in response to the triggering operation of the video call between the first terminal and the second terminal comprises:
responding to a triggering operation of a multi-person video call between a first terminal and a plurality of second terminals, and displaying an interaction area of the multi-person video call;
the interactive area displays a first video picture acquired by the first camera in real time and a video picture acquired by the second terminal in real time, and blends the first video picture and the video picture into a fusion picture of a second video picture acquired by the second camera in real time, and the method comprises the following steps:
and displaying a first video picture acquired by the first camera in real time and a video picture acquired by each second terminal in real time in the interaction area, and fusing a fused picture of a second video picture acquired by the second camera in real time.
15. The method according to claim 1, wherein the second terminal is provided with a third camera and a fourth camera which work simultaneously when the video call is made; the third camera and the fourth camera have different visual fields;
the displaying, in the interaction area, at least a part of a first video picture acquired by the first camera in real time, displaying at least a part of a second video picture acquired by the second camera in real time, and displaying at least a part of a video picture acquired by at least one camera in real time at the second terminal includes:
displaying a fusion picture of a third video picture acquired by the third camera in real time and a fourth video picture acquired by the fourth camera in real time in a main picture area of the interaction area;
and displaying a first video picture acquired by the first camera in real time and a second video picture acquired by the second camera in real time in a secondary picture area of the interactive area.
16. The method according to claim 1, wherein the displaying the interaction area of the video call in response to the triggering operation of the video call between the first terminal and the second terminal comprises:
displaying a session page for carrying out instant session with the session object;
and responding to the triggering operation of the video call between the session page and the session object, and entering a video call interaction interface for performing video call interaction between the first terminal and a second terminal corresponding to the session object from the session page.
17. The method of any one of claims 1 to 16, wherein the first camera and the second camera are each one of a rear camera and a front camera of the first terminal.
18. A video call interaction apparatus, the apparatus comprising:
the response module is used for responding to the triggering operation of the video call between the first terminal and the second terminal and displaying the interaction area of the video call, the first terminal is provided with a first camera and a second camera which work simultaneously when the video call is carried out, and the visual fields of the first camera and the second camera are different;
and the display module is used for displaying at least one part of a first video picture acquired by the first camera in real time, displaying at least one part of a second video picture acquired by the second camera in real time and displaying at least one part of a video picture acquired by at least one camera in real time at the second terminal in the interaction area.
19. A computer device comprising a memory and a processor, the memory storing a computer program, characterized in that the processor, when executing the computer program, implements the steps of the method of any of claims 1 to 17.
20. A computer-readable storage medium, in which a computer program is stored which, when being executed by a processor, carries out the steps of the method according to any one of claims 1 to 17.
CN202110178087.1A 2021-02-09 2021-02-09 Video call interaction method, device, computer equipment and storage medium Active CN114915852B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110178087.1A CN114915852B (en) 2021-02-09 2021-02-09 Video call interaction method, device, computer equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110178087.1A CN114915852B (en) 2021-02-09 2021-02-09 Video call interaction method, device, computer equipment and storage medium

Publications (2)

Publication Number Publication Date
CN114915852A true CN114915852A (en) 2022-08-16
CN114915852B CN114915852B (en) 2023-07-25

Family

ID=82761752

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110178087.1A Active CN114915852B (en) 2021-02-09 2021-02-09 Video call interaction method, device, computer equipment and storage medium

Country Status (1)

Country Link
CN (1) CN114915852B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116661656A (en) * 2023-08-02 2023-08-29 安科优选(深圳)技术有限公司 Picture interaction method and shooting display system

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104365088A (en) * 2012-06-08 2015-02-18 三星电子株式会社 Multiple channel communication using multiple cameras
CN105159578A (en) * 2015-08-24 2015-12-16 小米科技有限责任公司 Video display mode switching method and apparatus
CN106303229A (en) * 2016-08-04 2017-01-04 努比亚技术有限公司 A kind of photographic method and device
CN107317993A (en) * 2017-08-08 2017-11-03 维沃移动通信有限公司 A kind of video call method and mobile terminal
CN107528938A (en) * 2017-07-26 2017-12-29 维沃移动通信有限公司 A kind of video call method, terminal and computer-readable recording medium
CN109672843A (en) * 2018-12-25 2019-04-23 努比亚技术有限公司 Video communication method, terminal and computer readable storage medium
CN109788359A (en) * 2017-11-14 2019-05-21 腾讯科技(深圳)有限公司 A kind of video data handling procedure and relevant apparatus
CN110475015A (en) * 2018-09-03 2019-11-19 王闯 A kind of mobile phone front camera and rear camera synchronous working dual display function
CN110809100A (en) * 2019-10-30 2020-02-18 北京字节跳动网络技术有限公司 Video processing method and device, terminal and storage medium
CN112118415A (en) * 2020-09-18 2020-12-22 瑞然(天津)科技有限公司 Remote diagnosis and treatment method and device, patient side terminal and doctor side terminal
CN112153272A (en) * 2019-06-28 2020-12-29 华为技术有限公司 Image shooting method and electronic equipment
CN112312055A (en) * 2019-07-26 2021-02-02 华为技术有限公司 Electronic device and image processing method

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104365088A (en) * 2012-06-08 2015-02-18 三星电子株式会社 Multiple channel communication using multiple cameras
CN105159578A (en) * 2015-08-24 2015-12-16 小米科技有限责任公司 Video display mode switching method and apparatus
CN106303229A (en) * 2016-08-04 2017-01-04 努比亚技术有限公司 A kind of photographic method and device
CN107528938A (en) * 2017-07-26 2017-12-29 维沃移动通信有限公司 A kind of video call method, terminal and computer-readable recording medium
CN107317993A (en) * 2017-08-08 2017-11-03 维沃移动通信有限公司 A kind of video call method and mobile terminal
CN109788359A (en) * 2017-11-14 2019-05-21 腾讯科技(深圳)有限公司 A kind of video data handling procedure and relevant apparatus
CN110475015A (en) * 2018-09-03 2019-11-19 王闯 A kind of mobile phone front camera and rear camera synchronous working dual display function
CN109672843A (en) * 2018-12-25 2019-04-23 努比亚技术有限公司 Video communication method, terminal and computer readable storage medium
CN112153272A (en) * 2019-06-28 2020-12-29 华为技术有限公司 Image shooting method and electronic equipment
CN112312055A (en) * 2019-07-26 2021-02-02 华为技术有限公司 Electronic device and image processing method
CN110809100A (en) * 2019-10-30 2020-02-18 北京字节跳动网络技术有限公司 Video processing method and device, terminal and storage medium
CN112118415A (en) * 2020-09-18 2020-12-22 瑞然(天津)科技有限公司 Remote diagnosis and treatment method and device, patient side terminal and doctor side terminal

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116661656A (en) * 2023-08-02 2023-08-29 安科优选(深圳)技术有限公司 Picture interaction method and shooting display system
CN116661656B (en) * 2023-08-02 2024-03-12 安科优选(深圳)技术有限公司 Picture interaction method and shooting display system

Also Published As

Publication number Publication date
CN114915852B (en) 2023-07-25

Similar Documents

Publication Publication Date Title
CN107528938B (en) Video call method, terminal and computer readable storage medium
US10334002B2 (en) Communication device and method
WO2019096027A1 (en) Communication processing method, terminal, and storage medium
WO2018153267A1 (en) Group video session method and network device
TWI650976B (en) Video communication method and system in instant messaging process
CN116320263A (en) Image processing apparatus, camera apparatus, and image processing method
CN112153400B (en) Live broadcast interaction method and device, electronic equipment and storage medium
CN108933965A (en) screen content sharing method, device and storage medium
CN109150690B (en) Interactive data processing method and device, computer equipment and storage medium
JP7177476B2 (en) Information processing system
JP2023009137A (en) Information processing system
CN106961558B (en) Photographing processing method and device
WO2022134684A1 (en) Interaction method and apparatus based on live streaming application program, and device and storage medium
CN112905074B (en) Interactive interface display method, interactive interface generation method and device and electronic equipment
JP2018515979A (en) Communication processing method and electronic apparatus
CN110168630B (en) Augmented video reality
CN109788359A (en) A kind of video data handling procedure and relevant apparatus
CN111770298A (en) Video call method and device, electronic equipment and storage medium
CN114915852B (en) Video call interaction method, device, computer equipment and storage medium
CN114785977A (en) Controlling video data content using computer vision
CN109039851B (en) Interactive data processing method and device, computer equipment and storage medium
CN116939275A (en) Live virtual resource display method and device, electronic equipment, server and medium
CN112533052A (en) Video sharing method and device, electronic equipment and storage medium
WO2022247766A1 (en) Image processing method and apparatus, and electronic device
CN115314728A (en) Information display method, system, device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
REG Reference to a national code

Ref country code: HK

Ref legal event code: DE

Ref document number: 40074373

Country of ref document: HK

GR01 Patent grant
GR01 Patent grant