CN113112986A - Audio synthesis method, apparatus, device, medium, and program product - Google Patents

Audio synthesis method, apparatus, device, medium, and program product Download PDF

Info

Publication number
CN113112986A
CN113112986A CN202110523959.3A CN202110523959A CN113112986A CN 113112986 A CN113112986 A CN 113112986A CN 202110523959 A CN202110523959 A CN 202110523959A CN 113112986 A CN113112986 A CN 113112986A
Authority
CN
China
Prior art keywords
control
audio
setting
user
sound
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110523959.3A
Other languages
Chinese (zh)
Inventor
陈映宜
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing ByteDance Network Technology Co Ltd
Original Assignee
Beijing ByteDance Network Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing ByteDance Network Technology Co Ltd filed Critical Beijing ByteDance Network Technology Co Ltd
Priority to CN202110523959.3A priority Critical patent/CN113112986A/en
Publication of CN113112986A publication Critical patent/CN113112986A/en
Priority to PCT/CN2022/087491 priority patent/WO2022237464A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/02Methods for producing synthetic speech; Speech synthesisers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • General Engineering & Computer Science (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

According to the audio synthesis method, the audio synthesis device, the audio synthesis equipment, the audio synthesis medium and the program product, when the preset display condition is met, the synthesis control is displayed on the display screen in a first form, then the audio synthesis operation of a user on the synthesis control is received, finally, in the live broadcast equipment with the live broadcast application installed, the audio to be synthesized and the initial audio are synthesized, and the synthesized target audio is used as the input audio of the live broadcast application. The embodiment of the disclosure solves the technical problem that the prior art can only depend on complex professional equipment and professional technical staff team to add the synthetic audio when facing the high-quality live broadcast requirement. The technical effect that the anchor can conveniently and quickly add the synthetic audio in the live audio stream on the live terminal is achieved, the live cost and the technical threshold are reduced, and the use experience of a user is improved.

Description

Audio synthesis method, apparatus, device, medium, and program product
Technical Field
The embodiments of the present disclosure relate to the field of broadcast television and internet live broadcast technologies, and in particular, to an audio synthesis method, apparatus, device, medium, and program product.
Background
With the continuous and deep development of social informatization, the timeliness requirement of people on information acquisition is continuously improved. The live broadcast technology can enable people to break through the limitation of space and obtain various kinds of information of a live broadcast site in real time. In recent years, the emerging live webcasting also becomes a platform for people to show themselves, and great social popularity is obtained. The live broadcast of the mobile phone reduces the technical threshold of live broadcast, so that common people without professional technology can simply carry out live broadcast. However, as the demands of viewers on the content and quality of live broadcast are increasing, a great number of peripheral auxiliary products are derived around mobile live broadcast devices such as mobile phones.
Currently, a high-quality live broadcast of a mobile phone requires hardware devices including: desktop computers, sound cards, several intermediate converters/adapters, mobile phones, headsets, microphones, etc., in addition to the anchor, special technicians are also needed to be equipped to take care of the real-time addition of various live effects.
That is, the prior art has a technical problem that the synthetic audio can only be added by depending on complex professional equipment and a professional technical staff when meeting the requirement of high-quality live broadcasting. This undoubtedly raises the technical threshold of live broadcast, and increases the live broadcast cost of the anchor.
Disclosure of Invention
The embodiment of the disclosure provides an audio synthesis method, an audio synthesis device, an audio synthesis medium and a program product, which are used for solving the technical problem that the existing technology can only depend on complex professional equipment and professional technical staff to add synthesized audio when meeting the requirement of high-quality live broadcasting.
In a first aspect, an embodiment of the present disclosure provides an audio synthesis method, including:
when the preset display condition is met, displaying a synthesis control on the display screen in a first form, wherein the visual influence of the first form on the display content on the display screen meets the preset visual requirement;
receiving audio synthesis operation of a user on a synthesis control;
in the live broadcast equipment provided with the live broadcast application, audio to be synthesized is synthesized with initial audio, and the synthesized target audio is used as input audio of the live broadcast application, wherein the audio to be synthesized is audio indicated by audio synthesis operation, and the initial audio is a sound signal sent by a main broadcast and/or in a live broadcast environment received by the live broadcast equipment through audio acquisition equipment.
In a second aspect, an embodiment of the present disclosure provides an audio synthesizing apparatus, including:
the display module is used for displaying the synthesis control on the display screen in a first form when a preset display condition is met, wherein the visual influence of the first form on the display content on the display screen meets a preset visual requirement;
the receiving module is used for receiving the audio synthesis operation of the user on the synthesis control;
and the processing module is used for synthesizing audio to be synthesized and initial audio in the live broadcast equipment provided with the live broadcast application, and taking the synthesized target audio as input audio of the live broadcast application, wherein the audio to be synthesized is audio indicated by audio synthesis operation, and the initial audio is a sound signal sent by a main broadcast and/or in a live broadcast environment received by the live broadcast equipment through audio acquisition equipment.
In a third aspect, an embodiment of the present disclosure provides an electronic device, including:
at least one processor and memory;
the memory stores a computer program;
the at least one processor executes the computer program stored by the memory such that the at least one processor performs the audio synthesis method as described above in the first aspect and various possible designs of the first aspect.
In a fourth aspect, an embodiment of the present disclosure provides a live all-in-one machine, which includes any one of the possible electronic devices provided in the third aspect.
In a fifth aspect, the embodiments of the present disclosure provide a computer-readable storage medium, in which a computer program is stored, and when the computer program is executed by a processor, the audio synthesis method according to the first aspect and various possible designs of the first aspect are implemented.
In a sixth aspect, the embodiments of the present disclosure provide a computer program product comprising a computer program that, when executed by a processor, implements the audio synthesis method as described above in the first aspect and in various possible designs of the first aspect.
According to the audio synthesis method, the audio synthesis device, the audio synthesis equipment, the audio synthesis medium and the program product, when the preset display condition is met, the synthesis control is displayed on the display screen in a first form, then the audio synthesis operation of a user on the synthesis control is received, finally, in the live broadcast equipment with the live broadcast application installed, the audio to be synthesized and the initial audio are synthesized, and the synthesized target audio is used as the input audio of the live broadcast application. The embodiment of the disclosure solves the technical problem that the prior art can only depend on complex professional equipment and professional technical staff team to add the synthetic audio when facing the high-quality live broadcast requirement. The technical effect that the anchor can conveniently and quickly add the synthetic audio in the live audio stream on the live terminal is achieved, the live cost and the technical threshold are reduced, and the use experience of a user is improved.
Drawings
In order to more clearly illustrate the embodiments of the present disclosure or the technical solutions in the prior art, the drawings needed to be used in the description of the embodiments or the prior art will be briefly introduced below, and it is obvious that the drawings in the following description are some embodiments of the present disclosure, and for those skilled in the art, other drawings can be obtained according to the drawings without inventive exercise.
Fig. 1 is a schematic structural diagram of a live broadcast equipment suite used by an existing indoor anchor according to an embodiment of the present disclosure;
fig. 2 is a first flowchart illustrating an audio synthesizing method according to an embodiment of the disclosure;
fig. 3 is a second flowchart illustrating an audio synthesizing method according to an embodiment of the disclosure;
4a-4c are schematic diagrams of displaying a composite control and user interaction operations on a screen provided by an embodiment of the present disclosure;
fig. 5 is a third schematic flowchart of an audio synthesizing method according to an embodiment of the present disclosure;
6a-6b are schematic diagrams of application scenarios for pop-up window setting through a composition control according to an embodiment of the present disclosure;
fig. 7 is a fourth schematic flowchart of an audio synthesizing method according to an embodiment of the present disclosure;
FIG. 8 is a schematic view of a sound setting interface provided by an embodiment of the present disclosure;
fig. 9 is a scene schematic diagram of an external shortcut key keyboard of a live broadcast integrated device according to an embodiment of the present disclosure;
10a-10c are schematic diagrams illustrating a composition control displaying a playing progress when an audio to be composed is played according to an embodiment of the disclosure;
fig. 11 is a block diagram of an audio synthesizing apparatus according to an embodiment of the present disclosure;
fig. 12 is a schematic structural diagram of an electronic device according to an embodiment of the present disclosure.
Detailed Description
To make the objects, technical solutions and advantages of the embodiments of the present disclosure more clear, the technical solutions of the embodiments of the present disclosure will be described clearly and completely with reference to the drawings in the embodiments of the present disclosure, and it is obvious that the described embodiments are some, but not all embodiments of the present disclosure. All other embodiments, including but not limited to combinations of embodiments, which can be derived by one of ordinary skill in the art from the embodiments disclosed herein without making any inventive step are within the scope of the present disclosure.
The terms "first," "second," "third," "fourth," and the like in the description and in the claims of the present disclosure and in the drawings described above, if any, are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the disclosure described herein are, for example, capable of operation in sequences other than those illustrated or otherwise described herein. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed, but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
With the continuous development of the mobile internet technology, people's daily life is highly informationized, and the live broadcast technology solves an efficient means for people to receive live broadcast information in different places in different geographic spaces. The network live broadcast which is created in recent years increases much fun for entertainment life of people and receives the pursuit of people. A large number of anchor broadcasters join the line and row of network live broadcasting. Live broadcasting is performed through a mobile device, and is favored by many anchor broadcasters due to simplicity. However, as the requirement of viewers on the quality of live broadcasting is continuously increased, the anchor needs to perform audio synthesis when live broadcasting, such as adding audio effects of "enter sound", "laughter in warm spot", "clapping" and the like to the live broadcasting application.
In response to this need, most anchor broadcasters can only purchase a specialized set of live devices, including: the audio synthesis requirement in live broadcasting is realized by professional live broadcasting auxiliary equipment such as a desktop computer, a professional sound card, a plurality of intermediate converters/adapters and the like.
Fig. 1 is a schematic structural diagram of a live device suite used by an existing indoor anchor according to an embodiment of the present disclosure. As shown in fig. 1, a live device suite includes: mobile terminal 101, intermediate adapter 102, intermediate converter 103, sound card 104, desktop computer 105.
Wherein, the mobile terminal 101 is installed with a live application.
The intermediate adapter 102 is used to solve the problem that the charging interface is occupied by the data line and cannot be charged when the mobile terminal 101 is broadcast for a long time.
The intermediate converter 103 is used for transferring special effect data in the desktop computer 105 to the mobile terminal 101, and can provide an ear return function for a client after being connected with an earphone, and can solve the problem that the sound card 104 can only output in a single path and cannot reversely receive audio data output by live application.
The sound card 104 is connected to a microphone to convert an audio analog signal into a digital signal, and the digital signal is input to the desktop computer 105 to perform various sound processing.
However, the mode shown in fig. 1 not only increases the live broadcast cost, but also requires configuring a plurality of auxiliary devices, and most anchor broadcasters do not have professional technology and cannot configure the set of devices by themselves, so that a special technical staff team is required to construct, operate and maintain the set of professional live broadcast devices, which all improve the live broadcast threshold that has been reduced by mobile terminals such as mobile phones.
Moreover, due to the fact that a plurality of devices are connected, the stability of the whole set of live broadcast equipment is easily affected by various factors, the socket connector of any one device is in a problem, or a circuit is broken due to some reason, and the system cannot work.
Based on the technical problems, the invention idea of the application aims to:
the method and the system have the advantages that a live terminal directly enables a main broadcast to quickly synthesize required audio into a live audio stream through a screen or a key during live broadcasting. The method and the system have the advantages that the audio synthesis can be completed by only one small device without a plurality of devices in live broadcasting, professional technical personnel team is not needed to participate in operation and maintenance, and the main broadcasting can add the synthesized audio in real time by one person in live broadcasting.
The audio synthesis method provided by the present application is described in detail below with reference to the accompanying drawings.
Referring to fig. 2, fig. 2 is a first schematic flowchart of an audio synthesizing method according to an embodiment of the present disclosure. The method of the embodiment is applied to a live broadcast integrated terminal, and the audio synthesis method comprises the following steps:
s201: and when the preset display condition is met, displaying the synthesis control on the display screen in a first form.
In this step, the preset display conditions include: the method comprises the steps of detecting at least one of a camera equipment starting instruction, detecting that the live broadcast terminal moves according to a preset movement mode, receiving a preset voice instruction and receiving a preset button switch starting instruction.
In one possible design, the composition control includes a floating window control, the composition control displayed on the display screen in a first modality, including:
and displaying the floating window control at the edge of the display screen according to the preset transparency, the preset size and the preset shape.
In another possible design, the visual impact of the first modality on the content displayed on the display screen satisfies a preset visual requirement, and includes: the shielding rate of the audio synthesis control on the live broadcast interface is smaller than or equal to a preset shielding threshold value, and/or the comprehensive transparency of the audio synthesis control is larger than or equal to a preset transparent threshold value;
and determining the comprehensive transparency according to the transparency of each display element in the audio synthesis control by using a preset model. Specifically, the transparency of each display element on a certain control is multiplied by the corresponding preset weight value, and then the products are added to obtain the comprehensive weight value.
For example: and the transparency of the background color of the button control and the transparency of the text on the button control are multiplied by the corresponding weight values respectively, and the sum of the transparency of the background color and the transparency of the text on the button control is obtained, namely the comprehensive transparency of the button control.
In one possible design, the display screen includes: touch screen and non-touch screen.
In this embodiment, the live broadcast integrated terminal includes at least one display screen thereon. In one possible design, the live broadcast integrated terminal may be connected to the plurality of display screens through an expansion interface, where the expansion interface includes a wireless interface and a wired interface.
S202: and receiving the audio synthesis operation of the user on the synthesis control.
In this step, the audio synthesizing operation on the synthesizing control comprises: single click, double click, rocking-shaking, sliding along a predetermined path, etc
In this embodiment, when the display screen is a touch screen, it is specifically detected in real time that the user approaches or touches the composite control on the touch screen by using a preset operation object.
The preset operation object comprises: any part of the user's body, a touch object (e.g., a touch pen), a thermal sensor (e.g., a heating stick), an electromagnetic sensor (e.g., an electromagnetic sensor pen), an RFID (Radio Frequency Identification) tag key, and the like.
And the touch object directly touches the synthesized control.
The thermal sensor is installed on the live broadcast integrated terminal as long as the thermal sensor is close to the preset sensing area corresponding to the touch screen, and when the thermal sensor senses that the heat source is close to the preset sensing area, the audio synthesis operation of the user is received.
The electromagnetic induction object is installed on the live broadcast integrated terminal as long as being close to a preset induction area corresponding to the touch screen, and when a magnetic field is induced to be close to the preset induction area, the received audio synthesis operation of a user is confirmed.
The RFID label key is installed on a card reader or a card reading circuit on the live broadcast integrated terminal as long as the RFID label key is close to a preset induction area corresponding to the touch screen, and the audio information to be synthesized contained in the RFID label is identified, namely, the audio synthesis operation is received.
S203: in the live broadcast equipment with the live broadcast application, synthesizing audio to be synthesized and initial audio, and taking the synthesized target audio as input audio of the live broadcast application.
In this step, the audio to be synthesized is the audio indicated by the audio synthesis operation, and the initial audio is a sound signal sent by the anchor and/or in a live environment received by the live broadcast device through the audio acquisition device.
And after the target audio is coded by the live broadcast application, the target audio is transmitted to a live broadcast platform server through a network and then is sent to each audience by the live broadcast platform server.
Note that the audio to be synthesized includes: laughter, sound of going/coming, bird scream (e.g., crow scream), clapping, earring, gunshot, and the like.
Note that, the live device in which the live application is installed includes: live broadcast camera equipment (a professional camera for live broadcast of news media), a personal computer (pc), a mobile terminal (a mobile phone and a tablet computer), network live broadcast integrated equipment, live broadcast carriers (such as live broadcast vehicles and live broadcast shopping carts) and the like.
According to the audio synthesis method provided by the embodiment of the disclosure, firstly, when a preset display condition is met, a synthesis control is displayed on a display screen in a first form, then, an audio synthesis operation of a user on the synthesis control is received, finally, in a live broadcast device with a live broadcast application installed, audio to be synthesized and initial audio are synthesized, and synthesized target audio is used as input audio of the live broadcast application. The embodiment of the disclosure solves the technical problem that the prior art can only depend on complex professional equipment and professional technical staff team to add the synthetic audio when facing the high-quality live broadcast requirement. The technical effect that the anchor can conveniently and quickly add the synthetic audio in the live audio stream on the live terminal is achieved, the live cost and the technical threshold are reduced, and the use experience of a user is improved.
Referring to fig. 3, fig. 3 is a schematic flowchart illustrating a second audio synthesizing method according to an embodiment of the disclosure. The method of the embodiment is applied to a live broadcast integrated terminal, and the audio synthesis method comprises the following steps:
s301: and when the camera equipment starting instruction is detected, displaying the synthesis control on the display screen in a first form.
In this step, the visual impact of the first modality on the content displayed on the display screen meets a preset visual requirement.
In this embodiment, after the user starts the live application on the live integrated terminal, the user clicks to start live broadcasting, and then the live application starts the camera on the live integrated terminal. And when the live broadcast integrated terminal detects that the camera is opened, the composite control is displayed at the edge of the display screen.
In one possible design, the composite control includes a floating window control, and displaying the composite control on the display screen in a first modality includes:
and displaying the floating window control at the edge of the display screen according to the preset transparency, the preset size and the preset shape.
Fig. 4a-4c are schematic diagrams illustrating the display of a composite control on a screen and the user interaction operation provided by the embodiment of the disclosure. As shown in fig. 4a, the composite control is a floating window 41, after it is detected that the camera is opened, the composite control is displayed in the form of a small floating window, and the floating window 41 is set to have a background color transparency of more than 50%, so that the visual impact of the floating window 41 on other contents on the live-cast interface is reduced to meet the preset requirement.
In one possible design, the composite control includes a plurality of child controls, as shown in FIG. 4b, the composite control includes: a popup window setting sub-control 411, a plurality of audio sub-controls 412 corresponding to the audio to be synthesized, and a stowing control 413. And after the camera is detected to be opened, directly displaying all the sub-controls.
In another possible design, the synthesized control includes multiple sub-controls and, however, the first modality does not show all the sub-controls, as shown in fig. 4c, the first modality only displays the audio sub-control 412, the pop-up window setting sub-control 411 and the expansion control 414, which partially correspond to the audio to be synthesized.
S302: and acquiring the state setting operation of the user on the synthesis control.
In this step, the state setting operation includes: and changing the display form of the composite control and the position of the composite control.
The operation of changing the display form of the synthesis control comprises the following steps: expand/collapse operations, operations to swap the order or position of sub-controls, compress, enlarge, etc.
The operation of changing the position of the synthesized control comprises the following steps: and dragging the synthesized control to other positions of the display screen.
It should be noted that, in a possible design, the synthesized control always needs to be displayed close to the edge, and after the user drags the synthesized control to be released, the distances from the synthesized control to the four edges of the display screen are calculated, and the synthesized control is adsorbed to the vicinity of the edge with the minimum distance to be displayed.
S303: and switching the form and/or the display position of the synthesized control according to the state setting operation.
For example, as shown in fig. 4a-4c, when the user clicks the expand control 414 in fig. 4c, the composite control is transformed from the configuration of fig. 4c to the configuration of fig. 4b, and when the user clicks the retract control 413, the composite control is transformed from the configuration of fig. 4b to the configuration of fig. 4 a.
S304: and detecting that the user approaches or touches the synthesized control on the touch screen by using a preset operation object in real time.
In this embodiment, the display screen includes at least one touch screen, and the preset operation object includes: any part of the body of the user, a touch object (such as a touch pen), a heat sensing object (such as a heating rod), an electromagnetic induction object (an electromagnetic induction pen), an RFID tag key and the like.
In a possible design, the synthesis control includes a plurality of sub-controls corresponding to the audio to be synthesized, as shown in fig. 4b, this step specifically includes:
if the preset operation object is detected to be close to or touch the sub-control, switching the sub-control to a selected form;
and determining a synthetic operation instruction corresponding to the child control.
S305: in the live broadcast equipment with the live broadcast application, synthesizing audio to be synthesized and initial audio, and taking the synthesized target audio as input audio of the live broadcast application.
The specific principle and noun explanation of this step refer to S203, which is not described herein again.
According to the audio synthesis method provided by the embodiment of the disclosure, firstly, when a preset display condition is met, a synthesis control is displayed on a display screen in a first form, then, an audio synthesis operation of a user on the synthesis control is received, finally, in a live broadcast device with a live broadcast application installed, audio to be synthesized and initial audio are synthesized, and synthesized target audio is used as input audio of the live broadcast application. The embodiment of the disclosure solves the technical problem that the prior art can only depend on complex professional equipment and professional technical staff team to add the synthetic audio when facing the high-quality live broadcast requirement. The technical effect that the anchor can conveniently and quickly add the synthetic audio in the live audio stream on the live terminal is achieved, the live cost and the technical threshold are reduced, and the use experience of a user is improved.
With reference to fig. 5, on the basis of the above two embodiments, the audio synthesis method provided by the present disclosure can further set various audio control parameters of the live broadcast integrated device in a form of pop-up window setting when the synthesis control is live broadcast, which is exemplified below by a specific embodiment.
Fig. 5 is a third schematic flowchart of an audio synthesizing method according to an embodiment of the present disclosure. The method of the embodiment is applied to a live broadcast integrated terminal, and the audio synthesis method comprises the following steps:
s501: and acquiring popup setting operation of the user on the synthesis control.
In this embodiment, as shown in the popup setting sub-control 411 shown in fig. 4b or fig. 4c, after the user clicks the control, the live broadcast integrated device acquires a control instruction of the popup setting operation.
S502: and responding to the popup window setting operation, and superposing and displaying a setting window on the display screen.
In this embodiment, the setting window includes a sound setting tab, and in a display area corresponding to the sound setting tab, the setting window includes: at least one of an ear return switch control, a dodging switch control, a loudspeaker selection control, a microphone volume setting control, a sound effect selection control, a sound effect volume control and the like.
S503: and responding to the operation instruction of the user to each control, and correspondingly setting each audio control parameter.
In this embodiment, specifically:
and S5031, setting a control parameter corresponding to the on-off state of the ear return effect by using a switching instruction of the ear return switch control, and correspondingly switching the on-off display state of the ear return switch control.
And/or the presence of a gas in the gas,
s5032, responding to a switching instruction of a user to the dodging switch control, setting a control parameter corresponding to the dodging effect switch state, and correspondingly switching the switch display state of the dodging switch control.
And/or the presence of a gas in the gas,
s5033, setting the working state of each speaker in the live broadcasting device in response to the selection instruction of the user for the speaker selection control, and correspondingly switching the speaker selection state of the speaker selection control.
And/or the presence of a gas in the gas,
s5034, in response to a selection instruction of the microphone selection control from the user, setting a working state of each microphone in the live broadcast device, and correspondingly switching a microphone selection state of the microphone selection control.
And/or the presence of a gas in the gas,
s5035, in response to a first volume setting operation of the user on the microphone connecting volume setting control, setting a volume control parameter of the opposite party during microphone connecting, and correspondingly changing the display state of the microphone connecting volume setting control;
and/or the presence of a gas in the gas,
s5036, responding to the selection instruction of the user to the sound effect selection control, setting a sound effect control parameter during target audio synthesis, and correspondingly switching the sound effect selection state of the sound effect selection control.
And/or the presence of a gas in the gas,
s5037, in response to the second volume setting operation of the sound effect volume control by the user, setting a volume control parameter corresponding to the sound effect, and correspondingly changing the display state of the sound effect volume control.
In one possible design, the setup window includes a shortcut setup tab that includes: if the shortcut key array and the shortcut key array to be selected are selected, the method further comprises the following steps:
s5038, in response to the delete instruction of the user for the selected shortcut key array, updating the display status of the selected shortcut key array and the shortcut key array to be selected.
And/or the presence of a gas in the gas,
s5039, in response to the selection instruction of the user for the shortcut key array to be selected, updating the selected shortcut key array and the display status of the shortcut key array to be selected.
It should be noted that, the sequence of S5031 to S5039 is not required, and the user can set several parameters at the same time.
Fig. 6a to 6b are schematic diagrams of application scenarios for pop-up window setting through a composite control according to an embodiment of the present disclosure. One implementation of the sound setting tab is shown in fig. 6a, after the user clicks the pop-up window setting sub-control 411, a setting window pops up below the display screen, and on the sound setting tab 421 of the setting window, the user sets to turn on or turn off the ear return effect or the dodging effect by toggling the switch of the ear return or dodging control. The speaker or microphone is selected by clicking the corresponding control. Various volume sizes are set by dialing the volume bar. And switching the sound effect mode by clicking the corresponding sound effect control.
It should be noted that the selection control can be selected by checking or by pulling down the selection control, in addition to the single button control for each option in fig. 6 a. The volume bar may also be provided in the form of an open arc, closed ring, knob, etc.
One implementation of the sound effect shortcut key tab 422 is shown in fig. 6b, where a box shows a shortcut key corresponding to a selected and displayed audio to be synthesized, in this embodiment, at most 8 shortcut keys are allowed to be added, and a "trash box" -shaped button on the right side of the selected shortcut key, i.e., a delete button, is clicked, so that the shortcut key may be set not to be displayed.
The lower part is arranged with a plurality of audio options to be synthesized in an array form, the audio to be synthesized can be added into the upper shortcut key box in the display state by clicking the icon of the plus sign on the right side, and after the audio is selected, the background color of the selection control is switched, and the option is indicated to be selected by changing the plus sign into the V sign.
In addition, the audio option to be synthesized can also be downloaded from the cloud server or the internet server through the network interface, for example, the audio to be synthesized corresponding to "whistle" or "scream" in fig. 6b is not pre-stored in the live broadcast integrated terminal, and only the download button on the right side of the control needs to be clicked, so that the audio can be downloaded from the server to the local storage.
It should be noted that, in a possible design, as long as the list of audio options to be synthesized is updated in the cloud server or the internet server, the audio shortcut tab 422 is also updated synchronously, and the user can click the download button to download after seeing the new option. Of course, the user may also set automatic downloading, or batch downloading, to download all the updated audio to be synthesized to the local storage at one time.
It should be noted that the setting of the sound effect shortcut key may also be selected by using a drop-down box, or other selection manners, and those skilled in the art may select the sound effect shortcut key according to actual situations, which is not limited in this application.
According to the audio synthesis method provided by the embodiment of the disclosure, on the basis of the embodiments shown in fig. 2 and fig. 3, the popup setting operation of the user on the synthesis control is acquired, then the setting window is displayed on the display screen in an overlapping manner in response to the popup setting operation, and then the audio control parameters are set correspondingly in response to the operation instruction of the user on each control, so that the technical problem that the setting parameters related to each sound cannot be changed simply, conveniently and quickly by the anchor in live broadcasting and the assistance of professional technicians is required is solved. The shortcut key of the sound setting and synthesis control is simple and convenient to set, the direction of sight does not need to be changed, the live broadcast effect is influenced, the live broadcast setting does not need to be suspended, and the use experience of a user on the live broadcast integrated terminal is improved.
On the basis of the three embodiments, the audio synthesis method provided by the present disclosure can also be used for setting the display state of the synthesis control and each sound control parameter on the live broadcast integrated terminal through a special sound setting interface before live broadcast, and the following describes specific steps with specific embodiments.
Fig. 7 is a fourth schematic flowchart of an audio synthesizing method according to an embodiment of the present disclosure. The method of the embodiment is applied to a live broadcast integrated terminal, and the audio synthesis method comprises the following steps:
s701: and acquiring the sound setting operation of the user before the camera is not started or the live broadcast is started.
In this step, the live broadcast integrated terminal detects that the user clicks a sound setting button, or performs an opening operation of a sound setting interface through a preset setting instruction when the live broadcast application starts.
S702: in response to a sound setting operation by a user, a sound setting interface is displayed on the display screen.
In this embodiment, the sound setting interface includes: at least one of an ear return switch control, a dodging switch control, a loudspeaker selection control, a microphone volume setting control, a sound effect selection control, a sound effect volume control, a synthesis control (such as a floating window) display switch control, a selected shortcut key array, a shortcut key array to be selected and the like.
S703: and correspondingly setting various audio control parameters according to the setting instruction of the user on each parameter control on the sound setting interface.
In this embodiment, specifically:
s7031, setting a control parameter corresponding to the ear return effect switch state according to a switch instruction of the ear return switch control, and correspondingly switching the switch display state of the ear return switch control.
And/or the presence of a gas in the gas,
s7032, responding to a switch instruction of a user to the dodging switch control, setting control parameters corresponding to the dodging effect switch state, and correspondingly switching the switch display state of the dodging switch control.
And/or the presence of a gas in the gas,
s7033, responding to a selection instruction of the user to the loudspeaker selection control, setting the working state of each loudspeaker in the live broadcast equipment, and correspondingly switching the loudspeaker selection state of the loudspeaker selection control.
And/or the presence of a gas in the gas,
s7034, setting the operating states of the microphones in the live broadcast device in response to a selection instruction of the microphone selection control from the user, and correspondingly switching the states of the microphones in the microphone selection control.
And/or the presence of a gas in the gas,
s7035, responding to a first volume setting operation of a user on the microphone connecting volume setting control, setting a volume control parameter of the opposite party during microphone connecting, and correspondingly changing the display state of the microphone connecting volume setting control;
and/or the presence of a gas in the gas,
s7036, responding to a selection instruction of a user to the sound effect selection control, setting sound effect control parameters during target audio synthesis, and correspondingly switching the sound effect selection state of the sound effect selection control.
And/or the presence of a gas in the gas,
s7037, responding to a second volume setting operation of the user on the sound effect volume control, setting a volume size control parameter corresponding to the sound effect, and correspondingly changing the display state of the sound effect volume control.
And/or the presence of a gas in the gas,
s7038, in response to a deletion instruction of the selected shortcut key array by the user, updating the display states of the selected shortcut key array and the shortcut key array to be selected.
And/or the presence of a gas in the gas,
s7039, in response to a selection instruction of the user on the shortcut key array to be selected, updating the selected shortcut key array and the display state of the shortcut key array to be selected.
And/or the presence of a gas in the gas,
s70310, responding to a switch instruction of a user to display a switch on a composite control (e.g., a floating window), and correspondingly setting whether the composite control is displayed on a display screen.
It should be noted that S7031-S70310 do not have the requirement of a sequential order, and a user may set several parameters at the same time.
In particular, one possible implementation may refer to fig. 8.
Fig. 8 is a schematic view of a sound setting interface according to an embodiment of the disclosure. As shown in fig. 8, a user can select a speaker carried by a "wired headset" or a live broadcast integrated device, i.e., "local speaker", through a speaker selection control on a sound setting interface; through the microphone selection control, a user can select 'wired headset', 'local microphone', 'Cannon microphone'; through the ear return switch control, a user turns on or off the ear return effect by shifting the switch control to the right/left, and the ear return switch control also comprises a prompt text, if the 'local loudspeaker does not support the ear return function', the user is prompted that the ear return function and the loudspeaker selection directly have associated setting, so that the influence on the use effect caused by improper user parameter selection is avoided; similarly, the evasion switch control and the floating window switch control also have similar operations, and the specific contents refer to fig. 8.
The aesthetic sound option in fig. 8 is a selection control for the sound beautifying effect, and by clicking the right expansion icon ">", the sound list can be expanded for the user to select the corresponding sound beautifying effect, in each embodiment of the present disclosure, the sound beautifying effect includes: sound characteristics such as primary sound, KTV, magnetism, old record, air crash, theater, every kind of audio is corresponding a plurality of audio setting parameters, through selecting specific audio, can set up each parameter fast, has reduced the professional threshold that the audio set up for the anchor who does not possess professional technique also can carry out simple and convenient audio setting.
The lowermost part of fig. 8 is a floating window, that is, a shortcut key selection array of a synthesis control, and a user only needs to select each button control, and then sets up a shortcut key for displaying a corresponding audio to be synthesized, so that the user can directly perform audio synthesis in a live broadcast interface through the shortcut key after the live broadcast starts.
On the basis of the embodiments shown in fig. 2, fig. 3, and fig. 5, the audio synthesis method provided by the embodiment of the present disclosure obtains the sound setting operation of the user before the camera is not turned on or the live broadcast is turned on, then displays the sound setting interface on the display screen in response to the sound setting operation of the user, and sets the audio control parameters correspondingly according to the setting instruction of the user on each parameter control on the sound setting interface, thereby solving the technical problem that the anchor can not change the setting parameters related to each sound easily, conveniently, and quickly, and needs the assistance of a professional. The shortcut key setting of the sound setting and the synthesis control is simple and convenient, and the use experience of the user on the live broadcast integrated terminal is improved.
In a possible design, on the basis of the embodiments of the present disclosure, the live all-in-one machine, that is, the live broadcast integrated device, is connected to at least one shortcut keyboard through an external interface, step S202: receiving audio synthesis operation of a user on a synthesis control, including;
receiving an operation instruction sent by a shortcut keyboard, wherein the operation instruction is determined by the shortcut keyboard in response to the operation of a user;
and determining the corresponding audio to be synthesized according to the operation instruction.
It should be noted that the external interface includes: a wireless interface and a wired interface.
Fig. 9 is a scene schematic diagram of an external shortcut key keyboard of a live broadcast integrated device according to an embodiment of the present disclosure. As shown in fig. 9, the live all-in-one machine 91 is connected to a shortcut keyboard 92 through a wired interface or a wireless interface, and is configured to implement that a plurality of sub-controls in a synthesis control 911 of an audio synthesis method according to each embodiment of the present disclosure are in one-to-one correspondence with keys on the keyboard.
In a possible design, the corresponding relation between each key on the shortcut keyboard and each sub-control on the display screen can be set arbitrarily. The shortcut key keyboard includes: mechanical button type keyboards, touch screens, etc. The user clicks the shortcut key, namely, clicks the corresponding sub-control on the display screen, and simultaneously, the sub-control is switched and displayed to be in a selected state.
It should be further noted that, after the sub-control is clicked and selected, the playing progress of the audio to be synthesized may be synchronously displayed in the form of a bar-shaped progress bar or a sector-shaped progress chart.
Fig. 10a to 10c are schematic diagrams illustrating a composition control displaying a playing progress when an audio to be synthesized is played according to an embodiment of the present disclosure. As shown in fig. 10a, after the user clicks the "laugh 1" sub-control, a play progress bar for playing the audio corresponding to "laugh 1" is displayed on the sub-control.
As shown in fig. 10b, after the user clicks the "laugh 1" sub-control, the background color or the background pattern of the sub-control displays the audio playing progress bar corresponding to the "laugh 1" in a rotating and gradually changing manner.
As shown in fig. 10c, after the user clicks the "laugh 1" child control, the background color or the background pattern of the child control displays the audio playing progress bar corresponding to the "laugh 1" in a gradual change form in any one of left to right, right to left, top to bottom, bottom to top, left to top, right to bottom to left, and the like.
Therefore, the anchor can know the playing progress of the audio to be synthesized, further, the playing can be finished by clicking the sub-control again, or other sub-controls are clicked to play a plurality of audio to be synthesized in a mixed mode, the operation flexibility of the live broadcast audio synthesis is improved, the live audio effect during live broadcast is improved, and the user experience is improved.
Fig. 11 is a block diagram of an audio synthesis apparatus 1100 according to an embodiment of the present disclosure. For ease of illustration, only portions that are relevant to embodiments of the present disclosure are shown. Referring to fig. 11, the apparatus includes:
the display module 1101 is configured to display a composite control on a display screen in a first form when a preset display condition is met, where a visual influence of the first form on a content displayed on the display screen meets a preset visual requirement;
a receiving module 1102, configured to receive an audio synthesis operation performed on the synthesis control by a user;
the processing module 1103 is configured to, in a live device in which a live application is installed, synthesize an audio to be synthesized and an initial audio, and use a synthesized target audio as an input audio of the live application, where the audio to be synthesized is an audio indicated by the audio synthesis operation, and the initial audio is a sound signal sent by a main broadcast and/or in a live environment received by the live device through an audio acquisition device.
In an embodiment of the present disclosure, the display screen includes a touch screen, and the receiving module 1102 is configured to detect in real time that the user approaches or touches the composite control on the touch screen by using a preset operation object.
In an embodiment of the present disclosure, the synthesis control includes a sub-control corresponding to the audio to be synthesized, and the receiving module 1102 is specifically configured to:
if the preset operation object is detected to be close to or touch the sub-control, switching the sub-control to a selected form;
and determining a synthetic operation instruction corresponding to the child control.
In one embodiment of the present disclosure, the preset display condition includes: the method comprises the steps of detecting at least one of a camera equipment starting instruction, detecting that the live broadcast terminal moves according to a preset movement mode, receiving a preset voice instruction and receiving a preset button switch starting instruction.
In an embodiment of the present disclosure, the synthesized control includes a floating window control, and the display module 1101 is configured to display the floating window control on an edge of a display screen according to a preset transparency, a preset size, and a preset shape.
In an embodiment of the present disclosure, the receiving module 1102 is further configured to obtain a state setting operation of the user on the composite control;
the display module 1101 is further configured to switch a form and/or a display position of the synthesized control according to the state setting operation.
In an embodiment of the present disclosure, the display module 1101 is further configured to, in response to a popup setting operation performed by the user on the composition control, superimpose and display a setting window on a display screen;
the processing module 1103 is further configured to set each audio control parameter correspondingly according to a setting instruction of the user for each parameter control on the setting window.
In one embodiment of the present disclosure, the setup window includes a sound setup tab, the sound setup tab including: the ear return switch control and/or the evasion switch control, the processing module 1103 is specifically configured to:
responding to a switch instruction of the user to the ear return switch control, setting a control parameter corresponding to an ear return effect switch state, and correspondingly switching a switch display state of the ear return switch control;
and/or the presence of a gas in the gas,
responding to the switching instruction of the user to the dodging switch control, setting control parameters corresponding to the dodging effect switching state, and correspondingly switching the switching display state of the dodging switch control.
In one embodiment of the present disclosure, the setup window includes a sound setup tab, the sound setup tab including: the processing module 1103 is specifically configured to:
and responding to a selection instruction of the user on the loudspeaker selection control, setting the working state of each loudspeaker in the live broadcast equipment, and correspondingly switching the loudspeaker selection state of the loudspeaker selection control.
In one embodiment of the present disclosure, the setup window includes a sound setup tab, the sound setup tab including: the processing module 1103 is specifically configured to:
and responding to a selection instruction of the user on the microphone selection control, setting the working state of each microphone in the live broadcast equipment, and correspondingly switching the microphone selection state of the microphone selection control.
In one embodiment of the present disclosure, the setup window includes a sound setup tab, the sound setup tab including: the processing module 1103 is specifically configured to:
responding to the first volume setting operation of the user on the microphone connecting volume setting control, setting a volume control parameter of the opposite side during microphone connecting, and correspondingly changing the display state of the microphone connecting volume setting control.
In one embodiment of the present disclosure, the setup window includes a sound setup tab, the sound setup tab including: an audio selection control, and/or an audio volume control, the processing module 1103 is specifically configured to:
responding to a selection instruction of the user to the sound effect selection control, setting a sound effect control parameter during target audio synthesis, and correspondingly switching a sound effect selection state of the sound effect selection control;
and/or the presence of a gas in the gas,
responding to the second volume setting operation of the sound effect volume control, setting a volume size control parameter corresponding to the sound effect, and correspondingly changing the display state of the sound effect volume control.
In one embodiment of the present disclosure, the setting window includes a shortcut setting tab, and the shortcut setting tab includes: the processing module 1103 is specifically configured to, after the shortcut key array and the shortcut key array to be selected are selected:
responding to a deleting instruction of the user on the selected shortcut key array, and updating the display states of the selected shortcut key array and the shortcut key array to be selected;
and/or the presence of a gas in the gas,
and updating the selected shortcut key array and the display state of the shortcut key array to be selected in response to the selection instruction of the user on the shortcut key array to be selected.
In an embodiment of the present disclosure, before displaying the synthesized control on the display screen in the first form when the preset display condition is met, the display module 1101 is further configured to display a sound setting interface on the display screen in response to a sound setting operation of the user;
the processing module 1103 is further configured to set, according to a setting instruction of the user on each parameter control on the sound setting interface, each audio control parameter correspondingly.
In one embodiment of the present disclosure, the sound setting interface includes: an ear return switch control, and/or an evasion switch control, the processing module 1103 is further configured to:
responding to a switch instruction of the user to the ear return switch control, setting a control parameter corresponding to an ear return effect switch state, and correspondingly switching a switch display state of the ear return switch control;
and/or the presence of a gas in the gas,
responding to the switching instruction of the user to the dodging switch control, setting control parameters corresponding to the dodging effect switching state, and correspondingly switching the switching display state of the dodging switch control.
In one embodiment of the present disclosure, the sound setting interface includes: a speaker selection control, the processing module 1103 further configured to:
and responding to a selection instruction of the user on the loudspeaker selection control, setting the working state of each loudspeaker in the live broadcast equipment, and correspondingly switching the loudspeaker selection state of the loudspeaker selection control.
In one embodiment of the present disclosure, the sound setting interface includes: a microphone selection control, the processing module 1103 is further configured to:
and responding to a selection instruction of the user on the microphone selection control, setting the working state of each microphone in the live broadcast equipment, and correspondingly switching the microphone selection state of the microphone selection control.
In one embodiment of the present disclosure, the sound setting interface includes: the processing module 1103 is further configured to:
responding to the first volume setting operation of the user on the microphone connecting volume setting control, setting a volume control parameter of the opposite side during microphone connecting, and correspondingly changing the display state of the microphone connecting volume setting control.
In one embodiment of the present disclosure, the sound setting interface includes: an audio selection control, and/or an audio volume control, the processing module 1103 is further configured to:
responding to a selection instruction of the user to the sound effect selection control, setting a sound effect control parameter during target audio synthesis, and correspondingly switching a sound effect selection state of the sound effect selection control;
and/or the presence of a gas in the gas,
responding to the second volume setting operation of the sound effect volume control, setting a volume size control parameter corresponding to the sound effect, and correspondingly changing the display state of the sound effect volume control.
In one embodiment of the present disclosure, the sound setting interface includes: the processing module 1103 is further configured to:
responding to a deleting instruction of the user on the selected shortcut key array, and updating the display states of the selected shortcut key array and the shortcut key array to be selected;
and/or the presence of a gas in the gas,
and updating the selected shortcut key array and the display state of the shortcut key array to be selected in response to the selection instruction of the user on the shortcut key array to be selected.
In one embodiment of the present disclosure, the sound setting interface includes: a composition control display switch, and the processing module 1103 is further configured to:
and responding to a switch instruction of the user to the display switch of the synthesis control, and correspondingly setting whether the synthesis control is displayed on a display screen.
In an embodiment of the present disclosure, the audio synthesis apparatus 1100 is connected to at least one shortcut keyboard through an external interface, and the receiving of the audio synthesis operation of the user on the synthesis control includes;
the receiving module 1102 is further configured to receive an operation instruction sent by the shortcut keyboard, where the operation instruction is determined by the shortcut keyboard in response to an operation of the user;
the processing module 1103 is further configured to determine, according to the operation instruction, the corresponding audio to be synthesized.
The apparatus 1100 provided in this embodiment may be configured to perform steps related to the user side in the technical solution of the foregoing method embodiment, and the implementation principle and the technical effect are similar, which are not described herein again.
Referring to fig. 12, which shows a schematic structural diagram of an electronic device suitable for implementing the embodiment of the present disclosure, the electronic device 1200 may be a terminal device or a server. Among them, the terminal Device may include, but is not limited to, a mobile terminal such as a mobile phone, a notebook computer, a Digital broadcast receiver, a Personal Digital Assistant (PDA), a tablet computer (PAD), a Portable Multimedia Player (PMP), a car terminal (e.g., car navigation terminal), etc., and a fixed terminal such as a Digital TV, a desktop computer, etc. The electronic device shown in fig. 12 is only an example, and should not bring any limitation to the functions and the scope of use of the embodiments of the present disclosure.
As shown in fig. 12, the electronic device 1200 may include a processing means (e.g., a central processing unit, a graphics processor, etc.) 1201 which may perform various appropriate actions and processes according to a program stored in a Read Only Memory (ROM) 1202 or a program loaded from a storage device 1208 into a Random Access Memory (RAM) 1203. In the RAM 1203, various programs and data necessary for the operation of the electronic apparatus 1200 are also stored. The processing apparatus 1201, the ROM 1202, and the RAM 1203 are connected to each other by a bus 1204. An input/output (I/O) interface 1205 is also connected to bus 1204.
Generally, the following devices may be connected to the I/O interface 1205: input devices 12012 including, for example, a touch screen, a touch pad, a keyboard, a mouse, a camera, a microphone, an accelerometer, a gyroscope, and the like; an output device 1207 including, for example, a Liquid Crystal Display (LCD), a speaker, a vibrator, and the like; storage devices 1208 including, for example, magnetic tape, hard disk, etc.; and a communication device 1209. The communication device 1209 may allow the electronic apparatus 1200 to communicate wirelessly or by wire with other apparatuses to exchange data. While fig. 12 illustrates an electronic device 1200 having various means, it is to be understood that not all illustrated means are required to be implemented or provided. More or fewer devices may alternatively be implemented or provided.
In particular, according to an embodiment of the present disclosure, the processes described above with reference to the flowcharts may be implemented as computer software programs. For example, embodiments of the present disclosure include a computer program product comprising a computer program embodied on a computer readable medium, the computer program comprising program code for performing the method illustrated in the flow chart. In such an embodiment, the computer program may be downloaded and installed from a network through the communication device 1209, or installed from the storage device 1208, or installed from the ROM 1202. The computer program, when executed by the processing apparatus 1201, performs the above-described functions defined in the methods of the embodiments of the present disclosure.
It should be noted that the computer readable medium in the present disclosure can be a computer readable signal medium or a computer readable storage medium or any combination of the two. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples of the computer readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the present disclosure, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In contrast, in the present disclosure, a computer readable signal medium may comprise a propagated data signal with computer readable program code embodied therein, either in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: electrical wires, optical cables, RF (radio frequency), etc., or any suitable combination of the foregoing.
The computer readable medium may be embodied in the electronic device; or may exist separately without being assembled into the electronic device.
The computer readable medium carries one or more programs which, when executed by the electronic device, cause the electronic device to perform the methods shown in the above embodiments.
Computer program code for carrying out operations for aspects of the present disclosure may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C + +, and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any type of Network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet service provider).
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present disclosure. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The units described in the embodiments of the present disclosure may be implemented by software or hardware. Where the name of a unit does not in some cases constitute a limitation of the unit itself, for example, the first retrieving unit may also be described as a "unit for retrieving at least two internet protocol addresses".
The functions described herein above may be performed, at least in part, by one or more hardware logic components. For example, without limitation, exemplary types of hardware logic components that may be used include: field Programmable Gate Arrays (FPGAs), Application Specific Integrated Circuits (ASICs), Application Specific Standard Products (ASSPs), systems on a chip (SOCs), Complex Programmable Logic Devices (CPLDs), and the like.
In the context of this disclosure, a machine-readable medium may be a tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. The machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. A machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a machine-readable storage medium would include an electrical connection based on one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
A computer program product is also provided in the disclosed embodiments, comprising a computer program that, when executed by a processor, implements the methods in the embodiments described above.
The embodiment of the disclosure also provides a live all-in-one machine or live all-in-one device, which comprises the electronic device corresponding to fig. 12. It should be further noted that the control circuit of the live broadcast all-in-one machine includes: the device comprises a main control module and an audio processing module, wherein the main control module is provided with a live application, and the audio processing module is used for synthesizing audio to be synthesized into initial audio to form target audio. The audio to be synthesized can be stored in the audio processing module, the main control module and the memory of the live broadcast all-in-one machine.
In summary, the following steps:
in a first aspect, according to one or more embodiments of the present disclosure, there is provided an audio synthesis method including:
when a preset display condition is met, displaying a synthesis control on a display screen in a first form, wherein the visual influence of the first form on the display content on the display screen meets a preset visual requirement;
receiving an audio synthesis operation of a user on the synthesis control;
in the live broadcast equipment provided with the live broadcast application, synthesizing audio to be synthesized and initial audio, and taking the synthesized target audio as input audio of the live broadcast application, wherein the audio to be synthesized is audio indicated by audio synthesis operation, and the initial audio is a sound signal sent by a main broadcast and/or in a live broadcast environment received by the live broadcast equipment through audio acquisition equipment.
According to one or more embodiments of the present disclosure, the display screen includes a touch screen, and the receiving an audio synthesis operation of a user on the synthesis control includes:
and detecting that the user approaches or touches the synthetic control on the touch screen by using a preset operation object in real time.
According to one or more embodiments of the present disclosure, the synthesizing control includes a sub-control corresponding to the audio to be synthesized, and the detecting, in real time, that the user approaches or touches the synthesizing control on the touch screen by using a preset operator includes:
if the preset operation object is detected to be close to or touch the sub-control, switching the sub-control to a selected form;
and determining a synthetic operation instruction corresponding to the child control.
According to one or more embodiments of the present disclosure, the preset display condition includes: the method comprises the steps of detecting at least one of a camera equipment starting instruction, detecting that the live broadcast terminal moves according to a preset movement mode, receiving a preset voice instruction and receiving a preset button switch starting instruction.
According to one or more embodiments of the present disclosure, the composition control comprises a floating window control, the displaying the composition control on the display screen in the first modality comprises:
and displaying the floating window control at the edge of the display screen according to the preset transparency, the preset size and the preset shape.
In accordance with one or more embodiments of the present disclosure, the method further comprises:
acquiring the state setting operation of the user on the synthesis control;
and switching the form and/or the display position of the synthesized control according to the state setting operation.
In accordance with one or more embodiments of the present disclosure, the method further comprises:
responding to popup setting operation of the user on the synthesis control, and overlaying and displaying a setting window on a display screen;
and correspondingly setting various audio control parameters according to the setting instruction of the user on each parameter control on the setting window.
According to one or more embodiments of the present disclosure, the setup window includes a sound setup tab including: the ear return switch control and/or the dodging switch control correspondingly set various audio control parameters according to the setting instruction of the user on each parameter control on the setting window, and the method comprises the following steps:
responding to a switch instruction of the user to the ear return switch control, setting a control parameter corresponding to an ear return effect switch state, and correspondingly switching a switch display state of the ear return switch control;
and/or the presence of a gas in the gas,
responding to the switching instruction of the user to the dodging switch control, setting control parameters corresponding to the dodging effect switching state, and correspondingly switching the switching display state of the dodging switch control.
According to one or more embodiments of the present disclosure, the setup window includes a sound setup tab including: the loudspeaker selects the control, according to the user to the setting instruction of each parameter control on the setting window, correspondingly sets each audio control parameter, including:
and responding to a selection instruction of the user on the loudspeaker selection control, setting the working state of each loudspeaker in the live broadcast equipment, and correspondingly switching the loudspeaker selection state of the loudspeaker selection control.
According to one or more embodiments of the present disclosure, the setup window includes a sound setup tab including: the microphone selection control correspondingly sets various audio control parameters according to the setting instruction of the user on each parameter control on the setting window, and comprises the following steps:
and responding to a selection instruction of the user on the microphone selection control, setting the working state of each microphone in the live broadcast equipment, and correspondingly switching the microphone selection state of the microphone selection control.
According to one or more embodiments of the present disclosure, the setup window includes a sound setup tab including: and connecting the microphone volume setting control, correspondingly setting various audio control parameters according to the setting instruction of the user on each parameter control on the setting window, and comprising the following steps:
responding to the first volume setting operation of the user on the microphone connecting volume setting control, setting a volume control parameter of the opposite side during microphone connecting, and correspondingly changing the display state of the microphone connecting volume setting control.
According to one or more embodiments of the present disclosure, the setup window includes a sound setup tab including: the sound effect selection control and/or the sound effect volume control, according to the user is to the setting instruction of each parameter control on the setting window, each audio control parameter is correspondingly set, and the method comprises the following steps:
responding to a selection instruction of the user to the sound effect selection control, setting a sound effect control parameter during target audio synthesis, and correspondingly switching a sound effect selection state of the sound effect selection control;
and/or the presence of a gas in the gas,
responding to the second volume setting operation of the sound effect volume control, setting a volume size control parameter corresponding to the sound effect, and correspondingly changing the display state of the sound effect volume control.
According to one or more embodiments of the present disclosure, the setup window includes a shortcut setup tab, and the shortcut setup tab includes: the method includes that a shortcut key array and a shortcut key array to be selected are selected, and each audio control parameter is correspondingly set according to a setting instruction of the user to each parameter control on the setting window, and includes the following steps:
responding to a deleting instruction of the user on the selected shortcut key array, and updating the display states of the selected shortcut key array and the shortcut key array to be selected;
and/or the presence of a gas in the gas,
and updating the selected shortcut key array and the display state of the shortcut key array to be selected in response to the selection instruction of the user on the shortcut key array to be selected.
According to one or more embodiments of the present disclosure, before displaying the composition control on the display screen in the first form when the preset display condition is met, the method further includes:
displaying a sound setting interface on a display screen in response to the sound setting operation of the user;
and correspondingly setting various audio control parameters according to the setting instruction of the user on each parameter control on the sound setting interface.
According to one or more embodiments of the present disclosure, the sound setting interface includes: the ear return switch control and/or the dodging switch control correspondingly set various audio control parameters according to the setting instruction of the user on each parameter control on the sound setting interface, and the method comprises the following steps:
responding to a switch instruction of the user to the ear return switch control, setting a control parameter corresponding to an ear return effect switch state, and correspondingly switching a switch display state of the ear return switch control;
and/or the presence of a gas in the gas,
responding to the switching instruction of the user to the dodging switch control, setting control parameters corresponding to the dodging effect switching state, and correspondingly switching the switching display state of the dodging switch control.
According to one or more embodiments of the present disclosure, the sound setting interface includes: the loudspeaker selection control correspondingly sets various audio control parameters according to the setting instruction of the user on each parameter control on the sound setting interface, and the method comprises the following steps:
and responding to a selection instruction of the user on the loudspeaker selection control, setting the working state of each loudspeaker in the live broadcast equipment, and correspondingly switching the loudspeaker selection state of the loudspeaker selection control.
According to one or more embodiments of the present disclosure, the sound setting interface includes: the microphone selection control correspondingly sets various audio control parameters according to the setting instruction of the user on each parameter control on the sound setting interface, and comprises the following steps:
and responding to a selection instruction of the user on the microphone selection control, setting the working state of each microphone in the live broadcast equipment, and correspondingly switching the microphone selection state of the microphone selection control.
According to one or more embodiments of the present disclosure, the sound setting interface includes: connecting the microphone volume setting control, correspondingly setting various audio control parameters according to the setting instruction of the user on each parameter control on the sound setting interface, and comprising the following steps:
responding to the first volume setting operation of the user on the microphone connecting volume setting control, setting a volume control parameter of the opposite side during microphone connecting, and correspondingly changing the display state of the microphone connecting volume setting control.
According to one or more embodiments of the present disclosure, the sound setting interface includes: the sound effect selection control and/or the sound effect volume control correspondingly set various audio control parameters according to the setting instruction of each parameter control on the sound setting interface by the user, and comprises the following steps:
responding to a selection instruction of the user to the sound effect selection control, setting a sound effect control parameter during target audio synthesis, and correspondingly switching a sound effect selection state of the sound effect selection control;
and/or the presence of a gas in the gas,
responding to the second volume setting operation of the sound effect volume control, setting a volume size control parameter corresponding to the sound effect, and correspondingly changing the display state of the sound effect volume control.
According to one or more embodiments of the present disclosure, the sound setting interface includes: and correspondingly setting various audio control parameters according to the setting instruction of the user on each parameter control on the sound setting interface by the selected shortcut key array and the shortcut key array to be selected, wherein the setting instruction comprises the following steps:
responding to a deleting instruction of the user on the selected shortcut key array, and updating the display states of the selected shortcut key array and the shortcut key array to be selected;
and/or the presence of a gas in the gas,
and updating the selected shortcut key array and the display state of the shortcut key array to be selected in response to the selection instruction of the user on the shortcut key array to be selected.
According to one or more embodiments of the present disclosure, the sound setting interface includes: synthesizing a control display switch, and correspondingly setting various audio control parameters according to the setting instruction of the user on each parameter control on the sound setting interface, wherein the method comprises the following steps:
and responding to a switch instruction of the user to the display switch of the synthesis control, and correspondingly setting whether the synthesis control is displayed on a display screen.
According to one or more embodiments of the present disclosure, the method is applied to a live all-in-one machine, and the live all-in-one machine includes: the live broadcast application is installed in the main control module, and the audio processing module is used for synthesizing the target audio.
According to one or more embodiments of the present disclosure, the live all-in-one machine is connected to at least one shortcut key keyboard through an external interface, and the receiving of the audio synthesis operation of the user on the synthesis control includes;
receiving an operation instruction sent by the shortcut keyboard, wherein the operation instruction is determined by the shortcut keyboard in response to the operation of the user;
and determining the corresponding audio to be synthesized according to the operation instruction.
In a second aspect, according to one or more embodiments of the present disclosure, there is provided an audio synthesizing apparatus including:
the display module is used for displaying the synthesis control on the display screen in a first form when a preset display condition is met, wherein the visual influence of the first form on the display content on the display screen meets a preset visual requirement;
the receiving module is used for receiving the audio synthesis operation of the user on the synthesis control;
and the processing module is used for synthesizing audio to be synthesized and initial audio in live broadcast equipment provided with live broadcast application, and taking the synthesized target audio as input audio of the live broadcast application, wherein the audio to be synthesized is audio indicated by audio synthesis operation, and the initial audio is a sound signal sent by a main broadcast and/or in a live broadcast environment received by the live broadcast equipment through audio acquisition equipment.
In a third aspect, according to one or more embodiments of the present disclosure, there is provided an electronic device including:
at least one processor and memory;
the memory is used for storing a computer program;
the at least one processor executes the computer program stored by the memory, causing the at least one processor to perform the audio synthesis method as described above in the first aspect and in various possible designs of the first aspect.
In a fourth aspect, according to one or more embodiments of the present disclosure, there is provided a live all-in-one machine, including: an electronic device as claimed in the third aspect and various possible designs of the third aspect.
In a fifth aspect, according to one or more embodiments of the present disclosure, there is provided a computer-readable storage medium having stored therein a computer program which, when executed by a processor, implements an audio synthesis method as described above in the first aspect and various possible designs of the first aspect.
In a sixth aspect, according to one or more embodiments of the present disclosure, there is provided a computer program product comprising a computer program which, when executed by a processor, implements an audio synthesis method as described above in the first aspect and in various possible designs of the first aspect.
The foregoing description is only exemplary of the preferred embodiments of the disclosure and is illustrative of the principles of the technology employed. It will be appreciated by those skilled in the art that the scope of the disclosure herein is not limited to the particular combination of features described above, but also encompasses other embodiments in which any combination of the features described above or their equivalents does not depart from the spirit of the disclosure. For example, the above features and (but not limited to) the features disclosed in this disclosure having similar functions are replaced with each other to form the technical solution.
Further, while operations are depicted in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order. Under certain circumstances, multitasking and parallel processing may be advantageous. Likewise, while several specific implementation details are included in the above discussion, these should not be construed as limitations on the scope of the disclosure. Certain features that are described in the context of separate embodiments can also be implemented in combination in a single embodiment. Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable subcombination.
Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as example forms of implementing the claims.

Claims (20)

1. An audio synthesis method, comprising:
when a preset display condition is met, displaying a synthesis control on a display screen in a first form, wherein the visual influence of the first form on the display content on the display screen meets a preset visual requirement;
receiving an audio synthesis operation of a user on the synthesis control;
in the live broadcast equipment provided with the live broadcast application, synthesizing audio to be synthesized and initial audio, and taking the synthesized target audio as input audio of the live broadcast application, wherein the audio to be synthesized is audio indicated by audio synthesis operation, and the initial audio is a sound signal sent by a main broadcast and/or in a live broadcast environment received by the live broadcast equipment through audio acquisition equipment.
2. The audio synthesis method of claim 1, wherein the display screen comprises a touch screen, and the receiving an audio synthesis operation of the synthesis control by a user comprises:
and detecting that the user approaches or touches the synthetic control on the touch screen by using a preset operation object in real time.
3. The audio synthesis method according to claim 2, wherein the synthesis control includes a sub-control corresponding to the audio to be synthesized, and the detecting in real time that the user approaches or touches the synthesis control on the touch screen with a preset operator includes:
if the preset operation object is detected to be close to or touch the sub-control, switching the sub-control to a selected form;
and determining a synthetic operation instruction corresponding to the child control.
4. The audio synthesis method according to claim 1, wherein the preset display condition comprises: the method comprises the steps of detecting at least one of a camera equipment starting instruction, detecting that the live broadcast terminal moves according to a preset movement mode, receiving a preset voice instruction and receiving a preset button switch starting instruction.
5. The audio synthesis method of claim 1, wherein the synthesis control comprises a floating window control, and wherein displaying the synthesis control on the display screen in the first modality comprises:
and displaying the floating window control at the edge of the display screen according to the preset transparency, the preset size and the preset shape.
6. The audio synthesis method of claim 1, further comprising:
acquiring the state setting operation of the user on the synthesis control;
and switching the form and/or the display position of the synthesized control according to the state setting operation.
7. The audio synthesis method of claim 1, further comprising:
responding to the popup window setting operation of the user on the synthesis control during live broadcasting, and overlaying and displaying a setting window on a display screen;
and/or responding to sound setting operation triggered by the user before live broadcasting, and displaying a sound setting interface on a display screen;
and correspondingly setting various audio control parameters according to the setting instruction of the user on each parameter control on the setting window or the sound setting interface.
8. The audio synthesis method of claim 7, wherein the setup window comprises a sound setup tab, and wherein the sound setup tab, and/or wherein the sound setup interface comprises: the ear return switch control and/or the evasion switch control correspondingly set various audio control parameters according to the setting instruction of the user on each parameter control on the setting window or the sound setting interface, and the method comprises the following steps:
responding to a switch instruction of the user to the ear return switch control, setting a control parameter corresponding to an ear return effect switch state, and correspondingly switching a switch display state of the ear return switch control;
and/or the presence of a gas in the gas,
responding to the switching instruction of the user to the dodging switch control, setting control parameters corresponding to the dodging effect switching state, and correspondingly switching the switching display state of the dodging switch control.
9. The audio synthesis method of claim 7, wherein the setup window comprises a sound setup tab, and wherein the sound setup tab, and/or wherein the sound setup interface comprises: the loudspeaker selection control correspondingly sets various audio control parameters according to the setting instruction of the user on each parameter control on the setting window or the sound setting interface, and the method comprises the following steps:
and responding to a selection instruction of the user on the loudspeaker selection control, setting the working state of each loudspeaker in the live broadcast equipment, and correspondingly switching the loudspeaker selection state of the loudspeaker selection control.
10. The audio synthesis method of claim 7, wherein the setup window comprises a sound setup tab, and wherein the sound setup tab, and/or wherein the sound setup interface comprises: the microphone selection control correspondingly sets various audio control parameters according to the setting instruction of the user on each parameter control on the setting window or the sound setting interface, and comprises the following steps:
and responding to a selection instruction of the user on the microphone selection control, setting the working state of each microphone in the live broadcast equipment, and correspondingly switching the microphone selection state of the microphone selection control.
11. The audio synthesis method of claim 7, wherein the setup window comprises a sound setup tab, and wherein the sound setup tab, and/or wherein the sound setup interface comprises: and connecting the microphone volume setting control, correspondingly setting various audio control parameters according to the setting instruction of the user on each parameter control on the setting window or the sound setting interface, and comprising the following steps:
responding to the first volume setting operation of the user on the microphone connecting volume setting control, setting a volume control parameter of the opposite side during microphone connecting, and correspondingly changing the display state of the microphone connecting volume setting control.
12. The audio synthesis method of claim 7, wherein the setup window comprises a sound setup tab, and wherein the sound setup tab, and/or wherein the sound setup interface comprises: the sound effect selects the control, and/or, the sound effect volume control, according to the user is to set up the window or the setting instruction of each parameter control on the sound setting interface, correspond and set up each item audio frequency control parameter, include:
responding to a selection instruction of the user to the sound effect selection control, setting a sound effect control parameter during target audio synthesis, and correspondingly switching a sound effect selection state of the sound effect selection control;
and/or the presence of a gas in the gas,
responding to the second volume setting operation of the sound effect volume control, setting a volume size control parameter corresponding to the sound effect, and correspondingly changing the display state of the sound effect volume control.
13. The audio synthesis method according to claim 7, wherein the setup window comprises a shortcut setup tab, and/or wherein the sound setup interface comprises: the method includes that a shortcut key array and a shortcut key array to be selected are selected, and each audio control parameter is correspondingly set according to a setting instruction of the user to each parameter control on the setting window or the sound setting interface, and includes:
responding to a deleting instruction of the user on the selected shortcut key array, and updating the display states of the selected shortcut key array and the shortcut key array to be selected;
and/or the presence of a gas in the gas,
and updating the selected shortcut key array and the display state of the shortcut key array to be selected in response to the selection instruction of the user on the shortcut key array to be selected.
14. An audio synthesis method according to any of claims 1-13, applied to a live all-in-one machine comprising: the live broadcast application is installed in the main control module, and the audio processing module is used for synthesizing the target audio.
15. The audio synthesis method according to claim 14, wherein the live all-in-one machine is connected with at least one shortcut keyboard through an external interface, and the receiving of the audio synthesis operation of the user on the synthesis control comprises;
receiving an operation instruction sent by the shortcut keyboard, wherein the operation instruction is determined by the shortcut keyboard in response to the operation of the user;
and determining the corresponding audio to be synthesized according to the operation instruction.
16. An audio synthesizing apparatus, comprising:
the display module is used for displaying the synthesis control on the display screen in a first form when a preset display condition is met, wherein the visual influence of the first form on the display content on the display screen meets a preset visual requirement;
the receiving module is used for receiving the audio synthesis operation of the user on the synthesis control;
and the processing module is used for synthesizing audio to be synthesized and initial audio in live broadcast equipment provided with live broadcast application, and taking the synthesized target audio as input audio of the live broadcast application, wherein the audio to be synthesized is audio indicated by audio synthesis operation, and the initial audio is a sound signal sent by a main broadcast and/or in a live broadcast environment received by the live broadcast equipment through audio acquisition equipment.
17. An electronic device, comprising:
at least one processor and memory;
the memory is used for storing a computer program;
the at least one processor executing the computer program stored by the memory causes the at least one processor to perform the audio synthesis method of any of claims 1 to 15.
18. A live all-in-one machine, its characterized in that includes: the electronic device of claim 17.
19. A computer-readable storage medium, in which a computer program is stored which, when being executed by a processor, carries out an audio synthesis method according to any one of claims 1 to 15.
20. A computer program product comprising a computer program, characterized in that the computer program, when being executed by a processor, implements the audio synthesis method of any of claims 1 to 15.
CN202110523959.3A 2021-05-13 2021-05-13 Audio synthesis method, apparatus, device, medium, and program product Pending CN113112986A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202110523959.3A CN113112986A (en) 2021-05-13 2021-05-13 Audio synthesis method, apparatus, device, medium, and program product
PCT/CN2022/087491 WO2022237464A1 (en) 2021-05-13 2022-04-18 Audio synthesis method and apparatus, and device, medium and program product

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110523959.3A CN113112986A (en) 2021-05-13 2021-05-13 Audio synthesis method, apparatus, device, medium, and program product

Publications (1)

Publication Number Publication Date
CN113112986A true CN113112986A (en) 2021-07-13

Family

ID=76722536

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110523959.3A Pending CN113112986A (en) 2021-05-13 2021-05-13 Audio synthesis method, apparatus, device, medium, and program product

Country Status (2)

Country Link
CN (1) CN113112986A (en)
WO (1) WO2022237464A1 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113542792A (en) * 2021-07-14 2021-10-22 北京字节跳动网络技术有限公司 Audio merging method, audio uploading method, device and program product
CN114095480A (en) * 2022-01-24 2022-02-25 北京麦颂文化传播有限公司 KTV live broadcast wheat connecting method, device and system
WO2022237698A1 (en) * 2021-05-13 2022-11-17 北京字节跳动网络技术有限公司 Method, apparatus, and device for adjusting sound effect, medium, and program product
WO2022237464A1 (en) * 2021-05-13 2022-11-17 北京字节跳动网络技术有限公司 Audio synthesis method and apparatus, and device, medium and program product
WO2023237102A1 (en) * 2022-06-10 2023-12-14 北京字跳网络技术有限公司 Voice chat display method and apparatus, electronic device, and computer readable medium

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150078584A1 (en) * 2013-09-16 2015-03-19 Nancy Diane Moon Live Sound Mixer User Interface
CN105786439A (en) * 2014-12-26 2016-07-20 腾讯科技(深圳)有限公司 Method and device for controlling sound in terminal
CN105872253A (en) * 2016-05-31 2016-08-17 腾讯科技(深圳)有限公司 Live broadcast sound processing method and mobile terminal
CN207010753U (en) * 2017-07-28 2018-02-13 徐文波 A kind of intelligent mobile terminal with hardware audio effect processing function
CN109257493A (en) * 2018-09-15 2019-01-22 深圳市创成微电子有限公司 A kind of apparatus for processing audio control system based on APP
CN109814798A (en) * 2019-01-17 2019-05-28 Oppo广东移动通信有限公司 Ear returns function control method, device and mobile terminal
CN111208969A (en) * 2020-01-03 2020-05-29 海信视像科技股份有限公司 Selection control method of sound output equipment and display equipment

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111309214A (en) * 2020-03-17 2020-06-19 网易(杭州)网络有限公司 Video interface setting method and device, electronic equipment and storage medium
CN111510738B (en) * 2020-04-26 2023-08-11 北京字节跳动网络技术有限公司 Audio transmission method and device in live broadcast
CN111880694A (en) * 2020-07-22 2020-11-03 维沃移动通信有限公司 Display method, device, equipment and storage medium
CN112347273A (en) * 2020-11-05 2021-02-09 北京字节跳动网络技术有限公司 Audio playing method and device, electronic equipment and storage medium
CN112423009A (en) * 2020-11-09 2021-02-26 珠海格力电器股份有限公司 Method and equipment for controlling live broadcast audio
CN113112986A (en) * 2021-05-13 2021-07-13 北京字节跳动网络技术有限公司 Audio synthesis method, apparatus, device, medium, and program product

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150078584A1 (en) * 2013-09-16 2015-03-19 Nancy Diane Moon Live Sound Mixer User Interface
CN105786439A (en) * 2014-12-26 2016-07-20 腾讯科技(深圳)有限公司 Method and device for controlling sound in terminal
CN105872253A (en) * 2016-05-31 2016-08-17 腾讯科技(深圳)有限公司 Live broadcast sound processing method and mobile terminal
CN207010753U (en) * 2017-07-28 2018-02-13 徐文波 A kind of intelligent mobile terminal with hardware audio effect processing function
CN109257493A (en) * 2018-09-15 2019-01-22 深圳市创成微电子有限公司 A kind of apparatus for processing audio control system based on APP
CN109814798A (en) * 2019-01-17 2019-05-28 Oppo广东移动通信有限公司 Ear returns function control method, device and mobile terminal
CN111208969A (en) * 2020-01-03 2020-05-29 海信视像科技股份有限公司 Selection control method of sound output equipment and display equipment

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022237698A1 (en) * 2021-05-13 2022-11-17 北京字节跳动网络技术有限公司 Method, apparatus, and device for adjusting sound effect, medium, and program product
WO2022237464A1 (en) * 2021-05-13 2022-11-17 北京字节跳动网络技术有限公司 Audio synthesis method and apparatus, and device, medium and program product
CN113542792A (en) * 2021-07-14 2021-10-22 北京字节跳动网络技术有限公司 Audio merging method, audio uploading method, device and program product
CN113542792B (en) * 2021-07-14 2023-04-07 北京字节跳动网络技术有限公司 Audio merging method, audio uploading method, device and program product
CN114095480A (en) * 2022-01-24 2022-02-25 北京麦颂文化传播有限公司 KTV live broadcast wheat connecting method, device and system
CN114095480B (en) * 2022-01-24 2022-04-15 北京麦颂文化传播有限公司 KTV live broadcast wheat connecting method, device and system
WO2023237102A1 (en) * 2022-06-10 2023-12-14 北京字跳网络技术有限公司 Voice chat display method and apparatus, electronic device, and computer readable medium

Also Published As

Publication number Publication date
WO2022237464A1 (en) 2022-11-17

Similar Documents

Publication Publication Date Title
US10848704B2 (en) Remote controller and method for controlling screen thereof
CN113112986A (en) Audio synthesis method, apparatus, device, medium, and program product
US10353661B2 (en) Method for sharing screen between devices and device using the same
US20180356971A1 (en) Method of controlling a list scroll bar and an electronic device using the same
AU2013263767B2 (en) Apparatus and method of managing a plurality of objects displayed on touch screen
CN108139778A (en) The screen display method of portable device and portable device
CN103713843B (en) For generating the method and apparatus of task recommendation icon in the mobile device
CN108182016A (en) Mobile terminal and its control method
CN108600605A (en) Mobile terminal and its control method
EP3232307A1 (en) Method and mobile device for providing recommended items based on context awareness
CN108370395A (en) User terminal apparatus and its mode conversion method and audio system for controlling loudspeaker volume
KR20140139377A (en) Method and apparatus for controlling screen display using environmental information
CN105191330A (en) Display apparatus and graphic user interface screen providing method thereof
CN105872683A (en) Image display apparatus and method
CN105393202B (en) Method for providing the portable equipment of combined user interface component and controlling it
US10318093B2 (en) Method and apparatus for message processing
CN108370386A (en) Electronic equipment and the method for operating it
US20140298226A1 (en) Display apparatus displaying user interface and method of providing the user interface
CN108055490A (en) A kind of method for processing video frequency, device, mobile terminal and storage medium
EP4124052A1 (en) Video production method and apparatus, and device and storage medium
EP4343580A1 (en) Media file processing method and apparatus, device, readable storage medium, and product
CN112416485A (en) Information guiding method, device, terminal and storage medium
US10409478B2 (en) Method, apparatus, and recording medium for scrapping content
KR20140113643A (en) Information-sharing device, information-sharing method, information-sharing program and terminal device
CN112000267A (en) Information display method, device, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information

Address after: 100041 B-0035, 2 floor, 3 building, 30 Shixing street, Shijingshan District, Beijing.

Applicant after: Tiktok vision (Beijing) Co.,Ltd.

Address before: 100041 B-0035, 2 floor, 3 building, 30 Shixing street, Shijingshan District, Beijing.

Applicant before: BEIJING BYTEDANCE NETWORK TECHNOLOGY Co.,Ltd.

Address after: 100041 B-0035, 2 floor, 3 building, 30 Shixing street, Shijingshan District, Beijing.

Applicant after: Douyin Vision Co.,Ltd.

Address before: 100041 B-0035, 2 floor, 3 building, 30 Shixing street, Shijingshan District, Beijing.

Applicant before: Tiktok vision (Beijing) Co.,Ltd.

CB02 Change of applicant information