CN114143599A - Sound information processing method and related equipment - Google Patents

Sound information processing method and related equipment Download PDF

Info

Publication number
CN114143599A
CN114143599A CN202111401797.2A CN202111401797A CN114143599A CN 114143599 A CN114143599 A CN 114143599A CN 202111401797 A CN202111401797 A CN 202111401797A CN 114143599 A CN114143599 A CN 114143599A
Authority
CN
China
Prior art keywords
sound
audio data
channel
channels
track
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111401797.2A
Other languages
Chinese (zh)
Inventor
邓戊庚
李东
周红丽
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hunan Happly Sunshine Interactive Entertainment Media Co Ltd
Original Assignee
Hunan Happly Sunshine Interactive Entertainment Media Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hunan Happly Sunshine Interactive Entertainment Media Co Ltd filed Critical Hunan Happly Sunshine Interactive Entertainment Media Co Ltd
Priority to CN202111401797.2A priority Critical patent/CN114143599A/en
Publication of CN114143599A publication Critical patent/CN114143599A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • H04N21/4394Processing of audio elementary streams involving operations for analysing the audio stream, e.g. detecting features or characteristics in audio streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/485End-user interface for client configuration
    • H04N21/4852End-user interface for client configuration for modifying audio parameters, e.g. switching between mono and stereo
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8547Content authoring involving timestamps for synchronizing content

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Security & Cryptography (AREA)
  • Human Computer Interaction (AREA)
  • Stereophonic System (AREA)

Abstract

The invention provides a sound information processing method and related equipment, and the scheme comprises the following steps: acquiring audio data to be processed through a web end; identifying a soundtrack and a channel of the audio data; when the audio data has a sound track and 2 sound channels, performing intermodulation on the sound channels of the audio data with the same sound track and different sound channels; when the audio data has a plurality of sound tracks and 2 sound channels, combining the plurality of sound tracks into one sound track, and dividing the sound channels of the sound data in the audio data into a left channel, a right channel and a two-channel based on sound types; and carrying out sound mixing on the combined sound channels, and outputting the sound channels to the user through a web end. In the scheme, the audio track and the sound channel of the audio data are operated and played at the web end, so that the development cost of each PC desktop end is reduced, and meanwhile, the web end operates the audio track and the sound channel and can present the effect to a user in real time, so that the storage cost of server resources is reduced, and the effects of less cost and less server investment are finally achieved.

Description

Sound information processing method and related equipment
Technical Field
The invention relates to the technical field of sound processing, in particular to a sound information processing method and related equipment.
Background
In short video, video clip technology is rapidly developing, so-called video clip, i.e., a large amount of material shot in the production of a movie. Such as a picture, a mosaic, a sound effect, a transition and the like, and finally a coherent and smooth work with clear meaning, bright theme and artistic appeal is finished through selection, division, assembly and the like, thereby generating the clipping art. Editing is both an indispensable task in the film production process and the final re-creation in the film art creation process
The audio is the soul of the whole work for the video, and the track and the sound track in the audio are the components of the soul, so that the work has more infectivity, penetrability and perfection, the operations of merging, intercepting, adjusting, editing, processing, mixing and the like of the audio and the track sound track are needed.
Currently, the schemes for audio in mainstream clipping software include the following:
1. and (3) using PC desktop editing software to perform audio and video processing: the method has the advantages that the operation is convenient and simple, the defects are obvious, the editing has higher requirements on the performance of the machine, and a large amount of transcoding time is consumed;
2. the web end is used for editing, so that the method has the advantages of cross-platform use and reduction of development cost of a PC desktop end, and has the defect that the operation processing on audio tracks and sound channels of audio is not supported at present;
disclosure of Invention
In view of the above, embodiments of the present invention provide a sound information processing method and related device, so as to implement a low-cost audio processing scheme.
In order to achieve the above purpose, the embodiments of the present invention provide the following technical solutions:
a sound information processing method, the method comprising:
acquiring audio data to be processed through a web end;
identifying a soundtrack and a channel of the audio data;
when the audio data has a sound track and 2 sound channels, performing intermodulation on the sound channels of the audio data with the same sound track and different sound channels;
when the audio data has a plurality of sound tracks and 2 sound channels, combining the plurality of sound tracks into one sound track, and dividing the sound channels of the sound data in the audio data into a left channel, a right channel and a two-channel based on sound types;
and carrying out sound mixing on the combined sound channels, and outputting the sound channels to the user through a web end.
Optionally, in the sound information processing method, before the obtaining of the audio data to be processed through the web end, the method further includes:
and slicing the audio data, and splitting the audio track based on the slicing result.
Optionally, the sound information processing method further includes:
and receiving and caching the splitting result by adopting the created media node container, and judging whether the cached splitting result is complete or not based on the node data transmission state.
Optionally, in the sound information processing method, when the channel of the audio data is identified, the following steps are specifically adopted: and adopting a preset script and the AudioContext to identify the sound channel of the cache data of the media node container.
Optionally, in the sound information processing method, based on a sound type, dividing a channel of sound data in the audio data into a left channel, a right channel, and a binaural channel includes:
configuring a first person sound in the audio data as a left channel, configuring a second person sound in the audio data as a right channel, and configuring background music in the audio data as a dual channel.
A sound information processing apparatus comprising:
the data acquisition unit acquires audio data to be processed through a web terminal;
an identifying unit for identifying a soundtrack and a channel of the audio data;
a processing unit to:
when the audio data has a sound track and 2 sound channels, performing intermodulation on the sound channels of the audio data with the same sound track and different sound channels;
when the audio data has a plurality of sound tracks and 2 sound channels, combining the plurality of sound tracks into one sound track, and dividing the sound channels of the sound data in the audio data into a left channel, a right channel and a two-channel based on sound types;
and the sound mixing unit is used for mixing the sound of the combined sound channels and outputting the sound to the user through a web end.
Optionally, in the sound information processing apparatus, before the obtaining of the audio data to be processed through the web end, the method further includes:
slicing the audio data, and splitting an audio track based on a slicing result;
and receiving and caching the splitting result by adopting the created media node container, and judging whether the cached splitting result is complete or not based on the node data transmission state.
Optionally, in the sound information processing apparatus, based on a sound type, dividing a channel of sound data in the audio data into a left channel, a right channel, and a binaural channel, includes:
configuring a first person sound in the audio data as a left channel, configuring a second person sound in the audio data as a right channel, and configuring background music in the audio data as a dual channel.
A storage medium having stored thereon a computer program which, when executed by a processor, implements the steps of the sound information processing method as set forth in any one of the preceding claims.
A soundtrack and channel mixer comprising a memory and a processor;
the memory is used for storing programs;
the processor is configured to execute the program to implement the steps of the sound information processing method according to any one of the above.
Based on the technical scheme, in the scheme provided by the embodiment of the invention, the audio track and the audio channel of the audio data are operated and played at the web end, so that the development cost of each PC desktop end is reduced, and meanwhile, the web end operates the audio track and the audio channel and can present the effect to a user in real time, so that the storage cost of server resources is reduced, and the effects of less cost and less server investment are finally achieved.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the provided drawings without creative efforts.
Fig. 1 is a schematic flow chart illustrating a method for processing voice information according to an embodiment of the present disclosure;
FIG. 2 is a flow chart illustrating a method for processing audio information according to another embodiment of the present disclosure;
FIG. 3 is a flowchart illustrating a method for processing audio information according to another embodiment of the present disclosure;
FIG. 4 is a schematic diagram of a sound information processing apparatus according to an embodiment of the present application;
fig. 5 is a schematic structural diagram of a soundtrack and channel mixer disclosed in an embodiment of the present application;
fig. 6 is a schematic overall flow chart of the sound information processing scheme disclosed in the embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Based on the above situation, in order to play and operate the audio track and channel mixer on the web end cross platform, the invention provides a method for processing and playing the video audio track and channel mixer on the basis of the web end.
In order to reduce the development cost of the PC desktop, the sound information processing method disclosed in the embodiment of the present application may be applied to a web side, and referring to fig. 1, the method includes:
step S101: and acquiring audio data to be processed through a web terminal.
In the scheme, the acquired audio data to be processed can be pure sound data or video data.
Step S102: a soundtrack and a channel of the audio data are identified.
In the audio data, the audio track and the sound channel of different audio data can be different, and in the existing scheme, the sound track of audio data can be divided into two sound tracks, and in this application, these two sound tracks are respectively marked as first sound track and second sound track, and the sound channel of audio data can be divided into left channel, right channel.
Step S103: and when the audio data has a sound track and 2 sound channels, performing intermodulation on the sound channels of the audio data with the same sound track and different sound channels.
In this step, when it is detected that the audio data only has one audio track and two audio channels, in this scheme, the two audio channels of the acquired audio data are exchanged, that is, sound information originally output by using the left audio channel is output by using the right audio channel, sound information originally output by using the right audio channel is output by using the left audio channel, and finally, sound interaction of the left earphone and the right earphone is presented to the user, where a specific process is shown in fig. 2;
further, before the step is specifically executed, the processing mode and the description of the processing result of the step may be presented to the user in a preset mode, whether the user needs to execute the interactive instruction of the step is obtained through the interactive window, when a confirmation instruction entered through the interactive window by the user is obtained, the step S103 is executed, and when a cancel instruction entered through the interactive window by the user is obtained, the audio data is not operated.
Step S104: when the audio data has a plurality of sound tracks and 2 channels, the plurality of sound tracks are combined into one sound track, and the channels of sound data in the audio data are divided into a left channel, a right channel, and a binaural channel based on a sound type.
In this step, when it is detected that the audio data includes a plurality of sound tracks and two sound channels, the plurality of sound tracks are combined into one sound track, and the sound channels of the sound data in the audio data are divided into a left channel, a right channel, and a binaural channel based on the sound type of the audio data in the audio data.
For example, a first human voice in the audio data is configured as a left channel, a second human voice in the audio data is configured as a right channel, and background music in the audio data is configured as a binaural.
The splitting process will now be described with a specific example:
if it is determined that the audio data has a first soundtrack and a second soundtrack, each soundtrack having a left channel and a right channel, for example, a first track: wherein the left track is the dominant speaking sound, the right track is the background sound, the second audio track: wherein the left channel is a vocal chord of the vocal chord, the right channel is a scene sound, and when processing, the whole audio is synthesized into a track, for example, the audio data of the second track can be switched into a first vocal track, the vocal track of the chief vocal chord is set as the left channel, the vocal chord is set as the right channel, a piece of binaural (and left and right channels) background music is added, the background sound is set as the binaural, and the scene lift is deleted, specifically, when setting the vocal track and the vocal tracks, setting can be performed based on user execution sent by the user, for example, a first instruction input by the user is obtained, the first instruction corresponds to a vocal track to be reserved, based on the first instruction, the vocal track of the audio data is set as the vocal track corresponding to the first instruction, a second instruction input by the user is obtained, and the second instruction can be configured with corresponding relations between different audio types and the vocal tracks, and when the second instruction is acquired, configuring the sound channels corresponding to the audio data of each type based on the second branch collar.
After step S104 is finished, the volume corresponding to the audio data may be further adjusted based on a volume adjustment instruction input by the user, so that the output volume of the audio data can be adapted to the requirement of the designer.
Step S105: and carrying out sound mixing on the combined sound channels, and outputting the sound channels to the user through a web end.
After the operation of step S104 is finished, the required channels are combined together, and then audio mixing is performed through AudioContext, so as to achieve the final effect. That is, the AudioContext is used to perform audio mixing processing on the audio data processed in step S104, and the processing result is displayed for each user.
According to the technical scheme disclosed by the embodiment of the application, the audio tracks and the sound channels of the audio data are operated and played at the web end, so that the development cost of each PC desktop end is reduced, and meanwhile, the web end operates the audio tracks and the sound channels and can present the effect to a user in real time, so that the storage cost of server resources is reduced, and the effects of less cost and less server investment are finally achieved.
In the technical solution disclosed in another embodiment of the present application, in order to reduce transmission of a network bandwidth, in the technical solution, the obtained audio data may be sliced in advance at a Web end, and therefore, referring to fig. 2, before obtaining audio data to be processed through the Web end, the technical solution disclosed in the embodiment of the present application may further include:
step S201: slicing the audio data;
in identifying the soundtrack of the audio data, audio track splitting is performed based on the slicing result (sliced audio data).
Wherein, the slicing processing of the audio data may adopt ffmpeg for slicing processing.
In the technical solution disclosed in another embodiment of the present application, the slicing processing of the audio data may be performed on the audio data
In a technical solution disclosed in another embodiment of the present application, the slicing processing on the audio data may further create a media node container at the web end, and use the media node container to store a plurality of audio data for caching, specifically, referring to fig. 3, in the foregoing method, after the slicing processing on the audio data, the method may further include:
step S301: and receiving and caching the splitting result by adopting the created media node container, and judging whether the cached splitting result is complete or not based on the node data transmission state.
Specifically, in this step, the Web end uses a JavaScript native interface to create a media (media) node container according to the audio track split by the fragment processing, and uses the media node container to receive and cache the fragment result, and analyzes the transmission state of the node data to confirm the integrity of the data in the transmission process, so as to prevent the data from being lost in the transmission process.
In the technical solution disclosed in another embodiment of the present application, when the audio channel of the audio data is identified, the audio channel may be identified by using a script AudioContext to perform audio channel identification on the cache data in the media (media) node, and the cache data may be split into separate audio channels (creategan).
In this embodiment, a sound information processing apparatus corresponding to the method is also disclosed, and the detailed working content of each unit in the apparatus is referred to the content of the method embodiment.
The following describes a sound information processing apparatus according to an embodiment of the present invention, and the sound information processing apparatus described below and the sound information processing method described above may be referred to in correspondence with each other.
The sound information processing device disclosed in the embodiment of the present application can be applied to a web server, and referring to fig. 4, the device includes: a data acquisition unit 100, a recognition unit 200, a processing unit 300 and a mixing unit 400;
the data acquisition unit 100 corresponds to the step S101 in the method, and is configured to acquire audio data to be processed through a web end;
the audio data to be processed, which is acquired by the data acquisition unit 100, may be pure sound data or video data.
An identifying unit 200, corresponding to step S102 in the above method, for identifying the soundtrack and the channel of the audio data;
a processing unit 300, corresponding to steps S103 and S104 in the above method, for:
when the audio data has a sound track and 2 sound channels, performing intermodulation on the sound channels of the audio data with the same sound track and different sound channels;
when the audio data has a plurality of sound tracks and 2 sound channels, combining the plurality of sound tracks into one sound track, and dividing the sound channels of the sound data in the audio data into a left channel, a right channel and a two-channel based on sound types;
the mixing unit 400, corresponding to step S105 in the above method, is configured to perform sound mixing on the combined channels, and output the sound mixed channels to the user through the web.
Corresponding to the method, the method for obtaining the audio data to be processed through the web end further comprises the following steps:
slicing the audio data, and splitting an audio track based on a slicing result;
and receiving and caching the splitting result by adopting the created media node container, and judging whether the cached splitting result is complete or not based on the node data transmission state.
When the channels of the sound data in the audio data are divided into a left channel, a right channel, and a binaural channel based on the sound type, the processing unit 300 is specifically configured to:
configuring a first person sound in the audio data as a left channel, configuring a second person sound in the audio data as a right channel, and configuring background music in the audio data as a dual channel.
Corresponding to the above method, the present application also discloses a storage medium applied to a web side and having a computer program stored thereon, wherein the computer program, when executed by a processor, implements the steps of the sound information processing method as described in any one of the above.
Corresponding to the above method, the present application also discloses a soundtrack and channel mixer comprising: at least one processor 100, at least one memory 300, and may further include, at least one communication interface 200, at least one communication bus 400;
in the embodiment of the present invention, the number of the processor 100, the communication interface 200, the memory 300, and the communication bus 400 is at least one, and the processor 100, the communication interface 200, and the memory 300 complete the communication with each other through the communication bus 400; it is clear that the communication connections shown by the processor 100, the communication interface 200, the memory 300 and the communication bus 400 shown in fig. 6 are only optional;
optionally, the communication interface 200 may be an interface of a communication module, such as an interface of a GSM module;
the processor 100 may be a central processing unit CPU or an application Specific Integrated circuit asic or one or more Integrated circuits configured to implement embodiments of the present invention.
Memory 300 may comprise high-speed RAM memory and may also include non-volatile memory (non-volatile memory), such as at least one disk memory.
Wherein, the processor 100 is specifically configured to:
acquiring audio data to be processed through a web end;
identifying a soundtrack and a channel of the audio data;
when the audio data has a sound track and 2 sound channels, performing intermodulation on the sound channels of the audio data with the same sound track and different sound channels;
when the audio data has a plurality of sound tracks and 2 sound channels, combining the plurality of sound tracks into one sound track, and dividing the sound channels of the sound data in the audio data into a left channel, a right channel and a two-channel based on sound types;
and carrying out sound mixing on the combined sound channels, and outputting the sound channels to the user through a web end.
By combining the above embodiments, the present application further adopts a specific embodiment to explain the technical solutions disclosed in the embodiments of the present application.
In particular, with reference to figure 6,
the embodiment of the application discloses that after a Web end receives audio data to be processed (video or audio), the Web end performs fragment processing by adopting an FFmpeg technology, performs audio track splitting and outputs a fragment result and an audio track splitting result, receives and caches the fragment result and the audio track splitting result through a common media node container, performs audio track separation on the cached result, performs mixing processing by adopting an AudioContext after the audio track separation, and outputs a processing result to a user through the Web end.
For convenience of description, the above system is described with the functions divided into various modules, which are described separately. Of course, the functionality of the various modules may be implemented in the same one or more software and/or hardware implementations of the invention.
The embodiments in the present specification are described in a progressive manner, and the same and similar parts among the embodiments are referred to each other, and each embodiment focuses on the differences from the other embodiments. In particular, the system or system embodiments are substantially similar to the method embodiments and therefore are described in a relatively simple manner, and reference may be made to some of the descriptions of the method embodiments for related points. The above-described system and system embodiments are only illustrative, wherein the units described as separate parts may or may not be physically separate, and the parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of the present embodiment. One of ordinary skill in the art can understand and implement it without inventive effort.
Those of skill would further appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, computer software, or combinations of both, and that the various illustrative components and steps have been described above generally in terms of their functionality in order to clearly illustrate this interchangeability of hardware and software. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.
The steps of a method or algorithm described in connection with the embodiments disclosed herein may be embodied directly in hardware, in a software module executed by a processor, or in a combination of the two. A software module may reside in Random Access Memory (RAM), memory, Read Only Memory (ROM), electrically programmable ROM, electrically erasable programmable ROM, registers, hard disk, a removable disk, a CD-ROM, or any other form of storage medium known in the art.
It is further noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present invention. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the invention. Thus, the present invention is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims (10)

1. A method for processing sound information, the method comprising:
acquiring audio data to be processed through a web end;
identifying a soundtrack and a channel of the audio data;
when the audio data has a sound track and 2 sound channels, performing intermodulation on the sound channels of the audio data with the same sound track and different sound channels;
when the audio data has a plurality of sound tracks and 2 sound channels, combining the plurality of sound tracks into one sound track, and dividing the sound channels of the sound data in the audio data into a left channel, a right channel and a two-channel based on sound types;
and carrying out sound mixing on the combined sound channels, and outputting the sound channels to the user through a web end.
2. The sound information processing method according to claim 1, wherein before the obtaining the audio data to be processed through the web side, the method further comprises:
and slicing the audio data, and splitting the audio track based on the slicing result.
3. The sound information processing method according to claim 2, characterized by further comprising:
and receiving and caching the splitting result by adopting the created media node container, and judging whether the cached splitting result is complete or not based on the node data transmission state.
4. The sound information processing method according to claim 3, wherein in identifying the channel of the audio data, specifically: and adopting a preset script and the AudioContext to identify the sound channel of the cache data of the media node container.
5. The sound information processing method according to claim 3, wherein dividing channels of sound data in the audio data into a left channel, a right channel, and a binaural channel based on a sound type comprises:
configuring a first person sound in the audio data as a left channel, configuring a second person sound in the audio data as a right channel, and configuring background music in the audio data as a dual channel.
6. An audio information processing apparatus, characterized by comprising:
the data acquisition unit acquires audio data to be processed through a web terminal;
an identifying unit for identifying a soundtrack and a channel of the audio data;
a processing unit to:
when the audio data has a sound track and 2 sound channels, performing intermodulation on the sound channels of the audio data with the same sound track and different sound channels;
when the audio data has a plurality of sound tracks and 2 sound channels, combining the plurality of sound tracks into one sound track, and dividing the sound channels of the sound data in the audio data into a left channel, a right channel and a two-channel based on sound types;
and the sound mixing unit is used for mixing the sound of the combined sound channels and outputting the sound to the user through a web end.
7. The sound information processing apparatus according to claim 6, further comprising, before the obtaining the audio data to be processed through the web side:
slicing the audio data, and splitting an audio track based on a slicing result;
and receiving and caching the splitting result by adopting the created media node container, and judging whether the cached splitting result is complete or not based on the node data transmission state.
8. The sound information processing apparatus according to claim 6, wherein dividing channels of sound data in the audio data into a left channel, a right channel, and a binaural channel based on a sound type comprises:
configuring a first person sound in the audio data as a left channel, configuring a second person sound in the audio data as a right channel, and configuring background music in the audio data as a dual channel.
9. A storage medium having a computer program stored thereon, wherein the computer program, when executed by a processor, is implemented as
The sound information processing method according to any one of claims 1 to 5, wherein the steps are performed in a single step.
10. A soundtrack and channel mixer comprising a memory and a processor;
the memory is used for storing programs;
the processor, which executes the program, realizes the respective steps of the sound information processing method according to any one of claims 1 to 5.
CN202111401797.2A 2021-11-19 2021-11-19 Sound information processing method and related equipment Pending CN114143599A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111401797.2A CN114143599A (en) 2021-11-19 2021-11-19 Sound information processing method and related equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111401797.2A CN114143599A (en) 2021-11-19 2021-11-19 Sound information processing method and related equipment

Publications (1)

Publication Number Publication Date
CN114143599A true CN114143599A (en) 2022-03-04

Family

ID=80391288

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111401797.2A Pending CN114143599A (en) 2021-11-19 2021-11-19 Sound information processing method and related equipment

Country Status (1)

Country Link
CN (1) CN114143599A (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103632703A (en) * 2012-08-23 2014-03-12 蒂雅克股份有限公司 Multitrack recorder
CN105228041A (en) * 2015-09-24 2016-01-06 联想(北京)有限公司 A kind of information processing method and audio output device
CN106454684A (en) * 2016-10-18 2017-02-22 北京小米移动软件有限公司 Multimedia playing control method and device
CN110992970A (en) * 2019-12-13 2020-04-10 腾讯音乐娱乐科技(深圳)有限公司 Audio synthesis method and related device

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103632703A (en) * 2012-08-23 2014-03-12 蒂雅克股份有限公司 Multitrack recorder
CN105228041A (en) * 2015-09-24 2016-01-06 联想(北京)有限公司 A kind of information processing method and audio output device
CN106454684A (en) * 2016-10-18 2017-02-22 北京小米移动软件有限公司 Multimedia playing control method and device
CN110992970A (en) * 2019-12-13 2020-04-10 腾讯音乐娱乐科技(深圳)有限公司 Audio synthesis method and related device

Similar Documents

Publication Publication Date Title
CN113891113B (en) Video clip synthesis method and electronic equipment
CN102447839B (en) A kind of video creating method based on Quartz Composer and device
US11595731B2 (en) Implementation method and system of real-time subtitle in live broadcast and device
KR101528367B1 (en) Sound control system and method as the same
CN112965701B (en) Front-end code generation method and device
CN108053696A (en) A kind of method, apparatus and terminal device that sound broadcasting is carried out according to reading content
CN108241602A (en) A kind of online method and apparatus for making PowerPoint
CN104464743B (en) Method for playing background music in voice chat room and mobile terminal
US20220319548A1 (en) Video processing method for application and electronic device
CN112860944A (en) Video rendering method, device, equipment, storage medium and computer program product
CN108241601A (en) A kind of online method and apparatus for making PowerPoint
US10469975B2 (en) Personalization of spatial audio for streaming platforms
CN114143599A (en) Sound information processing method and related equipment
Comunità et al. Web-based binaural audio and sonic narratives for cultural heritage
CA3175347A1 (en) System and method of dynamic random access rendering
JP2009503757A (en) Method and apparatus for controlling reproduction of optical disc program
CN114582348A (en) Voice playing system, method, device and equipment
CN111048063A (en) Audio synthesis method and device
US8014883B2 (en) Templates and style sheets for audio broadcasts
KR20200033083A (en) Making system for video using blocking and method therefor
US7743318B2 (en) Order independent batched updates on a text buffer
CN112017261A (en) Sticker generation method and device, electronic equipment and computer readable storage medium
CN114286179B (en) Video editing method, apparatus, and computer-readable storage medium
CN113079397A (en) Multimedia resource playing method and device
CN115086742B (en) Audio and video generation method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination