CN111417030A - Method, device, system, equipment and storage equipment for setting score - Google Patents
Method, device, system, equipment and storage equipment for setting score Download PDFInfo
- Publication number
- CN111417030A CN111417030A CN202010351173.3A CN202010351173A CN111417030A CN 111417030 A CN111417030 A CN 111417030A CN 202010351173 A CN202010351173 A CN 202010351173A CN 111417030 A CN111417030 A CN 111417030A
- Authority
- CN
- China
- Prior art keywords
- type
- video
- score
- song
- song audio
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 49
- 230000000694 effects Effects 0.000 claims abstract description 68
- 230000015654 memory Effects 0.000 claims description 18
- 238000012544 monitoring process Methods 0.000 claims description 10
- 238000012545 processing Methods 0.000 description 15
- 230000002093 peripheral effect Effects 0.000 description 10
- 230000001133 acceleration Effects 0.000 description 9
- 230000006870 function Effects 0.000 description 8
- 238000010586 diagram Methods 0.000 description 7
- 238000004891 communication Methods 0.000 description 6
- 230000003287 optical effect Effects 0.000 description 6
- 230000008569 process Effects 0.000 description 5
- 241000282326 Felis catus Species 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 3
- 239000000919 ceramic Substances 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 238000009877 rendering Methods 0.000 description 2
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000005055 memory storage Effects 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000006641 stabilisation Effects 0.000 description 1
- 238000011105 stabilization Methods 0.000 description 1
- 230000002194 synthesizing effect Effects 0.000 description 1
- 230000001960 triggered effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/8106—Monomedia components thereof involving special audio data, e.g. different tracks for different languages
- H04N21/8113—Monomedia components thereof involving special audio data, e.g. different tracks for different languages comprising music, e.g. song in MP3 format
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/233—Processing of audio elementary streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
- H04N21/23418—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/25—Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
- H04N21/262—Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission, generating play-lists
- H04N21/26258—Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission, generating play-lists for generating a list of items to be played back in a given order, e.g. playlist, or scheduling item distribution according to such list
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/433—Content storage operation, e.g. storage operation in response to a pause request, caching operations
- H04N21/4334—Recording operations
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/84—Generation or processing of descriptive data, e.g. content descriptors
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Databases & Information Systems (AREA)
- Television Signal Processing For Recording (AREA)
Abstract
The application discloses a method, a device, a system, equipment and storage equipment for setting score, and belongs to the technical field of computers. The method comprises the following steps: acquiring a first type label corresponding to a special effect added in a target video; sending a video score obtaining request to a server, wherein the video score obtaining request carries the first type label; receiving and displaying a song list sent by the server, wherein the song list comprises at least one song audio option; and when a music adding instruction corresponding to the target song audio is detected, setting the target song audio as the music of the target video. The efficiency of finding the score with high matching degree with the image content can be improved through the method and the device.
Description
Technical Field
The present application relates to the field of computer technologies, and in particular, to a method, an apparatus, a system, a device, and a storage device for setting a score.
Background
With the development of computer technology, people are more and more common to use short video applications, and when the short video applications are used:
the user records videos through the terminal, after the recording is completed, the user manually selects one audio frequency from the music library as the music score, and the short video application program synthesizes the shot images and the music score into a short video, so that the short video can be manufactured.
In the process of implementing the present application, the inventor finds that the prior art has at least the following problems:
when the user manually selects the score, it takes a long time for the user to find the score having a high degree of matching with the image content, which is inefficient.
Disclosure of Invention
The embodiment of the application provides a method, a device, a system, equipment and storage equipment for setting score, which can solve the problem of low score efficiency caused by high matching degree with image content. The technical scheme is as follows:
in one aspect, a method for setting a score is provided, and the method is applied to a terminal, and includes:
acquiring a first type label corresponding to a special effect added in a target video;
sending a video score obtaining request to a server, wherein the video score obtaining request carries the first type label;
receiving and displaying a song list sent by the server, wherein the song list comprises at least one song audio option;
and when a music adding instruction corresponding to the target song audio is detected, setting the target song audio as the music of the target video.
Optionally, before sending the video score obtaining request to the server, the method further includes:
inputting the target video into an image recognition model to obtain a second type label corresponding to the target video;
the video score obtaining request also carries the second type label.
Optionally, when the music adding instruction corresponding to the target song audio is detected, before setting the target song audio as the music of the target video, the method further includes:
and when monitoring a listening trial instruction corresponding to the target song audio, playing the target song audio.
In another aspect, a method for setting a score is provided, and the method is applied to a server, and includes:
receiving a video score obtaining request sent by a terminal, wherein the video score obtaining request carries a first type label corresponding to a special effect added in a target video;
determining at least one song audio matching the type of the special effect based on the first type tag;
generating a song list including an option for the at least one song audio;
and sending the song list to the terminal.
Optionally, the video score obtaining request further carries a second type tag corresponding to the target video;
the determining, based on the first type tag, at least one song audio that matches the type of the special effect includes:
determining at least one song audio matching the special effect and the type of the target video based on the first type tag and the second type tag.
Optionally, the determining, based on the first type tag and the second type tag, at least one song audio matching the special effect and the type of the target video includes:
and acquiring at least one song audio of which the corresponding type label comprises the first type label and the second type label.
Optionally, the obtaining of the video score includes that the video score obtaining request carries a plurality of first type tags and a plurality of second type tags, and the obtaining of the corresponding type tags includes at least one song audio of the first type tags and the second type tags, including:
and acquiring at least one song audio of which the number of the first type tags and the second type tags in the corresponding type tags exceeds a preset threshold value.
In another aspect, there is provided an apparatus for setting a score, which may be a terminal, the apparatus comprising:
the acquisition module is used for acquiring a first type label corresponding to a special effect added in a target video;
the sending module is used for sending a video score obtaining request to a server, wherein the video score obtaining request carries the first type label;
the display module is used for receiving and displaying a song list sent by the server, wherein the song list comprises at least one song audio option;
and the setting module is used for setting the target song audio as the score of the target video when the score adding instruction corresponding to the target song audio is detected.
Optionally, the sending module is further configured to:
inputting the target video into an image recognition model to obtain a second type label corresponding to the target video;
the video score obtaining request also carries the second type label.
Optionally, the apparatus further includes a monitoring module, where the monitoring module is configured to:
and when monitoring a listening trial instruction corresponding to the target song audio, playing the target song audio.
In another aspect, there is provided an apparatus for setting a score, which may be a server, the apparatus comprising:
the receiving module is used for receiving a video score obtaining request sent by a terminal, wherein the video score obtaining request carries a first type label corresponding to a special effect added in a target video;
a determining module for determining at least one song audio matching the type of the special effect based on the first type tag;
a generating module for generating a song list comprising options of the at least one song audio;
and the sending module is used for sending the song list to the terminal.
Optionally, the video score obtaining request further carries a second type tag corresponding to the target video;
the determining module is configured to:
determining at least one song audio matching the special effect and the type of the target video based on the first type tag and the second type tag.
Optionally, the determining module is configured to:
and acquiring at least one song audio of which the corresponding type label comprises the first type label and the second type label.
Optionally, the video score obtaining request carries a plurality of first type tags and a plurality of second type tags, and the determining module is configured to:
and acquiring at least one song audio of which the number of the first type tags and the second type tags in the corresponding type tags exceeds a preset threshold value.
In yet another aspect, a computer device is provided, the computer device comprising a processor and a memory, the memory having instructions stored therein, the processor executing the instructions to cause the computer device to implement the method of setting a score.
In yet another aspect, a computer-readable storage medium is provided, the computer-readable storage medium storing instructions, execution of which by a computer device causes the computer device to implement the method of setting a score.
The technical scheme provided by the embodiment of the application has the following beneficial effects:
according to the method and the device, a first type label corresponding to a special effect added in a target video is obtained through a terminal, a video music obtaining request is sent to a server, the server determines at least one song audio matched with the type of the special effect based on the first type label in the video music obtaining request, then a song list comprising at least one song audio option is generated and sent to the terminal, after the terminal receives the song list, a user can trigger a music adding instruction corresponding to the target song audio, and then the terminal sets the target song audio as the music of the target video. Therefore, the terminal can automatically find the score with high matching degree for the target video, so that a user does not need to manually find the score with high matching degree, and the efficiency of finding the score with high matching degree with the image content is improved.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present application, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present application, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
FIG. 1 is a schematic diagram of an implementation environment provided by an embodiment of the present application;
fig. 2 is a flowchart of a terminal side of a method for setting a score according to an embodiment of the present application;
fig. 3 is a server-side flowchart of a method for setting a score according to an embodiment of the present application;
fig. 4 is an interaction flowchart of a method for setting a score according to an embodiment of the present application;
fig. 5 is a schematic structural diagram of a device for setting a score according to an embodiment of the present application;
fig. 6 is a schematic structural diagram of a device for setting a score according to an embodiment of the present application;
fig. 7 is a schematic structural diagram of a terminal provided in an embodiment of the present application;
fig. 8 is a schematic structural diagram of a server according to an embodiment of the present application.
Detailed Description
To make the objects, technical solutions and advantages of the present application more clear, embodiments of the present application will be described in further detail below with reference to the accompanying drawings.
The embodiment of the application provides a method for setting a score, which can be realized by a terminal and a server together. The terminal can be a mobile phone, a desktop computer, a tablet computer, an intelligent wearable device and the like, and can be provided with a screen, a loudspeaker, a camera and the like. The terminal may have a function of transmitting data, and the terminal may be installed with an application program, for example, a short video application program. The server may be a background server of the short video application program, the server may be a single server or a server group, if the server is a single server, the server may be responsible for all processing required by the server in the following scheme, if the server is a server group, different servers in the server group may be responsible for different processing in the following scheme, respectively, and the specific processing allocation condition may be set arbitrarily by a technical person according to actual requirements, which is not described herein again.
When the user uses the short video application program, the user can click the recording interface jump control, and then the terminal can jump to the short video recording interface, and display the video shot by the camera on the short video recording interface, wherein the short video recording interface can be the interface shown in fig. 1. The user can select to click the special effect control on the short video recording interface, and select a special effect, and then the special effect can be displayed at a corresponding position in the video, for example, the special effect is a cat ear special effect, and the cat ear special effect can be displayed at the top of the head of a person. The user can click the video recording control, and then the terminal can start to record and store the video carrying the special effect.
Fig. 2 is a flowchart of a terminal side of a method for setting a score according to an embodiment of the present application. Referring to fig. 2, the process includes:
The video score obtaining request carries a first type label.
And step 203, receiving and displaying the song list sent by the server.
Wherein the song list comprises at least one song audio option.
And step 204, when a music adding instruction corresponding to the target song audio is detected, setting the target song audio as the music of the target video.
Fig. 3 is a server-side flowchart of a method for setting a score according to an embodiment of the present application. Referring to fig. 3, the process includes:
The video score obtaining request carries a first type label corresponding to a special effect added in the target video.
Step 303 generates a song list including at least one selection of song audio.
Fig. 4 is an interaction flowchart of a method for setting a score according to an embodiment of the present application. Referring to fig. 4, the process includes:
step 401, the terminal obtains a first type label corresponding to a special effect added in a target video.
In implementation, a user may open a short video application and display a short video recording interface where the user may trigger a special effect control. And then the terminal can display a special effect list on the lower half part of the short video recording interface, the special effect can be divided into an expression special effect and a background special effect, and a user can select any special effect control in the special effect list. The user may also select any number of non-conflicting effect controls in the list of effects, for example, the user selects an emoji effect control and a background effect control. And then the terminal can take the special effect corresponding to the selected special effect control as the added special effect. When the user finishes selecting the special effect control and clicks the video recording control on the short video recording interface, the terminal can input the target video into the image recognition model to obtain the position of the target image, wherein the target image can be a content main body of the target video, for example, a face in the target video. And the terminal displays the special effect at the position of the target image, so that the target video added with the special effect can be obtained, for example, the special effect of the cat ear is displayed on the head of a person in the target video. Then, the terminal obtains a first type label corresponding to the stored special effect according to the corresponding relationship between the special effect and the first type label, where the obtained first type label may be multiple, that is, one special effect may correspond to multiple first type labels.
Optionally, the terminal may further input the target video into the image recognition model to obtain a second type tag corresponding to the target video.
After the target video is input into the image recognition model, the image recognition model may also output a second type tag corresponding to the target video, such as cat, dog, person, weather. And then the terminal may obtain a plurality of second type tags.
Step 402, the terminal sends a video score obtaining request to the server.
The video score obtaining request carries the first type label.
In an implementation, after the processing of step 401 is completed, the terminal may generate a video score obtaining request and send the video score obtaining request to the server.
Optionally, the video score obtaining request may further carry a second type tag.
Step 403, the server determines at least one song audio matching the type of the special effect based on the first type tag, and generates a song list including an option of the at least one song audio.
In implementation, after the server receives the video score obtaining request sent by the terminal, the following processing may be performed:
in a first step, the server determines at least one song audio matching the type of the special effect based on the first type tag.
The server may obtain a plurality of first type tags included in the video soundtrack obtaining request, and obtain a plurality of type tags corresponding to the song audio, where the type tags corresponding to the song audio may also be a plurality of type tags. And further detecting the number of the first type tags contained in the type tags corresponding to the song audios. And if the number is larger than a preset threshold value, acquiring the song audio as the pre-selected song audio.
For example, the video soundtrack acquisition request includes A, B, C three first type tags, and if the preset threshold is 2, song audio including A, B, C, A, B, B, C and A, C in the corresponding type tags is acquired and used as the pre-selected song audio.
Optionally, the specific processing that the server may obtain at least one song audio in which the number of the corresponding type tags, including the first type tag and the second type tag, exceeds the preset threshold may be as follows:
the server may obtain a plurality of first type tags and a plurality of second type tags included in the video soundtrack obtaining request, and obtain a type tag corresponding to the song audio, which may also be a plurality of type tags. And further detecting the number of the first type tags and the number of the second type tags contained in the type tags corresponding to the song audio. And if the two numbers are both larger than a preset threshold value, acquiring the song audio, and taking the song audio as the pre-selected song audio.
For example, the video soundtrack acquisition request includes A, B, C three first type tags and a, b, and c three second type tags, and the preset threshold is 2, then the audio of the song to be selected including any two or three first tags in A, B, C in the corresponding type tags may be acquired first, and then the audio of the song including any two or three second tags in a, b, and c in the type tag corresponding to the audio of the song to be selected is acquired as the audio of the preselected song.
Optionally, the server may further obtain at least one song audio, where the number of the first type tags and the number of the second type tags in the corresponding type tags both exceed a first preset threshold and the total number of the first type tags and the second type tags included in the corresponding type tags exceed a second preset threshold, and the processing may be as follows:
the server may obtain a plurality of first type tags and a plurality of second type tags included in the video soundtrack obtaining request, and obtain a type tag corresponding to the song audio, which may also be a plurality of type tags. And further detecting the number of the first type tags and the number of the second type tags contained in the type tags corresponding to the song audio. And if the two numbers are both larger than a first preset threshold value, acquiring the song audio, and taking the song audio as the audio of the song to be selected. And then acquiring the total number of the first type tags and the second type tags contained in the type tags corresponding to the audio of the song to be selected, judging the relation between the total number and a second preset threshold value, and if the total number is greater than the second preset threshold value, taking the audio of the song to be selected as the audio of the preselected song.
For example, the video score obtaining request includes A, B, C three first type tags and a, b, and c three second type tags, the first preset threshold is 2, and the second preset threshold is 5, then the screened song audio including any two or three first type tags in A, B, C in the corresponding type tags may be obtained first, and then the song audio including any two or three second type tags in a, b, and c in the type tags corresponding to the screened song audio is obtained as the song audio to be selected. Then, a song audio containing A, B, C three tags of the first type and containing any two or three tags of the second type among a, b, c, or a song audio containing A, B, C three tags of the second type among a, b, c, or A, B, C three tags of the first type and containing three tags of the second type among a, b, c is taken as the pre-selected song audio.
In a second step, the server generates a song list including at least one selection of song audio.
After the server acquires the preselected song audio, the user can acquire song attribute information and song links of the preselected song audio, wherein the song attribute information can include the name of a singer, the name of a song and an album. The server may then generate an option to pre-select song audio based on the song attribute information and the song link, and the server may generate a song list of the option to pre-select song audio.
Optionally, the server may further sort the options of the song audio in the song list based on the total number of the first type tag and the second type tag included in the corresponding type tag.
And acquiring the total number of the first type tags and the second type tags in the type tags corresponding to each preselected song audio, and corresponding the total number with the options of the preselected song audio, so that the options of each preselected song audio can be sorted according to the total number.
Optionally, the server may also sort the options of the song audio in the song list based on the number of times the acquired respective song audio is selected as the score.
Each time a preselected song audio is selected as a soundtrack, the server may record an increase of 1 in the number of times the preselected song audio is selected as a soundtrack. And after the server acquires the audio frequency of each preselected song, the server can sort the options of the preselected song audio frequency in the song list according to the corresponding frequency of the preselected song audio frequency selected as the score, and further obtain the sorted song list.
Step 404, the server sends the song list to the terminal.
In an implementation, after generating a song list including at least one option to pre-select song audio, the server may transmit the song list to the terminal.
Step 405, the terminal displays the song list.
Wherein the song list includes at least one option to pre-select song audio.
In implementation, after receiving the displayed song list sent by the server, the terminal may display the song list, and the options in the list may display information such as the name of the singer, the name of the song, and the album.
And step 406, when the terminal detects a music adding instruction corresponding to the target song audio, setting the target song audio as the music of the target video.
In implementation, after the terminal displays the song list, the user can trigger any one of the options of the pre-selected song audio in the song list, and the terminal determines the option of the pre-selected song audio as the option of the target song audio. Then, the user can click the determination control and generate a score adding instruction corresponding to the target song audio, and the terminal can set the target song audio as the score of the target video according to the score adding instruction.
Optionally, after the terminal determines the option of the preselected song audio as the option of the target song audio, when a listening trial instruction corresponding to the target song audio is monitored, the target song audio is played.
After the terminal determines the option of the pre-selected song audio as the option of the target song audio, a listening trial instruction may be triggered. When the terminal monitors a listening trial instruction corresponding to the target song audio, the terminal can access the song link corresponding to the option, acquire song audio data, decode the song audio data, transmit the decoded song audio data to the loudspeaker, and then the terminal can play the target song audio.
After the above steps are completed, the short video application may cut the score based on the time length of the recorded video, for example, if the time length of the recorded video is 30 seconds, the first 30 seconds of cutting the score correspond to the recorded video. And further synthesizing the cut soundtrack, the recorded video and the special effect into a short video.
According to the method and the device, a first type label corresponding to a special effect added in a target video is obtained through a terminal, a video music obtaining request is sent to a server, the server determines at least one song audio matched with the type of the special effect based on the first type label in the video music obtaining request, then a song list comprising at least one song audio option is generated and sent to the terminal, after the terminal receives the song list, a user can trigger a music adding instruction corresponding to the target song audio, and then the terminal sets the target song audio as the music of the target video. Therefore, the terminal can automatically find the score with high matching degree for the target video, so that a user does not need to manually find the score with high matching degree, and the efficiency of finding the score with high matching degree with the image content is improved.
All the above optional technical solutions may be combined arbitrarily to form the optional embodiments of the present disclosure, and are not described herein again.
Fig. 5 is an image display device according to an embodiment of the present disclosure, which may be the terminal.
Referring to fig. 5, the apparatus includes:
an obtaining module 510, configured to obtain a first type tag corresponding to a special effect added in a target video;
a sending module 520, configured to send a video score obtaining request to a server, where the video score obtaining request carries the first type tag;
a display module 530, configured to receive and display a song list sent by the server, where the song list includes at least one song audio option;
and the setting module 540 is configured to set the target song audio as the score of the target video when the score adding instruction corresponding to the target song audio is detected.
Optionally, the sending module 520 is further configured to:
inputting the target video into an image recognition model to obtain a second type label corresponding to the target video;
the video score obtaining request also carries the second type label.
Optionally, the apparatus further includes a monitoring module, where the monitoring module is configured to:
and when monitoring a listening trial instruction corresponding to the target song audio, playing the target song audio.
Fig. 6 is an apparatus for displaying an image according to an embodiment of the present application, where the apparatus may be the server. Referring to fig. 6, the apparatus includes:
the receiving module 610 is configured to receive a video score obtaining request sent by a terminal, where the video score obtaining request carries a first type tag corresponding to a special effect added in a target video;
a determining module 620, configured to determine at least one song audio matching the type of the special effect based on the first type tag;
a generating module 630 for generating a song list comprising options of the at least one song audio;
a sending module 640, configured to send the song list to the terminal.
Optionally, the video score obtaining request further carries a second type tag corresponding to the target video;
the determining module 620 is configured to:
determining at least one song audio matching the special effect and the type of the target video based on the first type tag and the second type tag.
Optionally, the determining module 620 is configured to:
and acquiring at least one song audio of which the corresponding type label comprises the first type label and the second type label.
Optionally, the video score obtaining request carries a plurality of first type tags and a plurality of second type tags, and the determining module 620 is configured to:
and acquiring at least one song audio of which the number of the first type tags and the second type tags in the corresponding type tags exceeds a preset threshold value.
According to the method and the device, a first type label corresponding to a special effect added in a target video is obtained through a terminal, a video music obtaining request is sent to a server, the server determines at least one song audio matched with the type of the special effect based on the first type label in the video music obtaining request, then a song list comprising at least one song audio option is generated and sent to the terminal, after the terminal receives the song list, a user can trigger a music adding instruction corresponding to the target song audio, and then the terminal sets the target song audio as the music of the target video. Therefore, the terminal can automatically find the score with high matching degree for the target video, so that a user does not need to manually find the score with high matching degree, and the efficiency of finding the score with high matching degree with the image content is improved.
It should be noted that: in the device for setting a score provided in the above embodiment, when setting a score, only the division of the above functional modules is exemplified, and in practical applications, the above function distribution may be completed by different functional modules according to needs, that is, the internal structure of the device is divided into different functional modules to complete all or part of the above described functions. In addition, the method embodiments for setting the score provided by the above embodiments belong to the same concept, and specific implementation processes thereof are described in the method embodiments in detail and are not described herein again.
Fig. 7 shows a block diagram of a terminal 700 provided in an exemplary embodiment of the present application, where the terminal 700 may be the above-mentioned terminal, and the terminal 700 may be a smart phone, a tablet computer, an MP3 player (Moving Picture experts Group Audio L layer III, mpeg Audio layer 3), an MP4 player (Moving Picture experts Group Audio L layer IV, mpeg Audio layer 4), a notebook computer, or a desktop computer.
In general, terminal 700 includes: a processor 701 and a memory 702.
In some embodiments, the terminal 700 may further optionally include: a peripheral interface 703 and at least one peripheral. The processor 701, the memory 702, and the peripheral interface 703 may be connected by buses or signal lines. Various peripheral devices may be connected to peripheral interface 703 via a bus, signal line, or circuit board. Specifically, the peripheral device includes: at least one of radio frequency circuitry 704, touch screen display 705, camera 706, audio circuitry 707, positioning components 708, and power source 709.
The peripheral interface 703 may be used to connect at least one peripheral related to I/O (Input/Output) to the processor 701 and the memory 702. In some embodiments, processor 701, memory 702, and peripheral interface 703 are integrated on the same chip or circuit board; in some other embodiments, any one or two of the processor 701, the memory 702, and the peripheral interface 703 may be implemented on a separate chip or circuit board, which is not limited in this embodiment.
The Radio Frequency circuit 704 is used for receiving and transmitting RF (Radio Frequency) signals, also called electromagnetic signals. The radio frequency circuitry 704 communicates with communication networks and other communication devices via electromagnetic signals. The rf circuit 704 converts an electrical signal into an electromagnetic signal to transmit, or converts a received electromagnetic signal into an electrical signal. Optionally, the radio frequency circuit 704 includes: an antenna system, an RF transceiver, one or more amplifiers, a tuner, an oscillator, a digital signal processor, a codec chipset, a subscriber identity module card, and so forth. The radio frequency circuitry 704 may communicate with other terminals via at least one wireless communication protocol. The wireless communication protocols include, but are not limited to: metropolitan area networks, various generation mobile communication networks (2G, 3G, 4G, and 5G), Wireless local area networks, and/or WiFi (Wireless Fidelity) networks. In some embodiments, the radio frequency circuit 704 may also include NFC (Near Field Communication) related circuits, which are not limited in this application.
The Display 705 is used to Display a UI (User Interface) that may include graphics, text, icons, video, and any combination thereof, when the Display 705 is a touch Display, the Display 705 also has the ability to capture touch signals on or over the surface of the Display 705. the touch signals may be input to the processor 701 for processing as control signals.
The camera assembly 706 is used to capture images or video. Optionally, camera assembly 706 includes a front camera and a rear camera. Generally, a front camera is disposed at a front panel of the terminal, and a rear camera is disposed at a rear surface of the terminal. In some embodiments, the number of the rear cameras is at least two, and each rear camera is any one of a main camera, a depth-of-field camera, a wide-angle camera and a telephoto camera, so that the main camera and the depth-of-field camera are fused to realize a background blurring function, and the main camera and the wide-angle camera are fused to realize panoramic shooting and VR (Virtual Reality) shooting functions or other fusion shooting functions. In some embodiments, camera assembly 706 may also include a flash. The flash lamp can be a monochrome temperature flash lamp or a bicolor temperature flash lamp. The double-color-temperature flash lamp is a combination of a warm-light flash lamp and a cold-light flash lamp, and can be used for light compensation at different color temperatures.
The audio circuitry 707 may include a microphone and a speaker. The microphone is used for collecting sound waves of a user and the environment, converting the sound waves into electric signals, and inputting the electric signals to the processor 701 for processing or inputting the electric signals to the radio frequency circuit 704 to realize voice communication. For the purpose of stereo sound collection or noise reduction, a plurality of microphones may be provided at different portions of the terminal 700. The microphone may also be an array microphone or an omni-directional pick-up microphone. The speaker is used to convert electrical signals from the processor 701 or the radio frequency circuit 704 into sound waves. The loudspeaker can be a traditional film loudspeaker or a piezoelectric ceramic loudspeaker. When the speaker is a piezoelectric ceramic speaker, the speaker can be used for purposes such as converting an electric signal into a sound wave audible to a human being, or converting an electric signal into a sound wave inaudible to a human being to measure a distance. In some embodiments, the audio circuitry 707 may also include a headphone jack.
The positioning component 708 is used to locate the current geographic location of the terminal 700 to implement navigation or L BS (L o geographic based Service.) the positioning component 708 may be a positioning component based on the united states GPS (global positioning System), the beidou System of china, the greiner System of russia, or the galileo System of the european union.
In some embodiments, terminal 700 also includes one or more sensors 710. The one or more sensors 710 include, but are not limited to: acceleration sensor 711, gyro sensor 712, pressure sensor 713, fingerprint sensor 714, optical sensor 715, and proximity sensor 716.
The acceleration sensor 711 can detect the magnitude of acceleration in three coordinate axes of a coordinate system established with the terminal 700. For example, the acceleration sensor 711 may be used to detect components of the gravitational acceleration in three coordinate axes. The processor 701 may control the touch screen 705 to display the user interface in a landscape view or a portrait view according to the gravitational acceleration signal collected by the acceleration sensor 711. The acceleration sensor 711 may also be used for acquisition of motion data of a game or a user.
The gyro sensor 712 may detect a body direction and a rotation angle of the terminal 700, and the gyro sensor 712 may cooperate with the acceleration sensor 711 to acquire a 3D motion of the terminal 700 by the user. From the data collected by the gyro sensor 712, the processor 701 may implement the following functions: motion sensing (such as changing the UI according to a user's tilting operation), image stabilization at the time of photographing, game control, and inertial navigation.
Pressure sensors 713 may be disposed on a side bezel of terminal 700 and/or an underlying layer of touch display 705. When the pressure sensor 713 is disposed on a side frame of the terminal 700, a user's grip signal on the terminal 700 may be detected, and the processor 701 performs right-left hand recognition or shortcut operation according to the grip signal collected by the pressure sensor 713. When the pressure sensor 713 is disposed at a lower layer of the touch display 705, the processor 701 controls the operability control on the UI interface according to the pressure operation of the user on the touch display 705. The operability control comprises at least one of a button control, a scroll bar control, an icon control and a menu control.
The fingerprint sensor 714 is used for collecting fingerprints of a user, the identity of the user is identified by the processor 701 according to the fingerprints collected by the fingerprint sensor 714, or the identity of the user is identified by the fingerprint sensor 714 according to the collected fingerprints, when the identity of the user is identified as a credible identity, the user is authorized to execute relevant sensitive operations by the processor 701, the sensitive operations comprise screen unlocking, encrypted information viewing, software downloading, payment, setting change and the like, the fingerprint sensor 714 can be arranged on the front side, the back side or the side of the terminal 700, when a physical key or a manufacturer L ogo is arranged on the terminal 700, the fingerprint sensor 714 can be integrated with the physical key or the manufacturer L ogo.
The optical sensor 715 is used to collect the ambient light intensity. In one embodiment, the processor 701 may control the display brightness of the touch display 705 based on the ambient light intensity collected by the optical sensor 715. Specifically, when the ambient light intensity is high, the display brightness of the touch display screen 705 is increased; when the ambient light intensity is low, the display brightness of the touch display 705 is turned down. In another embodiment, processor 701 may also dynamically adjust the shooting parameters of camera assembly 706 based on the ambient light intensity collected by optical sensor 715.
A proximity sensor 716, also referred to as a distance sensor, is typically disposed on a front panel of the terminal 700. The proximity sensor 716 is used to collect the distance between the user and the front surface of the terminal 700. In one embodiment, when the proximity sensor 716 detects that the distance between the user and the front surface of the terminal 700 gradually decreases, the processor 701 controls the touch display 705 to switch from the bright screen state to the dark screen state; when the proximity sensor 716 detects that the distance between the user and the front surface of the terminal 700 gradually becomes larger, the processor 701 controls the touch display 705 to switch from the breath screen state to the bright screen state.
Those skilled in the art will appreciate that the configuration shown in fig. 7 is not intended to be limiting of terminal 700 and may include more or fewer components than those shown, or some components may be combined, or a different arrangement of components may be used.
Fig. 8 is a schematic structural diagram of a server according to an embodiment of the present disclosure, where the server may be the above server, and the server 800 may generate a relatively large difference due to a difference in configuration or performance, and may include one or more processors (CPUs) 801 and one or more memories 802, where the memories 802 store at least one instruction, and the at least one instruction is loaded and executed by the processors 801 to implement the methods provided by the above method embodiments. Of course, the server may also have components such as a wired or wireless network interface, a keyboard, and an input/output interface, so as to perform input/output, and the server may also include other components for implementing the functions of the device, which are not described herein again.
In an exemplary embodiment, a computer-readable storage medium, such as a memory, including instructions executable by a processor in a terminal to perform the method of setting a score in the above embodiments is also provided. For example, the computer-readable storage medium may be a Read-only Memory (ROM), a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.
It will be understood by those skilled in the art that all or part of the steps for implementing the above embodiments may be implemented by hardware, or may be implemented by a program instructing relevant hardware, where the program may be stored in a computer-readable storage medium, and the above-mentioned storage medium may be a read-only memory, a magnetic disk or an optical disk, etc.
The above description is only exemplary of the present application and should not be taken as limiting the present application, as any modification, equivalent replacement, or improvement made within the spirit and principle of the present application should be included in the protection scope of the present application.
Claims (17)
1. A method for setting score, which is applied to a terminal, the method comprising:
acquiring a first type label corresponding to a special effect added in a target video;
sending a video score obtaining request to a server, wherein the video score obtaining request carries the first type label;
receiving and displaying a song list sent by the server, wherein the song list comprises at least one song audio option;
and when a music adding instruction corresponding to the target song audio is detected, setting the target song audio as the music of the target video.
2. The method of claim 1, wherein before sending the video score acquisition request to the server, further comprising:
inputting the target video into an image recognition model to obtain a second type label corresponding to the target video;
the video score obtaining request also carries the second type label.
3. The method of claim 1, wherein when the indication of the addition of the score corresponding to the target song audio is detected, setting the target song audio to be before the score of the target video, further comprises:
and when monitoring a listening trial instruction corresponding to the target song audio, playing the target song audio.
4. A method for setting a score, the method being applied to a server, the method comprising:
receiving a video score obtaining request sent by a terminal, wherein the video score obtaining request carries a first type label corresponding to a special effect added in a target video;
determining at least one song audio matching the type of the special effect based on the first type tag;
generating a song list including an option for the at least one song audio;
and sending the song list to the terminal.
5. The method according to claim 4, wherein the video score obtaining request further carries a second type tag corresponding to the target video;
the determining, based on the first type tag, at least one song audio that matches the type of the special effect includes:
determining at least one song audio matching the special effect and the type of the target video based on the first type tag and the second type tag.
6. The method of claim 5, wherein determining at least one song audio that matches the special effect and the type of the target video based on the first type tag and the second type tag comprises:
and acquiring at least one song audio of which the corresponding type label comprises the first type label and the second type label.
7. The method according to claim 6, wherein the video soundtrack acquisition request carries a plurality of first type tags and a plurality of second type tags, and the acquiring at least one song audio of the corresponding type tags that includes the first type tags and the second type tags comprises:
and acquiring at least one song audio of which the number of the first type tags and the second type tags in the corresponding type tags exceeds a preset threshold value.
8. An apparatus for setting a score, the apparatus comprising:
the acquisition module is used for acquiring a first type label corresponding to a special effect added in a target video;
the sending module is used for sending a video score obtaining request to a server, wherein the video score obtaining request carries the first type label;
the display module is used for receiving and displaying a song list sent by the server, wherein the song list comprises at least one song audio option;
and the setting module is used for setting the target song audio as the score of the target video when the score adding instruction corresponding to the target song audio is detected.
9. The apparatus of claim 8, wherein the sending module is further configured to:
inputting the target video into an image recognition model to obtain a second type label corresponding to the target video;
the video score obtaining request also carries the second type label.
10. The apparatus of claim 8, further comprising a monitoring module to:
and when monitoring a listening trial instruction corresponding to the target song audio, playing the target song audio.
11. An apparatus for setting a score, the apparatus comprising:
the receiving module is used for receiving a video score obtaining request sent by a terminal, wherein the video score obtaining request carries a first type label corresponding to a special effect added in a target video;
a determining module for determining at least one song audio matching the type of the special effect based on the first type tag;
a generating module for generating a song list comprising options of the at least one song audio;
and the sending module is used for sending the song list to the terminal.
12. The apparatus according to claim 11, wherein the video score obtaining request further carries a second type tag corresponding to the target video;
the determining module is configured to:
determining at least one song audio matching the special effect and the type of the target video based on the first type tag and the second type tag.
13. The apparatus of claim 12, wherein the determining module is configured to:
and acquiring at least one song audio of which the corresponding type label comprises the first type label and the second type label.
14. The apparatus of claim 13, wherein the video soundtrack acquisition request carries a plurality of first type tags and a plurality of second type tags, and the determining module is configured to:
and acquiring at least one song audio of which the number of the first type tags and the second type tags in the corresponding type tags exceeds a preset threshold value.
15. A system for setting a score, the apparatus comprising a terminal and a server, wherein:
the terminal is used for acquiring a first type tag corresponding to a special effect added in a target video, sending a video score acquisition request to the server, receiving and displaying a song list sent by the server, wherein the song list comprises at least one song audio option, and when a score addition instruction corresponding to the target song audio is detected, setting the target song audio as the score of the target video;
the server is used for receiving a video score acquisition request sent by the terminal, determining at least one song audio matched with the special effect type based on the first type label, generating a song list comprising options of the at least one song audio, and sending the song list to the terminal.
16. A computer device comprising a processor and a memory, the memory having stored therein at least one instruction that is loaded and executed by the processor to perform operations performed by the method of setting a score as claimed in any one of claims 1 to 7.
17. A computer-readable storage medium having stored therein at least one instruction which is loaded and executed by a processor to perform operations performed by the method of setting a score as claimed in any one of claims 1 to 7.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010351173.3A CN111417030A (en) | 2020-04-28 | 2020-04-28 | Method, device, system, equipment and storage equipment for setting score |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010351173.3A CN111417030A (en) | 2020-04-28 | 2020-04-28 | Method, device, system, equipment and storage equipment for setting score |
Publications (1)
Publication Number | Publication Date |
---|---|
CN111417030A true CN111417030A (en) | 2020-07-14 |
Family
ID=71494968
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010351173.3A Pending CN111417030A (en) | 2020-04-28 | 2020-04-28 | Method, device, system, equipment and storage equipment for setting score |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111417030A (en) |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109063163A (en) * | 2018-08-14 | 2018-12-21 | 腾讯科技(深圳)有限公司 | A kind of method, apparatus, terminal device and medium that music is recommended |
CN110704682A (en) * | 2019-09-26 | 2020-01-17 | 新华智云科技有限公司 | Method and system for intelligently recommending background music based on video multidimensional characteristics |
US20200034385A1 (en) * | 2018-07-25 | 2020-01-30 | Omfit LLC | Method and system for creating combined media and user-defined audio selection |
CN110740262A (en) * | 2019-10-31 | 2020-01-31 | 维沃移动通信有限公司 | Background music adding method and device and electronic equipment |
CN110830845A (en) * | 2018-08-09 | 2020-02-21 | 优视科技有限公司 | Video generation method and device and terminal equipment |
CN110958386A (en) * | 2019-11-12 | 2020-04-03 | 北京达佳互联信息技术有限公司 | Video synthesis method and device, electronic equipment and computer-readable storage medium |
CN111031391A (en) * | 2019-12-19 | 2020-04-17 | 北京达佳互联信息技术有限公司 | Video dubbing method, device, server, terminal and storage medium |
CN111061906A (en) * | 2019-11-12 | 2020-04-24 | 北京达佳互联信息技术有限公司 | Music information processing method and device, electronic equipment and computer readable storage medium |
-
2020
- 2020-04-28 CN CN202010351173.3A patent/CN111417030A/en active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20200034385A1 (en) * | 2018-07-25 | 2020-01-30 | Omfit LLC | Method and system for creating combined media and user-defined audio selection |
CN110830845A (en) * | 2018-08-09 | 2020-02-21 | 优视科技有限公司 | Video generation method and device and terminal equipment |
CN109063163A (en) * | 2018-08-14 | 2018-12-21 | 腾讯科技(深圳)有限公司 | A kind of method, apparatus, terminal device and medium that music is recommended |
CN110704682A (en) * | 2019-09-26 | 2020-01-17 | 新华智云科技有限公司 | Method and system for intelligently recommending background music based on video multidimensional characteristics |
CN110740262A (en) * | 2019-10-31 | 2020-01-31 | 维沃移动通信有限公司 | Background music adding method and device and electronic equipment |
CN110958386A (en) * | 2019-11-12 | 2020-04-03 | 北京达佳互联信息技术有限公司 | Video synthesis method and device, electronic equipment and computer-readable storage medium |
CN111061906A (en) * | 2019-11-12 | 2020-04-24 | 北京达佳互联信息技术有限公司 | Music information processing method and device, electronic equipment and computer readable storage medium |
CN111031391A (en) * | 2019-12-19 | 2020-04-17 | 北京达佳互联信息技术有限公司 | Video dubbing method, device, server, terminal and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110674022B (en) | Behavior data acquisition method and device and storage medium | |
CN107908929B (en) | Method and device for playing audio data | |
CN109327608B (en) | Song sharing method, terminal, server and system | |
CN108965757B (en) | Video recording method, device, terminal and storage medium | |
CN108320756B (en) | Method and device for detecting whether audio is pure music audio | |
CN110266982B (en) | Method and system for providing songs while recording video | |
CN111711838B (en) | Video switching method, device, terminal, server and storage medium | |
CN112084811A (en) | Identity information determining method and device and storage medium | |
CN111935516B (en) | Audio file playing method, device, terminal, server and storage medium | |
CN109102811B (en) | Audio fingerprint generation method and device and storage medium | |
CN111416996B (en) | Multimedia file detection method, multimedia file playing device, multimedia file equipment and storage medium | |
CN111081277B (en) | Audio evaluation method, device, equipment and storage medium | |
CN113918767A (en) | Video clip positioning method, device, equipment and storage medium | |
CN110798327B (en) | Message processing method, device and storage medium | |
CN111402844A (en) | Song chorusing method, device and system | |
CN109547847B (en) | Method and device for adding video information and computer readable storage medium | |
CN111092991A (en) | Lyric display method and device and computer storage medium | |
CN112667844A (en) | Method, device, equipment and storage medium for retrieving audio | |
CN109819308B (en) | Virtual resource acquisition method, device, terminal, server and storage medium | |
CN111428080A (en) | Storage method, search method and device for video files | |
CN111641853B (en) | Multimedia resource loading method and device, computer equipment and storage medium | |
CN111063372B (en) | Method, device and equipment for determining pitch characteristics and storage medium | |
CN110277105B (en) | Method, device and system for eliminating background audio data | |
CN114388001A (en) | Multimedia file playing method, device, equipment and storage medium | |
CN109344284B (en) | Song file playing method, device, equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20200714 |
|
RJ01 | Rejection of invention patent application after publication |