CN111417030A - Method, device, system, equipment and storage equipment for setting score - Google Patents

Method, device, system, equipment and storage equipment for setting score Download PDF

Info

Publication number
CN111417030A
CN111417030A CN202010351173.3A CN202010351173A CN111417030A CN 111417030 A CN111417030 A CN 111417030A CN 202010351173 A CN202010351173 A CN 202010351173A CN 111417030 A CN111417030 A CN 111417030A
Authority
CN
China
Prior art keywords
type
video
score
song
song audio
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010351173.3A
Other languages
Chinese (zh)
Inventor
刘春宇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou Kugou Computer Technology Co Ltd
Original Assignee
Guangzhou Kugou Computer Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangzhou Kugou Computer Technology Co Ltd filed Critical Guangzhou Kugou Computer Technology Co Ltd
Priority to CN202010351173.3A priority Critical patent/CN111417030A/en
Publication of CN111417030A publication Critical patent/CN111417030A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8106Monomedia components thereof involving special audio data, e.g. different tracks for different languages
    • H04N21/8113Monomedia components thereof involving special audio data, e.g. different tracks for different languages comprising music, e.g. song in MP3 format
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/233Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/23418Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/262Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission, generating play-lists
    • H04N21/26258Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission, generating play-lists for generating a list of items to be played back in a given order, e.g. playlist, or scheduling item distribution according to such list
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/433Content storage operation, e.g. storage operation in response to a pause request, caching operations
    • H04N21/4334Recording operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/84Generation or processing of descriptive data, e.g. content descriptors

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

The application discloses a method, a device, a system, equipment and storage equipment for setting score, and belongs to the technical field of computers. The method comprises the following steps: acquiring a first type label corresponding to a special effect added in a target video; sending a video score obtaining request to a server, wherein the video score obtaining request carries the first type label; receiving and displaying a song list sent by the server, wherein the song list comprises at least one song audio option; and when a music adding instruction corresponding to the target song audio is detected, setting the target song audio as the music of the target video. The efficiency of finding the score with high matching degree with the image content can be improved through the method and the device.

Description

Method, device, system, equipment and storage equipment for setting score
Technical Field
The present application relates to the field of computer technologies, and in particular, to a method, an apparatus, a system, a device, and a storage device for setting a score.
Background
With the development of computer technology, people are more and more common to use short video applications, and when the short video applications are used:
the user records videos through the terminal, after the recording is completed, the user manually selects one audio frequency from the music library as the music score, and the short video application program synthesizes the shot images and the music score into a short video, so that the short video can be manufactured.
In the process of implementing the present application, the inventor finds that the prior art has at least the following problems:
when the user manually selects the score, it takes a long time for the user to find the score having a high degree of matching with the image content, which is inefficient.
Disclosure of Invention
The embodiment of the application provides a method, a device, a system, equipment and storage equipment for setting score, which can solve the problem of low score efficiency caused by high matching degree with image content. The technical scheme is as follows:
in one aspect, a method for setting a score is provided, and the method is applied to a terminal, and includes:
acquiring a first type label corresponding to a special effect added in a target video;
sending a video score obtaining request to a server, wherein the video score obtaining request carries the first type label;
receiving and displaying a song list sent by the server, wherein the song list comprises at least one song audio option;
and when a music adding instruction corresponding to the target song audio is detected, setting the target song audio as the music of the target video.
Optionally, before sending the video score obtaining request to the server, the method further includes:
inputting the target video into an image recognition model to obtain a second type label corresponding to the target video;
the video score obtaining request also carries the second type label.
Optionally, when the music adding instruction corresponding to the target song audio is detected, before setting the target song audio as the music of the target video, the method further includes:
and when monitoring a listening trial instruction corresponding to the target song audio, playing the target song audio.
In another aspect, a method for setting a score is provided, and the method is applied to a server, and includes:
receiving a video score obtaining request sent by a terminal, wherein the video score obtaining request carries a first type label corresponding to a special effect added in a target video;
determining at least one song audio matching the type of the special effect based on the first type tag;
generating a song list including an option for the at least one song audio;
and sending the song list to the terminal.
Optionally, the video score obtaining request further carries a second type tag corresponding to the target video;
the determining, based on the first type tag, at least one song audio that matches the type of the special effect includes:
determining at least one song audio matching the special effect and the type of the target video based on the first type tag and the second type tag.
Optionally, the determining, based on the first type tag and the second type tag, at least one song audio matching the special effect and the type of the target video includes:
and acquiring at least one song audio of which the corresponding type label comprises the first type label and the second type label.
Optionally, the obtaining of the video score includes that the video score obtaining request carries a plurality of first type tags and a plurality of second type tags, and the obtaining of the corresponding type tags includes at least one song audio of the first type tags and the second type tags, including:
and acquiring at least one song audio of which the number of the first type tags and the second type tags in the corresponding type tags exceeds a preset threshold value.
In another aspect, there is provided an apparatus for setting a score, which may be a terminal, the apparatus comprising:
the acquisition module is used for acquiring a first type label corresponding to a special effect added in a target video;
the sending module is used for sending a video score obtaining request to a server, wherein the video score obtaining request carries the first type label;
the display module is used for receiving and displaying a song list sent by the server, wherein the song list comprises at least one song audio option;
and the setting module is used for setting the target song audio as the score of the target video when the score adding instruction corresponding to the target song audio is detected.
Optionally, the sending module is further configured to:
inputting the target video into an image recognition model to obtain a second type label corresponding to the target video;
the video score obtaining request also carries the second type label.
Optionally, the apparatus further includes a monitoring module, where the monitoring module is configured to:
and when monitoring a listening trial instruction corresponding to the target song audio, playing the target song audio.
In another aspect, there is provided an apparatus for setting a score, which may be a server, the apparatus comprising:
the receiving module is used for receiving a video score obtaining request sent by a terminal, wherein the video score obtaining request carries a first type label corresponding to a special effect added in a target video;
a determining module for determining at least one song audio matching the type of the special effect based on the first type tag;
a generating module for generating a song list comprising options of the at least one song audio;
and the sending module is used for sending the song list to the terminal.
Optionally, the video score obtaining request further carries a second type tag corresponding to the target video;
the determining module is configured to:
determining at least one song audio matching the special effect and the type of the target video based on the first type tag and the second type tag.
Optionally, the determining module is configured to:
and acquiring at least one song audio of which the corresponding type label comprises the first type label and the second type label.
Optionally, the video score obtaining request carries a plurality of first type tags and a plurality of second type tags, and the determining module is configured to:
and acquiring at least one song audio of which the number of the first type tags and the second type tags in the corresponding type tags exceeds a preset threshold value.
In yet another aspect, a computer device is provided, the computer device comprising a processor and a memory, the memory having instructions stored therein, the processor executing the instructions to cause the computer device to implement the method of setting a score.
In yet another aspect, a computer-readable storage medium is provided, the computer-readable storage medium storing instructions, execution of which by a computer device causes the computer device to implement the method of setting a score.
The technical scheme provided by the embodiment of the application has the following beneficial effects:
according to the method and the device, a first type label corresponding to a special effect added in a target video is obtained through a terminal, a video music obtaining request is sent to a server, the server determines at least one song audio matched with the type of the special effect based on the first type label in the video music obtaining request, then a song list comprising at least one song audio option is generated and sent to the terminal, after the terminal receives the song list, a user can trigger a music adding instruction corresponding to the target song audio, and then the terminal sets the target song audio as the music of the target video. Therefore, the terminal can automatically find the score with high matching degree for the target video, so that a user does not need to manually find the score with high matching degree, and the efficiency of finding the score with high matching degree with the image content is improved.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present application, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present application, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
FIG. 1 is a schematic diagram of an implementation environment provided by an embodiment of the present application;
fig. 2 is a flowchart of a terminal side of a method for setting a score according to an embodiment of the present application;
fig. 3 is a server-side flowchart of a method for setting a score according to an embodiment of the present application;
fig. 4 is an interaction flowchart of a method for setting a score according to an embodiment of the present application;
fig. 5 is a schematic structural diagram of a device for setting a score according to an embodiment of the present application;
fig. 6 is a schematic structural diagram of a device for setting a score according to an embodiment of the present application;
fig. 7 is a schematic structural diagram of a terminal provided in an embodiment of the present application;
fig. 8 is a schematic structural diagram of a server according to an embodiment of the present application.
Detailed Description
To make the objects, technical solutions and advantages of the present application more clear, embodiments of the present application will be described in further detail below with reference to the accompanying drawings.
The embodiment of the application provides a method for setting a score, which can be realized by a terminal and a server together. The terminal can be a mobile phone, a desktop computer, a tablet computer, an intelligent wearable device and the like, and can be provided with a screen, a loudspeaker, a camera and the like. The terminal may have a function of transmitting data, and the terminal may be installed with an application program, for example, a short video application program. The server may be a background server of the short video application program, the server may be a single server or a server group, if the server is a single server, the server may be responsible for all processing required by the server in the following scheme, if the server is a server group, different servers in the server group may be responsible for different processing in the following scheme, respectively, and the specific processing allocation condition may be set arbitrarily by a technical person according to actual requirements, which is not described herein again.
When the user uses the short video application program, the user can click the recording interface jump control, and then the terminal can jump to the short video recording interface, and display the video shot by the camera on the short video recording interface, wherein the short video recording interface can be the interface shown in fig. 1. The user can select to click the special effect control on the short video recording interface, and select a special effect, and then the special effect can be displayed at a corresponding position in the video, for example, the special effect is a cat ear special effect, and the cat ear special effect can be displayed at the top of the head of a person. The user can click the video recording control, and then the terminal can start to record and store the video carrying the special effect.
Fig. 2 is a flowchart of a terminal side of a method for setting a score according to an embodiment of the present application. Referring to fig. 2, the process includes:
step 201, obtaining a first type label corresponding to a special effect added in a target video.
Step 202, sending a video score obtaining request to a server.
The video score obtaining request carries a first type label.
And step 203, receiving and displaying the song list sent by the server.
Wherein the song list comprises at least one song audio option.
And step 204, when a music adding instruction corresponding to the target song audio is detected, setting the target song audio as the music of the target video.
Fig. 3 is a server-side flowchart of a method for setting a score according to an embodiment of the present application. Referring to fig. 3, the process includes:
step 301, receiving a video score obtaining request sent by a terminal.
The video score obtaining request carries a first type label corresponding to a special effect added in the target video.
Step 302, determining at least one song audio matched with the type of the special effect based on the first type label.
Step 303 generates a song list including at least one selection of song audio.
Step 304, transmitting the song list to the terminal.
Fig. 4 is an interaction flowchart of a method for setting a score according to an embodiment of the present application. Referring to fig. 4, the process includes:
step 401, the terminal obtains a first type label corresponding to a special effect added in a target video.
In implementation, a user may open a short video application and display a short video recording interface where the user may trigger a special effect control. And then the terminal can display a special effect list on the lower half part of the short video recording interface, the special effect can be divided into an expression special effect and a background special effect, and a user can select any special effect control in the special effect list. The user may also select any number of non-conflicting effect controls in the list of effects, for example, the user selects an emoji effect control and a background effect control. And then the terminal can take the special effect corresponding to the selected special effect control as the added special effect. When the user finishes selecting the special effect control and clicks the video recording control on the short video recording interface, the terminal can input the target video into the image recognition model to obtain the position of the target image, wherein the target image can be a content main body of the target video, for example, a face in the target video. And the terminal displays the special effect at the position of the target image, so that the target video added with the special effect can be obtained, for example, the special effect of the cat ear is displayed on the head of a person in the target video. Then, the terminal obtains a first type label corresponding to the stored special effect according to the corresponding relationship between the special effect and the first type label, where the obtained first type label may be multiple, that is, one special effect may correspond to multiple first type labels.
Optionally, the terminal may further input the target video into the image recognition model to obtain a second type tag corresponding to the target video.
After the target video is input into the image recognition model, the image recognition model may also output a second type tag corresponding to the target video, such as cat, dog, person, weather. And then the terminal may obtain a plurality of second type tags.
Step 402, the terminal sends a video score obtaining request to the server.
The video score obtaining request carries the first type label.
In an implementation, after the processing of step 401 is completed, the terminal may generate a video score obtaining request and send the video score obtaining request to the server.
Optionally, the video score obtaining request may further carry a second type tag.
Step 403, the server determines at least one song audio matching the type of the special effect based on the first type tag, and generates a song list including an option of the at least one song audio.
In implementation, after the server receives the video score obtaining request sent by the terminal, the following processing may be performed:
in a first step, the server determines at least one song audio matching the type of the special effect based on the first type tag.
The server may obtain a plurality of first type tags included in the video soundtrack obtaining request, and obtain a plurality of type tags corresponding to the song audio, where the type tags corresponding to the song audio may also be a plurality of type tags. And further detecting the number of the first type tags contained in the type tags corresponding to the song audios. And if the number is larger than a preset threshold value, acquiring the song audio as the pre-selected song audio.
For example, the video soundtrack acquisition request includes A, B, C three first type tags, and if the preset threshold is 2, song audio including A, B, C, A, B, B, C and A, C in the corresponding type tags is acquired and used as the pre-selected song audio.
Optionally, the specific processing that the server may obtain at least one song audio in which the number of the corresponding type tags, including the first type tag and the second type tag, exceeds the preset threshold may be as follows:
the server may obtain a plurality of first type tags and a plurality of second type tags included in the video soundtrack obtaining request, and obtain a type tag corresponding to the song audio, which may also be a plurality of type tags. And further detecting the number of the first type tags and the number of the second type tags contained in the type tags corresponding to the song audio. And if the two numbers are both larger than a preset threshold value, acquiring the song audio, and taking the song audio as the pre-selected song audio.
For example, the video soundtrack acquisition request includes A, B, C three first type tags and a, b, and c three second type tags, and the preset threshold is 2, then the audio of the song to be selected including any two or three first tags in A, B, C in the corresponding type tags may be acquired first, and then the audio of the song including any two or three second tags in a, b, and c in the type tag corresponding to the audio of the song to be selected is acquired as the audio of the preselected song.
Optionally, the server may further obtain at least one song audio, where the number of the first type tags and the number of the second type tags in the corresponding type tags both exceed a first preset threshold and the total number of the first type tags and the second type tags included in the corresponding type tags exceed a second preset threshold, and the processing may be as follows:
the server may obtain a plurality of first type tags and a plurality of second type tags included in the video soundtrack obtaining request, and obtain a type tag corresponding to the song audio, which may also be a plurality of type tags. And further detecting the number of the first type tags and the number of the second type tags contained in the type tags corresponding to the song audio. And if the two numbers are both larger than a first preset threshold value, acquiring the song audio, and taking the song audio as the audio of the song to be selected. And then acquiring the total number of the first type tags and the second type tags contained in the type tags corresponding to the audio of the song to be selected, judging the relation between the total number and a second preset threshold value, and if the total number is greater than the second preset threshold value, taking the audio of the song to be selected as the audio of the preselected song.
For example, the video score obtaining request includes A, B, C three first type tags and a, b, and c three second type tags, the first preset threshold is 2, and the second preset threshold is 5, then the screened song audio including any two or three first type tags in A, B, C in the corresponding type tags may be obtained first, and then the song audio including any two or three second type tags in a, b, and c in the type tags corresponding to the screened song audio is obtained as the song audio to be selected. Then, a song audio containing A, B, C three tags of the first type and containing any two or three tags of the second type among a, b, c, or a song audio containing A, B, C three tags of the second type among a, b, c, or A, B, C three tags of the first type and containing three tags of the second type among a, b, c is taken as the pre-selected song audio.
In a second step, the server generates a song list including at least one selection of song audio.
After the server acquires the preselected song audio, the user can acquire song attribute information and song links of the preselected song audio, wherein the song attribute information can include the name of a singer, the name of a song and an album. The server may then generate an option to pre-select song audio based on the song attribute information and the song link, and the server may generate a song list of the option to pre-select song audio.
Optionally, the server may further sort the options of the song audio in the song list based on the total number of the first type tag and the second type tag included in the corresponding type tag.
And acquiring the total number of the first type tags and the second type tags in the type tags corresponding to each preselected song audio, and corresponding the total number with the options of the preselected song audio, so that the options of each preselected song audio can be sorted according to the total number.
Optionally, the server may also sort the options of the song audio in the song list based on the number of times the acquired respective song audio is selected as the score.
Each time a preselected song audio is selected as a soundtrack, the server may record an increase of 1 in the number of times the preselected song audio is selected as a soundtrack. And after the server acquires the audio frequency of each preselected song, the server can sort the options of the preselected song audio frequency in the song list according to the corresponding frequency of the preselected song audio frequency selected as the score, and further obtain the sorted song list.
Step 404, the server sends the song list to the terminal.
In an implementation, after generating a song list including at least one option to pre-select song audio, the server may transmit the song list to the terminal.
Step 405, the terminal displays the song list.
Wherein the song list includes at least one option to pre-select song audio.
In implementation, after receiving the displayed song list sent by the server, the terminal may display the song list, and the options in the list may display information such as the name of the singer, the name of the song, and the album.
And step 406, when the terminal detects a music adding instruction corresponding to the target song audio, setting the target song audio as the music of the target video.
In implementation, after the terminal displays the song list, the user can trigger any one of the options of the pre-selected song audio in the song list, and the terminal determines the option of the pre-selected song audio as the option of the target song audio. Then, the user can click the determination control and generate a score adding instruction corresponding to the target song audio, and the terminal can set the target song audio as the score of the target video according to the score adding instruction.
Optionally, after the terminal determines the option of the preselected song audio as the option of the target song audio, when a listening trial instruction corresponding to the target song audio is monitored, the target song audio is played.
After the terminal determines the option of the pre-selected song audio as the option of the target song audio, a listening trial instruction may be triggered. When the terminal monitors a listening trial instruction corresponding to the target song audio, the terminal can access the song link corresponding to the option, acquire song audio data, decode the song audio data, transmit the decoded song audio data to the loudspeaker, and then the terminal can play the target song audio.
After the above steps are completed, the short video application may cut the score based on the time length of the recorded video, for example, if the time length of the recorded video is 30 seconds, the first 30 seconds of cutting the score correspond to the recorded video. And further synthesizing the cut soundtrack, the recorded video and the special effect into a short video.
According to the method and the device, a first type label corresponding to a special effect added in a target video is obtained through a terminal, a video music obtaining request is sent to a server, the server determines at least one song audio matched with the type of the special effect based on the first type label in the video music obtaining request, then a song list comprising at least one song audio option is generated and sent to the terminal, after the terminal receives the song list, a user can trigger a music adding instruction corresponding to the target song audio, and then the terminal sets the target song audio as the music of the target video. Therefore, the terminal can automatically find the score with high matching degree for the target video, so that a user does not need to manually find the score with high matching degree, and the efficiency of finding the score with high matching degree with the image content is improved.
All the above optional technical solutions may be combined arbitrarily to form the optional embodiments of the present disclosure, and are not described herein again.
Fig. 5 is an image display device according to an embodiment of the present disclosure, which may be the terminal.
Referring to fig. 5, the apparatus includes:
an obtaining module 510, configured to obtain a first type tag corresponding to a special effect added in a target video;
a sending module 520, configured to send a video score obtaining request to a server, where the video score obtaining request carries the first type tag;
a display module 530, configured to receive and display a song list sent by the server, where the song list includes at least one song audio option;
and the setting module 540 is configured to set the target song audio as the score of the target video when the score adding instruction corresponding to the target song audio is detected.
Optionally, the sending module 520 is further configured to:
inputting the target video into an image recognition model to obtain a second type label corresponding to the target video;
the video score obtaining request also carries the second type label.
Optionally, the apparatus further includes a monitoring module, where the monitoring module is configured to:
and when monitoring a listening trial instruction corresponding to the target song audio, playing the target song audio.
Fig. 6 is an apparatus for displaying an image according to an embodiment of the present application, where the apparatus may be the server. Referring to fig. 6, the apparatus includes:
the receiving module 610 is configured to receive a video score obtaining request sent by a terminal, where the video score obtaining request carries a first type tag corresponding to a special effect added in a target video;
a determining module 620, configured to determine at least one song audio matching the type of the special effect based on the first type tag;
a generating module 630 for generating a song list comprising options of the at least one song audio;
a sending module 640, configured to send the song list to the terminal.
Optionally, the video score obtaining request further carries a second type tag corresponding to the target video;
the determining module 620 is configured to:
determining at least one song audio matching the special effect and the type of the target video based on the first type tag and the second type tag.
Optionally, the determining module 620 is configured to:
and acquiring at least one song audio of which the corresponding type label comprises the first type label and the second type label.
Optionally, the video score obtaining request carries a plurality of first type tags and a plurality of second type tags, and the determining module 620 is configured to:
and acquiring at least one song audio of which the number of the first type tags and the second type tags in the corresponding type tags exceeds a preset threshold value.
According to the method and the device, a first type label corresponding to a special effect added in a target video is obtained through a terminal, a video music obtaining request is sent to a server, the server determines at least one song audio matched with the type of the special effect based on the first type label in the video music obtaining request, then a song list comprising at least one song audio option is generated and sent to the terminal, after the terminal receives the song list, a user can trigger a music adding instruction corresponding to the target song audio, and then the terminal sets the target song audio as the music of the target video. Therefore, the terminal can automatically find the score with high matching degree for the target video, so that a user does not need to manually find the score with high matching degree, and the efficiency of finding the score with high matching degree with the image content is improved.
It should be noted that: in the device for setting a score provided in the above embodiment, when setting a score, only the division of the above functional modules is exemplified, and in practical applications, the above function distribution may be completed by different functional modules according to needs, that is, the internal structure of the device is divided into different functional modules to complete all or part of the above described functions. In addition, the method embodiments for setting the score provided by the above embodiments belong to the same concept, and specific implementation processes thereof are described in the method embodiments in detail and are not described herein again.
Fig. 7 shows a block diagram of a terminal 700 provided in an exemplary embodiment of the present application, where the terminal 700 may be the above-mentioned terminal, and the terminal 700 may be a smart phone, a tablet computer, an MP3 player (Moving Picture experts Group Audio L layer III, mpeg Audio layer 3), an MP4 player (Moving Picture experts Group Audio L layer IV, mpeg Audio layer 4), a notebook computer, or a desktop computer.
In general, terminal 700 includes: a processor 701 and a memory 702.
Processor 701 may include one or more Processing cores, such as a 4-core processor, an 8-core processor, etc. processor 701 may be implemented in at least one hardware form of a DSP (Digital Signal Processing), an FPGA (Field-Programmable Gate Array), a P L a (Programmable logic Array), processor 701 may also include a main processor and a coprocessor, the main processor being a processor for Processing data in a wake-up state, also known as a CPU (Central Processing Unit), the coprocessor being a low-power processor for Processing data in a standby state, in some embodiments, processor 701 may be integrated with a GPU (Graphics Processing Unit) for rendering and rendering content for display, in some embodiments, processor 701 may also include an AI (intelligent processor) for learning operations related to an AI (Artificial Intelligence processor) for computing operations related to display screens.
Memory 702 may include one or more computer-readable storage media, which may be non-transitory. Memory 702 may also include high-speed random access memory, as well as non-volatile memory, such as one or more magnetic disk storage devices, flash memory storage devices. In some embodiments, a non-transitory computer readable storage medium in memory 702 is used to store at least one instruction for execution by processor 701 to implement a method of setting a score as provided by method embodiments herein.
In some embodiments, the terminal 700 may further optionally include: a peripheral interface 703 and at least one peripheral. The processor 701, the memory 702, and the peripheral interface 703 may be connected by buses or signal lines. Various peripheral devices may be connected to peripheral interface 703 via a bus, signal line, or circuit board. Specifically, the peripheral device includes: at least one of radio frequency circuitry 704, touch screen display 705, camera 706, audio circuitry 707, positioning components 708, and power source 709.
The peripheral interface 703 may be used to connect at least one peripheral related to I/O (Input/Output) to the processor 701 and the memory 702. In some embodiments, processor 701, memory 702, and peripheral interface 703 are integrated on the same chip or circuit board; in some other embodiments, any one or two of the processor 701, the memory 702, and the peripheral interface 703 may be implemented on a separate chip or circuit board, which is not limited in this embodiment.
The Radio Frequency circuit 704 is used for receiving and transmitting RF (Radio Frequency) signals, also called electromagnetic signals. The radio frequency circuitry 704 communicates with communication networks and other communication devices via electromagnetic signals. The rf circuit 704 converts an electrical signal into an electromagnetic signal to transmit, or converts a received electromagnetic signal into an electrical signal. Optionally, the radio frequency circuit 704 includes: an antenna system, an RF transceiver, one or more amplifiers, a tuner, an oscillator, a digital signal processor, a codec chipset, a subscriber identity module card, and so forth. The radio frequency circuitry 704 may communicate with other terminals via at least one wireless communication protocol. The wireless communication protocols include, but are not limited to: metropolitan area networks, various generation mobile communication networks (2G, 3G, 4G, and 5G), Wireless local area networks, and/or WiFi (Wireless Fidelity) networks. In some embodiments, the radio frequency circuit 704 may also include NFC (Near Field Communication) related circuits, which are not limited in this application.
The Display 705 is used to Display a UI (User Interface) that may include graphics, text, icons, video, and any combination thereof, when the Display 705 is a touch Display, the Display 705 also has the ability to capture touch signals on or over the surface of the Display 705. the touch signals may be input to the processor 701 for processing as control signals.
The camera assembly 706 is used to capture images or video. Optionally, camera assembly 706 includes a front camera and a rear camera. Generally, a front camera is disposed at a front panel of the terminal, and a rear camera is disposed at a rear surface of the terminal. In some embodiments, the number of the rear cameras is at least two, and each rear camera is any one of a main camera, a depth-of-field camera, a wide-angle camera and a telephoto camera, so that the main camera and the depth-of-field camera are fused to realize a background blurring function, and the main camera and the wide-angle camera are fused to realize panoramic shooting and VR (Virtual Reality) shooting functions or other fusion shooting functions. In some embodiments, camera assembly 706 may also include a flash. The flash lamp can be a monochrome temperature flash lamp or a bicolor temperature flash lamp. The double-color-temperature flash lamp is a combination of a warm-light flash lamp and a cold-light flash lamp, and can be used for light compensation at different color temperatures.
The audio circuitry 707 may include a microphone and a speaker. The microphone is used for collecting sound waves of a user and the environment, converting the sound waves into electric signals, and inputting the electric signals to the processor 701 for processing or inputting the electric signals to the radio frequency circuit 704 to realize voice communication. For the purpose of stereo sound collection or noise reduction, a plurality of microphones may be provided at different portions of the terminal 700. The microphone may also be an array microphone or an omni-directional pick-up microphone. The speaker is used to convert electrical signals from the processor 701 or the radio frequency circuit 704 into sound waves. The loudspeaker can be a traditional film loudspeaker or a piezoelectric ceramic loudspeaker. When the speaker is a piezoelectric ceramic speaker, the speaker can be used for purposes such as converting an electric signal into a sound wave audible to a human being, or converting an electric signal into a sound wave inaudible to a human being to measure a distance. In some embodiments, the audio circuitry 707 may also include a headphone jack.
The positioning component 708 is used to locate the current geographic location of the terminal 700 to implement navigation or L BS (L o geographic based Service.) the positioning component 708 may be a positioning component based on the united states GPS (global positioning System), the beidou System of china, the greiner System of russia, or the galileo System of the european union.
Power supply 709 is provided to supply power to various components of terminal 700. The power source 709 may be alternating current, direct current, disposable batteries, or rechargeable batteries. When power source 709 includes a rechargeable battery, the rechargeable battery may support wired or wireless charging. The rechargeable battery may also be used to support fast charge technology.
In some embodiments, terminal 700 also includes one or more sensors 710. The one or more sensors 710 include, but are not limited to: acceleration sensor 711, gyro sensor 712, pressure sensor 713, fingerprint sensor 714, optical sensor 715, and proximity sensor 716.
The acceleration sensor 711 can detect the magnitude of acceleration in three coordinate axes of a coordinate system established with the terminal 700. For example, the acceleration sensor 711 may be used to detect components of the gravitational acceleration in three coordinate axes. The processor 701 may control the touch screen 705 to display the user interface in a landscape view or a portrait view according to the gravitational acceleration signal collected by the acceleration sensor 711. The acceleration sensor 711 may also be used for acquisition of motion data of a game or a user.
The gyro sensor 712 may detect a body direction and a rotation angle of the terminal 700, and the gyro sensor 712 may cooperate with the acceleration sensor 711 to acquire a 3D motion of the terminal 700 by the user. From the data collected by the gyro sensor 712, the processor 701 may implement the following functions: motion sensing (such as changing the UI according to a user's tilting operation), image stabilization at the time of photographing, game control, and inertial navigation.
Pressure sensors 713 may be disposed on a side bezel of terminal 700 and/or an underlying layer of touch display 705. When the pressure sensor 713 is disposed on a side frame of the terminal 700, a user's grip signal on the terminal 700 may be detected, and the processor 701 performs right-left hand recognition or shortcut operation according to the grip signal collected by the pressure sensor 713. When the pressure sensor 713 is disposed at a lower layer of the touch display 705, the processor 701 controls the operability control on the UI interface according to the pressure operation of the user on the touch display 705. The operability control comprises at least one of a button control, a scroll bar control, an icon control and a menu control.
The fingerprint sensor 714 is used for collecting fingerprints of a user, the identity of the user is identified by the processor 701 according to the fingerprints collected by the fingerprint sensor 714, or the identity of the user is identified by the fingerprint sensor 714 according to the collected fingerprints, when the identity of the user is identified as a credible identity, the user is authorized to execute relevant sensitive operations by the processor 701, the sensitive operations comprise screen unlocking, encrypted information viewing, software downloading, payment, setting change and the like, the fingerprint sensor 714 can be arranged on the front side, the back side or the side of the terminal 700, when a physical key or a manufacturer L ogo is arranged on the terminal 700, the fingerprint sensor 714 can be integrated with the physical key or the manufacturer L ogo.
The optical sensor 715 is used to collect the ambient light intensity. In one embodiment, the processor 701 may control the display brightness of the touch display 705 based on the ambient light intensity collected by the optical sensor 715. Specifically, when the ambient light intensity is high, the display brightness of the touch display screen 705 is increased; when the ambient light intensity is low, the display brightness of the touch display 705 is turned down. In another embodiment, processor 701 may also dynamically adjust the shooting parameters of camera assembly 706 based on the ambient light intensity collected by optical sensor 715.
A proximity sensor 716, also referred to as a distance sensor, is typically disposed on a front panel of the terminal 700. The proximity sensor 716 is used to collect the distance between the user and the front surface of the terminal 700. In one embodiment, when the proximity sensor 716 detects that the distance between the user and the front surface of the terminal 700 gradually decreases, the processor 701 controls the touch display 705 to switch from the bright screen state to the dark screen state; when the proximity sensor 716 detects that the distance between the user and the front surface of the terminal 700 gradually becomes larger, the processor 701 controls the touch display 705 to switch from the breath screen state to the bright screen state.
Those skilled in the art will appreciate that the configuration shown in fig. 7 is not intended to be limiting of terminal 700 and may include more or fewer components than those shown, or some components may be combined, or a different arrangement of components may be used.
Fig. 8 is a schematic structural diagram of a server according to an embodiment of the present disclosure, where the server may be the above server, and the server 800 may generate a relatively large difference due to a difference in configuration or performance, and may include one or more processors (CPUs) 801 and one or more memories 802, where the memories 802 store at least one instruction, and the at least one instruction is loaded and executed by the processors 801 to implement the methods provided by the above method embodiments. Of course, the server may also have components such as a wired or wireless network interface, a keyboard, and an input/output interface, so as to perform input/output, and the server may also include other components for implementing the functions of the device, which are not described herein again.
In an exemplary embodiment, a computer-readable storage medium, such as a memory, including instructions executable by a processor in a terminal to perform the method of setting a score in the above embodiments is also provided. For example, the computer-readable storage medium may be a Read-only Memory (ROM), a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.
It will be understood by those skilled in the art that all or part of the steps for implementing the above embodiments may be implemented by hardware, or may be implemented by a program instructing relevant hardware, where the program may be stored in a computer-readable storage medium, and the above-mentioned storage medium may be a read-only memory, a magnetic disk or an optical disk, etc.
The above description is only exemplary of the present application and should not be taken as limiting the present application, as any modification, equivalent replacement, or improvement made within the spirit and principle of the present application should be included in the protection scope of the present application.

Claims (17)

1. A method for setting score, which is applied to a terminal, the method comprising:
acquiring a first type label corresponding to a special effect added in a target video;
sending a video score obtaining request to a server, wherein the video score obtaining request carries the first type label;
receiving and displaying a song list sent by the server, wherein the song list comprises at least one song audio option;
and when a music adding instruction corresponding to the target song audio is detected, setting the target song audio as the music of the target video.
2. The method of claim 1, wherein before sending the video score acquisition request to the server, further comprising:
inputting the target video into an image recognition model to obtain a second type label corresponding to the target video;
the video score obtaining request also carries the second type label.
3. The method of claim 1, wherein when the indication of the addition of the score corresponding to the target song audio is detected, setting the target song audio to be before the score of the target video, further comprises:
and when monitoring a listening trial instruction corresponding to the target song audio, playing the target song audio.
4. A method for setting a score, the method being applied to a server, the method comprising:
receiving a video score obtaining request sent by a terminal, wherein the video score obtaining request carries a first type label corresponding to a special effect added in a target video;
determining at least one song audio matching the type of the special effect based on the first type tag;
generating a song list including an option for the at least one song audio;
and sending the song list to the terminal.
5. The method according to claim 4, wherein the video score obtaining request further carries a second type tag corresponding to the target video;
the determining, based on the first type tag, at least one song audio that matches the type of the special effect includes:
determining at least one song audio matching the special effect and the type of the target video based on the first type tag and the second type tag.
6. The method of claim 5, wherein determining at least one song audio that matches the special effect and the type of the target video based on the first type tag and the second type tag comprises:
and acquiring at least one song audio of which the corresponding type label comprises the first type label and the second type label.
7. The method according to claim 6, wherein the video soundtrack acquisition request carries a plurality of first type tags and a plurality of second type tags, and the acquiring at least one song audio of the corresponding type tags that includes the first type tags and the second type tags comprises:
and acquiring at least one song audio of which the number of the first type tags and the second type tags in the corresponding type tags exceeds a preset threshold value.
8. An apparatus for setting a score, the apparatus comprising:
the acquisition module is used for acquiring a first type label corresponding to a special effect added in a target video;
the sending module is used for sending a video score obtaining request to a server, wherein the video score obtaining request carries the first type label;
the display module is used for receiving and displaying a song list sent by the server, wherein the song list comprises at least one song audio option;
and the setting module is used for setting the target song audio as the score of the target video when the score adding instruction corresponding to the target song audio is detected.
9. The apparatus of claim 8, wherein the sending module is further configured to:
inputting the target video into an image recognition model to obtain a second type label corresponding to the target video;
the video score obtaining request also carries the second type label.
10. The apparatus of claim 8, further comprising a monitoring module to:
and when monitoring a listening trial instruction corresponding to the target song audio, playing the target song audio.
11. An apparatus for setting a score, the apparatus comprising:
the receiving module is used for receiving a video score obtaining request sent by a terminal, wherein the video score obtaining request carries a first type label corresponding to a special effect added in a target video;
a determining module for determining at least one song audio matching the type of the special effect based on the first type tag;
a generating module for generating a song list comprising options of the at least one song audio;
and the sending module is used for sending the song list to the terminal.
12. The apparatus according to claim 11, wherein the video score obtaining request further carries a second type tag corresponding to the target video;
the determining module is configured to:
determining at least one song audio matching the special effect and the type of the target video based on the first type tag and the second type tag.
13. The apparatus of claim 12, wherein the determining module is configured to:
and acquiring at least one song audio of which the corresponding type label comprises the first type label and the second type label.
14. The apparatus of claim 13, wherein the video soundtrack acquisition request carries a plurality of first type tags and a plurality of second type tags, and the determining module is configured to:
and acquiring at least one song audio of which the number of the first type tags and the second type tags in the corresponding type tags exceeds a preset threshold value.
15. A system for setting a score, the apparatus comprising a terminal and a server, wherein:
the terminal is used for acquiring a first type tag corresponding to a special effect added in a target video, sending a video score acquisition request to the server, receiving and displaying a song list sent by the server, wherein the song list comprises at least one song audio option, and when a score addition instruction corresponding to the target song audio is detected, setting the target song audio as the score of the target video;
the server is used for receiving a video score acquisition request sent by the terminal, determining at least one song audio matched with the special effect type based on the first type label, generating a song list comprising options of the at least one song audio, and sending the song list to the terminal.
16. A computer device comprising a processor and a memory, the memory having stored therein at least one instruction that is loaded and executed by the processor to perform operations performed by the method of setting a score as claimed in any one of claims 1 to 7.
17. A computer-readable storage medium having stored therein at least one instruction which is loaded and executed by a processor to perform operations performed by the method of setting a score as claimed in any one of claims 1 to 7.
CN202010351173.3A 2020-04-28 2020-04-28 Method, device, system, equipment and storage equipment for setting score Pending CN111417030A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010351173.3A CN111417030A (en) 2020-04-28 2020-04-28 Method, device, system, equipment and storage equipment for setting score

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010351173.3A CN111417030A (en) 2020-04-28 2020-04-28 Method, device, system, equipment and storage equipment for setting score

Publications (1)

Publication Number Publication Date
CN111417030A true CN111417030A (en) 2020-07-14

Family

ID=71494968

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010351173.3A Pending CN111417030A (en) 2020-04-28 2020-04-28 Method, device, system, equipment and storage equipment for setting score

Country Status (1)

Country Link
CN (1) CN111417030A (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109063163A (en) * 2018-08-14 2018-12-21 腾讯科技(深圳)有限公司 A kind of method, apparatus, terminal device and medium that music is recommended
CN110704682A (en) * 2019-09-26 2020-01-17 新华智云科技有限公司 Method and system for intelligently recommending background music based on video multidimensional characteristics
US20200034385A1 (en) * 2018-07-25 2020-01-30 Omfit LLC Method and system for creating combined media and user-defined audio selection
CN110740262A (en) * 2019-10-31 2020-01-31 维沃移动通信有限公司 Background music adding method and device and electronic equipment
CN110830845A (en) * 2018-08-09 2020-02-21 优视科技有限公司 Video generation method and device and terminal equipment
CN110958386A (en) * 2019-11-12 2020-04-03 北京达佳互联信息技术有限公司 Video synthesis method and device, electronic equipment and computer-readable storage medium
CN111031391A (en) * 2019-12-19 2020-04-17 北京达佳互联信息技术有限公司 Video dubbing method, device, server, terminal and storage medium
CN111061906A (en) * 2019-11-12 2020-04-24 北京达佳互联信息技术有限公司 Music information processing method and device, electronic equipment and computer readable storage medium

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200034385A1 (en) * 2018-07-25 2020-01-30 Omfit LLC Method and system for creating combined media and user-defined audio selection
CN110830845A (en) * 2018-08-09 2020-02-21 优视科技有限公司 Video generation method and device and terminal equipment
CN109063163A (en) * 2018-08-14 2018-12-21 腾讯科技(深圳)有限公司 A kind of method, apparatus, terminal device and medium that music is recommended
CN110704682A (en) * 2019-09-26 2020-01-17 新华智云科技有限公司 Method and system for intelligently recommending background music based on video multidimensional characteristics
CN110740262A (en) * 2019-10-31 2020-01-31 维沃移动通信有限公司 Background music adding method and device and electronic equipment
CN110958386A (en) * 2019-11-12 2020-04-03 北京达佳互联信息技术有限公司 Video synthesis method and device, electronic equipment and computer-readable storage medium
CN111061906A (en) * 2019-11-12 2020-04-24 北京达佳互联信息技术有限公司 Music information processing method and device, electronic equipment and computer readable storage medium
CN111031391A (en) * 2019-12-19 2020-04-17 北京达佳互联信息技术有限公司 Video dubbing method, device, server, terminal and storage medium

Similar Documents

Publication Publication Date Title
CN110674022B (en) Behavior data acquisition method and device and storage medium
CN107908929B (en) Method and device for playing audio data
CN109327608B (en) Song sharing method, terminal, server and system
CN108965757B (en) Video recording method, device, terminal and storage medium
CN108320756B (en) Method and device for detecting whether audio is pure music audio
CN110266982B (en) Method and system for providing songs while recording video
CN111711838B (en) Video switching method, device, terminal, server and storage medium
CN112084811A (en) Identity information determining method and device and storage medium
CN111935516B (en) Audio file playing method, device, terminal, server and storage medium
CN109102811B (en) Audio fingerprint generation method and device and storage medium
CN111416996B (en) Multimedia file detection method, multimedia file playing device, multimedia file equipment and storage medium
CN111081277B (en) Audio evaluation method, device, equipment and storage medium
CN113918767A (en) Video clip positioning method, device, equipment and storage medium
CN110798327B (en) Message processing method, device and storage medium
CN111402844A (en) Song chorusing method, device and system
CN109547847B (en) Method and device for adding video information and computer readable storage medium
CN111092991A (en) Lyric display method and device and computer storage medium
CN112667844A (en) Method, device, equipment and storage medium for retrieving audio
CN109819308B (en) Virtual resource acquisition method, device, terminal, server and storage medium
CN111428080A (en) Storage method, search method and device for video files
CN111641853B (en) Multimedia resource loading method and device, computer equipment and storage medium
CN111063372B (en) Method, device and equipment for determining pitch characteristics and storage medium
CN110277105B (en) Method, device and system for eliminating background audio data
CN114388001A (en) Multimedia file playing method, device, equipment and storage medium
CN109344284B (en) Song file playing method, device, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20200714

RJ01 Rejection of invention patent application after publication