CN113076444A - Song identification method and device, electronic equipment and storage medium - Google Patents

Song identification method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN113076444A
CN113076444A CN202110351517.5A CN202110351517A CN113076444A CN 113076444 A CN113076444 A CN 113076444A CN 202110351517 A CN202110351517 A CN 202110351517A CN 113076444 A CN113076444 A CN 113076444A
Authority
CN
China
Prior art keywords
song
input
songs
library
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110351517.5A
Other languages
Chinese (zh)
Inventor
陈兵
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vivo Mobile Communication Co Ltd
Original Assignee
Vivo Mobile Communication Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vivo Mobile Communication Co Ltd filed Critical Vivo Mobile Communication Co Ltd
Priority to CN202110351517.5A priority Critical patent/CN113076444A/en
Publication of CN113076444A publication Critical patent/CN113076444A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/63Querying
    • G06F16/632Query formulation

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The application discloses a song identification method and device, electronic equipment and a storage medium, and belongs to the technical field of electronics. The method comprises the following steps: receiving a first input under the condition that the electronic equipment is in a target state, wherein the content of the first input is first audio; in response to the first input, outputting target song information of the first audio based on first text corresponding to the first audio determined by a voice assistant of the electronic equipment; the target state includes at least one of a screen-saving state and a use state.

Description

Song identification method and device, electronic equipment and storage medium
Technical Field
The application belongs to the technical field of electronics, and particularly relates to a song identification method and device, electronic equipment and a storage medium.
Background
When people listen to a good-listening song but not a good-listening song on the street, the square or other scenes, the electronic equipment such as a mobile phone can be taken out to recognize the song by using the function of listening to the song.
However, sometimes the user has not yet had time to take out the electronic device such as a mobile phone to start identification, and the song is already put out, so that the song which the user needs to identify cannot be identified in time.
Disclosure of Invention
The embodiment of the application provides a song identification method and device, electronic equipment and a storage medium, which can realize timely and automatic identification of songs played in the environment where the electronic equipment is located.
In a first aspect, an embodiment of the present application provides a song identification method, including:
receiving a first input under the condition that the electronic equipment is in a target state, wherein the content of the first input is first audio;
in response to the first input, outputting target song information of the first audio based on first text corresponding to the first audio determined by a voice assistant of the electronic equipment;
the target state includes at least one of a screen-saving state and a use state.
In a second aspect, an embodiment of the present application provides a song recognition apparatus, including:
the electronic equipment comprises a first receiving module, a second receiving module and a display module, wherein the first receiving module is used for receiving a first input under the condition that the electronic equipment is in a target state, and the content of the first input is first audio;
a first output module, configured to output, in response to the first input, target song information of the first audio based on a first text corresponding to the first audio determined by a voice assistant of the electronic device;
the target state includes at least one of a screen-saving state and a use state.
In a third aspect, an embodiment of the present application provides an electronic device, which includes a processor, a memory, and a program or instructions stored on the memory and executable on the processor, and when executed by the processor, the program or instructions implement the steps of the method according to the first aspect.
In a fourth aspect, embodiments of the present application provide a readable storage medium, on which a program or instructions are stored, which when executed by a processor implement the steps of the method according to the first aspect.
In a fifth aspect, an embodiment of the present application provides a chip, where the chip includes a processor and a communication interface, where the communication interface is coupled to the processor, and the processor is configured to execute a program or instructions to implement the method according to the first aspect.
In the embodiment of the application, the audio in the current environment is received in the screen-off state or the use state of the electronic equipment, the audio is identified by the voice assistant, the song information played in the current environment is determined and displayed, the intelligent song listening and identification of the voice assistant are realized, the song played in the environment where the electronic equipment is located is timely and automatically identified, and the user is prevented from missing the song.
Drawings
Fig. 1 is a schematic flowchart of a song identification method provided in an embodiment of the present application;
fig. 2 is a schematic structural diagram of a song recognition apparatus provided in an embodiment of the present application;
fig. 3 is a schematic hardware structure diagram of an electronic device according to an embodiment of the present disclosure;
fig. 4 is a second schematic diagram of a hardware structure of an electronic device according to an embodiment of the present disclosure.
Detailed Description
The technical solutions in the embodiments of the present application will be described clearly below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some, but not all, embodiments of the present application. All other embodiments that can be derived from the embodiments given herein by a person of ordinary skill in the art are intended to be within the scope of the present disclosure.
The terms first, second and the like in the description and in the claims of the present application are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the application are capable of operation in sequences other than those illustrated or described herein. In addition, "and/or" in the specification and claims means at least one of connected objects, a character "/" generally means that a preceding and succeeding related objects are in an "or" relationship.
The song recognition method and device provided by the embodiments of the present application are described in detail below with reference to the accompanying drawings through some embodiments and application scenarios thereof.
Fig. 1 is a schematic flowchart of a song identification method provided in an embodiment of the present application, and as shown in fig. 1, the method includes the following steps:
step 100, receiving a first input under the condition that the electronic equipment is in a target state, wherein the content of the first input is first audio;
step 110, in response to the first input, outputting target song information of the first audio based on a first text corresponding to the first audio determined by a voice assistant of the electronic equipment;
the target state includes at least one of a screen-saving state and a use state.
Songs play an important role in life and work of people, can relieve stress, can joyfully oneself, are companions when traveling on a journey, and are lyric intermediaries of emotion. Everyone also can have the song list of oneself, and the song list also can slowly enrich, for example when hearing the song that oneself wanted to listen, can take out electronic equipment and use the function of listening to the song and knowing the song.
When people hear a good-hearing but unknown song on the street, square or other outdoor scenes, the electronic equipment can be taken out to use the function of listening to the song and recognizing the song. However, sometimes the song may be drunk in the singing voice, the electronic device has not yet to be taken out, and the song list of the user can be added with a good song without reason.
Alternatively, the electronic device may include, but is not limited to:
the non-mobile electronic device may be a server, a Network Attached Storage (NAS), a Personal Computer (PC), a television (television, TV), a teller machine or a self-service machine, and the embodiments of the present application are not limited in particular.
Therefore, the lyrics of the climax part of some possible songs can be recorded into the awakening words of the intelligent voice assistant of the electronic equipment, and when the first text (namely, the lyrics of the song being played in the current scene) corresponding to the first audio is completely matched with the awakening words, the voice assistant can identify the song and output the song to the user.
Optionally, when the target song information is output to the user, the user can be actively pushed the song just recognized under the opening of the intelligent scene, or the user wakes up the voice assistant to actively ask what song just played and answers the user, and this way can effectively avoid the user missing the recognition of the good-hearing song.
Optionally, in the screen-off state and the use state of the electronic device, as long as music is played in the environment, the song can be identified, and the target song information is output, so that the situation that the user is finished before starting the electronic device to identify the song is avoided.
Alternatively, the breath-screen state of the electronic device may include, but is not limited to, a sleep state, a lock screen state, and a screen-off but active state of the electronic device.
Alternatively, the first input may be a song being played in the environment surrounding the electronic device.
Alternatively, the first input may be considered received when the electronic device recognizes that a song is playing in the surrounding environment.
In the embodiment of the application, the audio in the current environment is received in the screen-off state or the use state of the electronic equipment, the audio is identified by the voice assistant, the song information played in the current environment is determined and displayed, the intelligent song listening and identification of the voice assistant are realized, the song played in the environment where the electronic equipment is located is timely and automatically identified, and the user is prevented from missing the song.
Optionally, the outputting the target song information of the first audio based on the first text corresponding to the first audio includes:
taking a first song to which first lyrics matched with the first text belong as a target song, wherein the first song is one of songs in a reference song library;
and displaying the target song information.
Alternatively, a reference song library may be created and all or part of the lyrics of the songs in the reference song library may be saved as wake-up words to be entered into the voice assistant as wake-up words for the target song recognition.
Optionally, if the first text obtained by the voice assistant based on the first audio, that is, the lyrics of the song played in the environment where the user is located, completely matches the lyrics recorded in a certain song in the reference song library, the voice assistant may be considered to capture the lyrics of the climax part, the intelligent song listening recognition function is triggered (triggered by the voice assistant), and the information of the song may be output to the user; for example, the song library for listening to songs can be added as the first song in the song library for listening to songs and identifying songs, while the user is informed of the negative screen display or the inquiry of the user.
Alternatively, the songs in the song library for listening to the songs and recognizing the songs may all be the information of the songs obtained by the voice assistant for listening to the songs and recognizing the songs, wherein the songs may be arranged in the reverse order of the recognized time, the information of the songs recognized earlier is arranged later, and the information of the songs recognized latest is arranged first in the song library for listening to the songs and recognizing the songs. It should be noted that this arrangement is only an example, and the embodiments of the present application do not limit the order of the songs in the song library.
Alternatively, a fixed number of songs may be limited in the song library for listening to songs, and when the number of songs in the song library reaches the fixed number, if newly identified songs are to be added, the oldest identified song may be deleted, and then information of the newly identified song may be added.
Optionally, the target song information includes, but is not limited to, a song name, and at least one of:
a singer;
a person editing the koji;
making a word;
a copyright holder;
the original singer;
the album to which it belongs;
the release time.
Optionally, whether an input inquiring the target song information is received or not, a target screen of the electronic device of the user, such as minus one screen, may generate a recommended song library card for listening to songs, wherein a song library for listening to songs may be displayed, and may receive an input of a user selecting collection, approval, playing, downloading, commenting, stepping on, etc. on the card and perform a corresponding operation.
Optionally, no matter whether an input for inquiring the target song information is received or not, negative one screen of the electronic device may generate a recommended song library card for listening to songs and identifying songs, a maximum of 5 cards may be displayed at a time, each card displays song information for listening to songs and identifying songs, the cards may be arranged according to time sequence, the latest identified song is arranged at the first position, and other cards may be displayed in a sliding manner.
Optionally, the displaying the target song information includes:
receiving a second input of the user, wherein the second input is an operation of inquiring the target song information by the user;
in response to the second input, displaying the target song information on a current display interface or a display interface of a voice assistant.
Alternatively, if the user actively asks what song was just heard through the second input, the song-listening recognition result can be presented to the user; for example, the intelligent voice assistant may display the song library for listening to songs or the recently identified target song information on the intelligent voice interface, or display the song library for listening to songs or the recently identified target song information on the current display interface of the user in the form of a pop-up window, a text box, a card, or the like.
Optionally, it is also possible to receive input of selecting collection, approval, playing, downloading, comment, and the like of the displayed song by the user, and perform corresponding operations based on the input.
Alternatively, the second input may be a voice input by the user to a voice assistant; may be a special gesture input; or may be a special click or slide input to the screen or the body, which is not limited in this embodiment.
Optionally, the first input may be an operation of the user actively singing a certain sentence of lyrics to search for a song, or may be an input received by the background to receive music in the surrounding environment;
optionally, the sound characteristics of the owner of the electronic device may be stored;
optionally, in a case that the first input is an operation of searching for a song by the user actively singing a certain sentence of lyrics, the obtained voice feature corresponding to the singing voice may be matched with the voice feature of the owner, and if it is determined that the first input corresponds to the owner, the search result may be directly displayed by the voice assistant; if the matching determines that the first input corresponds to the owner, the search result may be directly displayed without directly displaying the search result, but in the case where the owner of the receiver inputs the search result by the voice assistant.
Optionally, the displaying the target song information includes:
displaying a song identification song library;
wherein the song identification song library comprises target song information of the first audio.
Optionally, if the first text obtained by the voice assistant based on the first audio, that is, the lyrics of the song played in the environment where the user is located, completely matches the lyrics recorded in a certain song in the reference song library, the voice assistant may be considered to capture the lyrics of the climax part, the intelligent song listening recognition function is triggered (triggered by the voice assistant), and the information of the song may be output to the user; for example, the song library for listening to songs can be added as the first song in the song library for listening to songs and showing the song library for listening to songs (i.e. song identification library) while displaying the song library for listening to songs or informing the user when the user inquires.
Alternatively, the songs in the song recognition song library may all be the information of the songs obtained by the voice assistant listening to the songs, wherein the songs may be arranged in the reverse order of the recognition time, the earlier the information of the recognized songs is arranged later, and the information of the latest recognized songs may be arranged first in the song library listening to the songs. It should be noted that this arrangement is only an example, and the embodiments of the present application do not limit the arrangement order of songs in the song recognition library.
Alternatively, a fixed number of songs may be defined in the song recognition song library, and when the number of songs in the song library reaches the fixed number, if newly recognized songs are to be added, the oldest recognized song may be deleted, and then information of the newly recognized song may be added.
Alternatively, if the user actively asks what song was just heard through the second input, the song-listening recognition result can be presented to the user; for example, the intelligent voice assistant may display the song recognition song library through displaying the song recognition song library on the intelligent voice interface, or through a pop-up window or a text box or a card on the current display interface of the user.
Optionally, it is also possible to receive input of selecting collection, approval, playing, downloading, comment, and the like of the displayed song by the user, and perform corresponding operations based on the input.
Alternatively, the maximum level of the song library for listening to songs may be 100, and the historical songs may be eliminated in a first-in first-out chronological order.
Optionally, the songs actively asked by the user can be eliminated from the song library for listening to songs and identifying songs.
Optionally, the method further comprises:
determining the reference library;
wherein the reference library of songs comprises:
a popular song library and/or a recommended song library.
Alternatively, the target song may be determined from the reference song library at the time of song identification, and thus the reference song library may be determined first.
Alternatively, the reference song library may include a popular song library and/or a recommended song library.
Optionally, the songs in the reference song library can be updated regularly, so that the timeliness of the reference song library is guaranteed; it is ensured that the songs of the environment in which the user is located can be identified as accurately as possible.
Optionally, the reference library of songs comprises:
a popular song library;
the popular song library comprises at least one popular song;
wherein the determining the reference library of songs comprises:
determining the at least one popular song from big to small in sequence based on the popular score of each first song to be selected;
the popular score of the first song to be selected is determined based on the historical playing amount of the song to be selected and the playing amplification of the song to be selected after the song to be selected is released, and the number of the at least one popular song is preset.
Optionally, a hit score of each first song to be selected may be calculated, and the at least one hit song may be determined sequentially from large to small.
Alternatively, the first song to be selected may be all songs in the music software or all songs having unique identifiers or preset songs.
For example, the top score of each first song to be selected may be calculated, and the top 100 songs may be obtained to form a top song library according to the reverse order of the scores.
Optionally, when the hit score of each first song to be selected is calculated, the hit score may be determined based on the historical play amount of the song to be selected and the play amplification of the song to be selected after being released.
Optionally, the song score is 0.5 the historical playing amount score +0.5 the playing amplification score;
optionally, the history play score is 0.99 of the history play amountTime of release(ii) a Alternatively, the release time may become a release age, such as a song released for 150 days, which is (150/365) years; a 10 year full song is released with a release age of 10 years.
Wherein, the parameters can be as shown in the following table 1 (parameter table 1);
table 1 parameters table 1
play_time Historical playback volume, eg 1000
pushlish_time Release time, eg:20
play_times_scroe 1000*0.9920
Optionally, the calculation formula of the song score and the historical playing score performs time attenuation on the historical playing amount, so that the old song score is prevented from being too high;
optionally, the play augmentation score is 0.5 absolute augmentation +0.5 relative augmentation +0.5 absincrase 2; wherein, the parameters can be as shown in the following table 2 (parameter table 2);
table 2 parameters table 2
Figure BDA0003002506360000091
Optionally, the reference library of songs comprises:
recommending song libraries;
the recommended song library comprises a first recommended song;
wherein the determining the reference library of songs comprises:
for each user song in the user song list, determining a first similar song list of each user song based on the playing amount of the user song, the playing amount of a second song to be selected and the common playing amount of the user song and the second song to be selected;
and determining the songs in the first similar song list as first recommended songs.
Alternatively, in determining the recommended song library, songs similar to the user's taste may be determined.
Optionally, for user songs in an existing song list of the user, a candidate similar song list can be established for each user song; for example, for the first N songs of the user song a in the existing song list of the user, that is, all songs whose similarity with the user song a exceeds the preset threshold, N may be 3, 10, or 100, that is, N is a positive integer, which is not limited in this embodiment.
Optionally, the top N of all songs for which the similarity to song a of the user exceeds a preset threshold may be determined from the second candidate songs;
alternatively, the second candidate song may be all songs in the music software or so uniquely identified songs or pre-set songs.
Optionally, after a candidate similar list is created for each user song, the songs in all the candidate similar lists may be removed from duplication, and finally a first similar list is formed, where the songs in the first similar list are the first recommended songs.
Optionally, the similarity calculation method is as follows:
Figure BDA0003002506360000101
wherein the parameters are shown in Table 3 below (parameter Table 3); the playing quantity of the songs a is the number of users playing the songs a, and the playing quantity of the songs b is the number of users playing the songs b; the amount of song ab common playback is the number of users who have both a and b played.
Table 3 parameters table 3
Figure BDA0003002506360000102
Optionally, the library of recommended songs includes a second recommended song;
wherein the determining the reference library of songs comprises:
determining repeated songs in all the song lists of at least one similar user as second recommended songs; wherein, the overlapping degree of the songs in the singing list of the same kind of users and the songs of the users in the singing list of the users is higher than a first threshold value.
Optionally, songs that most users like may also be included when determining the recommended song library.
Alternatively, the song lists of a plurality of users including the user may be clustered, wherein the intersection of the song lists is greater than a first threshold value, such as 90%, and then the user may be considered as a class of users, and then the intersection of the song lists of at least one other user of the same class as the user or all users of the same class as the user may be added to the recommended song library of the user.
Alternatively, the library of recommended songs may filter songs from the user's existing song list, as well as songs already in the reference library, and may also filter songs that have received input from the user for a specific step.
Optionally, the method further comprises at least one of:
under the condition that the number of times of receiving the first input within the preset time length exceeds a second threshold value, not responding to the first input;
in the event that a preset environment receives a first input, the first input is not responded to.
Optionally, the current position of the user may be obtained, if the user is at home, the user may play songs with other playing devices such as a computer, and if frequently triggered, the user may not need to listen to songs and recognize songs in such a scenario, and may set intelligent songs listening and recognizing without triggering, so as to avoid energy consumption of the electronic device and do no work.
Optionally, the number of times of receiving the first input within the preset time length exceeds a second threshold; song recognition is not performed in response to the first input.
Alternatively, a first input (i.e., the electronic device recognizes that there is a song playing around) may be received at a preset environment such as a home or a company, and song recognition may not be performed in response to the first input.
Optionally, song recognition may not be performed in response to the first input when any one of a condition that the number of times the first input is received within a preset time period exceeds a second threshold or a condition that the number of times the first input is received within the preset time period exceeds a second threshold is satisfied.
Optionally, the song recognition may not be performed in response to the first input when any one of the conditions that the number of times of receiving the first input within the preset time period exceeds the second threshold and the number of times of receiving the first input within the preset time period exceeds the second threshold is satisfied.
In the embodiment of the application, the audio in the current environment is received in the screen-off state or the use state of the electronic equipment, the audio is identified by the voice assistant, the song information played in the current environment is determined and displayed, the intelligent song listening and identification of the voice assistant are realized, the song played in the environment where the electronic equipment is located is timely and automatically identified, and the user is prevented from missing the song.
It should be noted that, in the song recognition method provided in the embodiment of the present application, the execution subject may be a song recognition apparatus, or alternatively, a control module in the song recognition apparatus for executing the loading song recognition method. In the embodiment of the present application, a song recognition apparatus executes a method of recognizing a loaded song as an example, and the song recognition apparatus provided in the embodiment of the present application is described.
Fig. 2 is a schematic structural diagram of a song recognition apparatus provided in an embodiment of the present application, and as shown in fig. 2, the apparatus includes: a first receiving module 210, and a first outputting module 220; wherein:
the first receiving module 210 is configured to receive a first input when the electronic device is in a target state, where content of the first input is first audio;
the first output module 220 is used for outputting target song information of the first audio based on first text corresponding to the first audio determined by a voice assistant of the electronic equipment in response to the first input;
the target state includes at least one of a screen-saving state and a use state.
Specifically, the song recognition apparatus receives a first input through the first receiving module 210 when the electronic device is in the target state, and the content of the first input is first audio; then responding to the first input through the first output module 220, and outputting target song information of the first audio based on first text corresponding to the first audio determined by a voice assistant of the electronic equipment; the target state includes at least one of a screen-saving state and a use state.
In the embodiment of the application, the audio in the current environment is received in the screen-off state or the use state of the electronic equipment, the audio is identified by the voice assistant, the song information played in the current environment is determined and displayed, the intelligent song listening and identification of the voice assistant are realized, the song played in the environment where the electronic equipment is located is timely and automatically identified, and the user is prevented from missing the song.
Optionally, the first output module is configured to:
taking a first song to which first lyrics matched with the first text belong as a target song, wherein the first song is one of songs in a reference song library;
and displaying the target song information.
Optionally, the first output module is configured to:
receiving a second input of the user, wherein the second input is an operation of inquiring the target song information by the user;
in response to the second input, displaying the target song information on a current display interface or a display interface of a voice assistant.
Optionally, the first output module is configured to:
displaying a song identification song library;
wherein the song identification song library comprises target song information of the first audio.
Optionally, the reference library of songs comprises:
a popular song library;
the popular song library comprises at least one popular song;
wherein the first determination module is to:
determining the at least one popular song from big to small in sequence based on the popular score of each first song to be selected;
the popular score of the first song to be selected is determined based on the historical playing amount of the song to be selected and the playing amplification of the song to be selected after the song to be selected is released, and the number of the at least one popular song is preset.
Optionally, the reference library of songs comprises:
recommending song libraries;
the recommended song library comprises a first recommended song;
wherein the first determination module is to:
for each user song in the user song list, determining a first similar song list of each user song based on the playing amount of the user song, the playing amount of a second song to be selected and the common playing amount of the user song and the second song to be selected;
and determining the songs in the first similar song list as first recommended songs.
Optionally, the library of recommended songs includes a second recommended song;
wherein the first determination module is to:
determining repeated songs in all the song lists of at least one similar user as second recommended songs; wherein, the overlapping degree of the songs in the singing list of the same kind of users and the songs of the users in the singing list of the users is higher than a first threshold value.
Optionally, the apparatus further comprises:
the second receiving module is used for not responding to the first input under the condition that the number of times of receiving the first input in the preset time length exceeds a second threshold value;
and the third receiving module is used for not responding to the first input under the condition that the first input is received in the preset environment.
The song recognition apparatus in the embodiment of the present application may be an apparatus or an electronic device having an operating system, or may be a component, an integrated circuit, or a chip in a terminal. The electronic device may be a mobile electronic device or a non-mobile electronic device. By way of example, the mobile electronic device may be an electronic device, a tablet computer, a notebook computer, a palm top computer, a vehicle-mounted electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook or a Personal Digital Assistant (PDA), and the like, and the non-mobile electronic device may be a server, a Network Attached Storage (NAS), a Personal Computer (PC), a Television (TV), a teller machine, a self-service machine, and the like, and the embodiments of the present application are not particularly limited.
The song recognition device provided in the embodiment of the present application can implement each process implemented by the song recognition device in the embodiment of the method in fig. 1, and is not described here again to avoid repetition.
In the embodiment of the application, the audio in the current environment is received in the screen-off state or the use state of the electronic equipment, the audio is identified by the voice assistant, the song information played in the current environment is determined and displayed, the intelligent song listening and identification of the voice assistant are realized, the song played in the environment where the electronic equipment is located is timely and automatically identified, and the user is prevented from missing the song.
Optionally, an electronic device is further provided in the embodiment of the present application, and fig. 3 is one of the hardware structure diagrams of the electronic device provided in the embodiment of the present application. As shown in fig. 3, the electronic device includes a processor 302, a memory 301, and a program or an instruction stored in the memory 301 and executable on the processor 302, where the program or the instruction when executed by the processor 302 implements the processes of the above-mentioned song recognition method embodiment, and can achieve the same technical effects, and no further description is provided herein for avoiding repetition.
It should be noted that the electronic devices in the embodiments of the present application include the mobile electronic devices and the non-mobile electronic devices described above.
Fig. 4 is a second schematic diagram of a hardware structure of an electronic device according to an embodiment of the present disclosure.
The electronic device 400 includes, but is not limited to: at least some of the components of the radio unit 401, the network module 402, the audio output unit 403, the input unit 404, the sensor 405, the display unit 406, the user input unit 407, the interface unit 408, the memory 409, and the processor 410, and the like.
Those skilled in the art will appreciate that the electronic device 400 may further include a power source (e.g., a battery) for supplying power to various components, and the power source may be logically connected to the processor 410 through a power management system, so as to implement functions of managing charging, discharging, and power consumption through the power management system. The electronic device structure shown in fig. 4 does not constitute a limitation of the electronic device, and the electronic device may include more or less components than those shown, or combine some components, or arrange different components, and thus, the description is omitted here.
It should be understood that, in the embodiment of the present application, the input Unit 404 may include a Graphics Processing Unit (GPU) 441 and a microphone 442, and the Graphics processor 441 processes image data of still pictures or videos obtained by an image capturing device (such as a camera) in a video capturing mode or an image capturing mode. The display unit 406 may include a display panel 461, and the display panel 461 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like. The user input unit 407 includes a touch panel 471 and other input devices 472. A touch panel 471, also referred to as a touch screen. The touch panel 471 can include two parts, a touch detection device and a touch controller. Other input devices 4072 may include, but are not limited to, a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, a mouse, and a joystick, which are not described in detail herein.
In the embodiment of the present application, the radio frequency unit 401 acquires information and then processes the information with the processor 410. Typically, radio unit 401 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like.
The memory 409 may be used to store software programs or instructions as well as various data. The memory 409 may mainly include a storage program or instruction area and a storage data area, wherein the storage program or instruction area may store an operating system, an application program or instruction (such as a sound playing function, an image playing function, etc.) required for at least one function, and the like. In addition, the Memory 409 may include a high-speed random access Memory, and may further include a nonvolatile Memory, wherein the nonvolatile Memory may be a Read-Only Memory (ROM), a Programmable Read-Only Memory (PROM), an Erasable Programmable Read-Only Memory (EPROM), an Electrically Erasable Programmable Read-Only Memory (EEPROM), or a flash Memory. Such as at least one magnetic disk storage device, flash memory device, or other non-volatile solid state storage device.
Processor 410 may include one or more processing units; alternatively, the processor 410 may integrate an application processor, which primarily handles operating systems, user interfaces, and applications or instructions, etc., and a modem processor, which primarily handles wireless communications, such as a baseband processor. It will be appreciated that the modem processor described above may not be integrated into the processor 410.
The input unit 404 is configured to receive a first input when the electronic device is in the target state, where content of the first input is first audio;
the processor 410 is configured to output target song information of the first audio based on a first text corresponding to the first audio determined by a voice assistant of the electronic device in response to the first input;
the target state includes at least one of a screen-saving state and a use state.
In the embodiment of the application, the audio in the current environment is received in the screen-off state or the use state of the electronic equipment, the audio is identified by the voice assistant, the song information played in the current environment is determined and displayed, the intelligent song listening and identification of the voice assistant are realized, the song played in the environment where the electronic equipment is located is timely and automatically identified, and the user is prevented from missing the song.
Optionally, the processor 410 is configured to:
taking a first song to which first lyrics matched with the first text belong as a target song, wherein the first song is one of songs in a reference song library;
and displaying the target song information.
Optionally, the processor 410 is configured to:
receiving a second input of the user, wherein the second input is an operation of inquiring the target song information by the user;
in response to the second input, displaying the target song information on a current display interface or a display interface of a voice assistant.
Optionally, the processor 410 is configured to:
displaying a song identification song library;
wherein the song identification song library comprises target song information of the first audio.
Optionally, the reference library of songs comprises:
a popular song library;
the popular song library comprises at least one popular song;
wherein the processor 410 is configured to:
determining the at least one popular song from big to small in sequence based on the popular score of each first song to be selected;
the popular score of the first song to be selected is determined based on the historical playing amount of the song to be selected and the playing amplification of the song to be selected after the song to be selected is released, and the number of the at least one popular song is preset.
Optionally, the reference library of songs comprises:
recommending song libraries;
the recommended song library comprises a first recommended song;
wherein the processor 410 is configured to:
for each user song in the user song list, determining a first similar song list of each user song based on the playing amount of the user song, the playing amount of a second song to be selected and the common playing amount of the user song and the second song to be selected;
and determining the songs in the first similar song list as first recommended songs.
Optionally, the library of recommended songs includes a second recommended song;
wherein the processor 410 is configured to:
determining repeated songs in all the song lists of at least one similar user as second recommended songs; wherein, the overlapping degree of the songs in the singing list of the same kind of users and the songs of the users in the singing list of the users is higher than a first threshold value.
Optionally, the processor 410 is configured to:
under the condition that the number of times of receiving the first input within the preset time length exceeds a second threshold value, not responding to the first input;
in the event that a preset environment receives a first input, the first input is not responded to.
In the embodiment of the application, the audio in the current environment is received in the screen-off state or the use state of the electronic equipment, the audio is identified by the voice assistant, the song information played in the current environment is determined and displayed, the intelligent song listening and identification of the voice assistant are realized, the song played in the environment where the electronic equipment is located is timely and automatically identified, and the user is prevented from missing the song.
The electronic device embodiment in the embodiment of the present application is a product embodiment corresponding to the method embodiment, and all implementation manners in the method embodiment are applicable to the electronic device embodiment, and may also achieve the same or similar technical effects, so that details are not described herein again.
The embodiment of the present application further provides a readable storage medium, where a program or an instruction is stored on the readable storage medium, and when the program or the instruction is executed by a processor, the program or the instruction implements each process of the foregoing song identification method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
The processor is the processor in the electronic device described in the above embodiment. The readable storage medium includes a computer readable storage medium, such as a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and so on.
The embodiment of the present application further provides a chip, where the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is configured to run a program or an instruction to implement each process of the above song identification method embodiment, and can achieve the same technical effect, and is not described here again to avoid repetition.
It should be understood that the chips mentioned in the embodiments of the present application may also be referred to as system-on-chip, system-on-chip or system-on-chip, etc.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other elements in a process, method, article, or apparatus that comprises the element. Further, it should be noted that the scope of the methods and apparatus of the embodiments of the present application is not limited to performing the functions in the order illustrated or discussed, but may include performing the functions in a substantially simultaneous manner or in an opposing order depending on the functionality involved, e.g., the methods described may be performed in an order different than that described, and various steps may be added, omitted, or combined. In addition, features described with reference to certain examples may be combined in other examples.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solutions of the present application may be embodied in the form of a software product, which is stored in a storage medium (e.g., ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal (e.g., an electronic device, a computer, a server, or a network device) to execute the method according to the embodiments of the present application.
While the present embodiments have been described with reference to the accompanying drawings, it is to be understood that the invention is not limited to the precise embodiments described above, which are meant to be illustrative and not restrictive, and that various changes may be made therein by those skilled in the art without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (15)

1. A song identification method, comprising:
receiving a first input under the condition that the electronic equipment is in a target state, wherein the content of the first input is first audio;
in response to the first input, outputting target song information of the first audio based on first text corresponding to the first audio determined by a voice assistant of the electronic equipment;
the target state includes at least one of a screen-saving state and a use state.
2. The song recognition method of claim 1, wherein outputting the target song information of the first audio based on the first text corresponding to the first audio comprises:
taking a first song to which first lyrics matched with the first text belong as a target song, wherein the first song is one of songs in a reference song library;
and displaying the target song information.
3. The song recognition method of claim 2, wherein the displaying the target song information comprises:
receiving a second input of the user, wherein the second input is an operation of inquiring the target song information by the user;
in response to the second input, displaying the target song information on a current display interface or a display interface of a voice assistant.
4. The song recognition method of claim 2 or 3, wherein the displaying the target song information comprises:
displaying a song identification song library;
wherein the song identification song library comprises target song information of the first audio.
5. The song recognition method of claim 4, wherein the reference song library comprises:
a popular song library;
the popular song library comprises at least one popular song;
wherein the determining the reference library of songs comprises:
determining the at least one popular song from big to small in sequence based on the popular score of each first song to be selected;
the popular score of the first song to be selected is determined based on the historical playing amount of the song to be selected and the playing amplification of the song to be selected after the song to be selected is released, and the number of the at least one popular song is preset.
6. The song recognition method of claim 4, wherein the reference song library comprises:
recommending song libraries;
the recommended song library comprises a first recommended song;
wherein the determining the reference library of songs comprises:
for each user song in the user song list, determining a first similar song list of each user song based on the playing amount of the user song, the playing amount of a second song to be selected and the common playing amount of the user song and the second song to be selected;
and determining the songs in the first similar song list as first recommended songs.
7. The song identification method of claim 5 or 6, wherein the library of recommended songs includes a second recommended song;
wherein the determining the reference library of songs comprises:
determining repeated songs in all the song lists of at least one similar user as second recommended songs; wherein, the overlapping degree of the songs in the singing list of the same kind of users and the songs of the users in the singing list of the users is higher than a first threshold value.
8. The song identification method of claim 1, further comprising at least one of:
under the condition that the number of times of receiving the first input within the preset time length exceeds a second threshold value, not responding to the first input;
in the event that a preset environment receives a first input, the first input is not responded to.
9. A song recognition apparatus, comprising:
the electronic equipment comprises a first receiving module, a second receiving module and a display module, wherein the first receiving module is used for receiving a first input under the condition that the electronic equipment is in a target state, and the content of the first input is first audio;
a first output module, configured to output, in response to the first input, target song information of the first audio based on a first text corresponding to the first audio determined by a voice assistant of the electronic device;
the target state includes at least one of a screen-saving state and a use state.
10. The song recognition apparatus of claim 9, wherein the first output module is to:
taking a first song to which first lyrics matched with the first text belong as a target song, wherein the first song is one of songs in a reference song library;
and displaying the target song information.
11. The song recognition apparatus of claim 10, wherein the first output module is to:
receiving a second input of the user, wherein the second input is an operation of inquiring the target song information by the user;
in response to the second input, displaying the target song information on a current display interface or a display interface of a voice assistant.
12. The song recognition apparatus of claim 10 or 11, wherein the first output module is configured to:
displaying a song identification song library;
wherein the song identification song library comprises target song information of the first audio.
13. The song recognition apparatus of claim 9, wherein the apparatus further comprises:
the second receiving module is used for not responding to the first input under the condition that the number of times of receiving the first input in the preset time length exceeds a second threshold value;
and the third receiving module is used for not responding to the first input under the condition that the first input is received in the preset environment.
14. An electronic device comprising a processor, a memory, and a program or instructions stored on the memory and executable on the processor, the program or instructions when executed by the processor implementing the steps of the song recognition method of any one of claims 1-8.
15. A readable storage medium, on which a program or instructions are stored, which program or instructions, when executed by a processor, carry out the steps of the song identification method according to any one of claims 1 to 8.
CN202110351517.5A 2021-03-31 2021-03-31 Song identification method and device, electronic equipment and storage medium Pending CN113076444A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110351517.5A CN113076444A (en) 2021-03-31 2021-03-31 Song identification method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110351517.5A CN113076444A (en) 2021-03-31 2021-03-31 Song identification method and device, electronic equipment and storage medium

Publications (1)

Publication Number Publication Date
CN113076444A true CN113076444A (en) 2021-07-06

Family

ID=76614274

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110351517.5A Pending CN113076444A (en) 2021-03-31 2021-03-31 Song identification method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN113076444A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115602154A (en) * 2022-12-15 2023-01-13 杭州网易云音乐科技有限公司(Cn) Audio recognition method and device, storage medium and computing equipment

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140172423A1 (en) * 2012-12-14 2014-06-19 Lenovo (Beijing) Co., Ltd. Speech recognition method, device and electronic apparatus
CN112015943A (en) * 2019-05-31 2020-12-01 华为技术有限公司 Humming recognition method and related equipment

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140172423A1 (en) * 2012-12-14 2014-06-19 Lenovo (Beijing) Co., Ltd. Speech recognition method, device and electronic apparatus
CN112015943A (en) * 2019-05-31 2020-12-01 华为技术有限公司 Humming recognition method and related equipment

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115602154A (en) * 2022-12-15 2023-01-13 杭州网易云音乐科技有限公司(Cn) Audio recognition method and device, storage medium and computing equipment
CN115602154B (en) * 2022-12-15 2023-08-11 杭州网易云音乐科技有限公司 Audio identification method, device, storage medium and computing equipment

Similar Documents

Publication Publication Date Title
CN109522419B (en) Session information completion method and device
CN106098063B (en) Voice control method, terminal device and server
CN105488112A (en) Information pushing method and device
WO2016165325A1 (en) Audio information recognition method and apparatus
CN103605656A (en) Music recommendation method and device and mobile terminal
CN104378441A (en) Schedule creating method and device
CN110147467A (en) A kind of generation method, device, mobile terminal and the storage medium of text description
CN104035995A (en) Method and device for generating group tags
CN110391966B (en) Message processing method and device and message processing device
CN109165292A (en) Data processing method, device and mobile terminal
CN105068976A (en) Ticket information exhibition method and device
CN111491123A (en) Video background processing method and device and electronic equipment
CN112287162A (en) Message searching method and device and electronic equipment
CN104615663A (en) File sorting method and device and terminal
WO2021135561A1 (en) Skill voice wake-up method and apparatus
CN111708943A (en) Search result display method and device and search result display device
CN103366010A (en) Method and device for searching audio file
CN108234758B (en) Application display method and device, storage medium and electronic equipment
CN113076444A (en) Song identification method and device, electronic equipment and storage medium
CN111739535A (en) Voice recognition method and device and electronic equipment
CN106407359A (en) Image playing method and mobile terminal
CN112286421A (en) Playlist processing method and device and electronic equipment
CN105843865B (en) Media file playing method and device
CN107729439A (en) Obtain the methods, devices and systems of multi-medium data
CN112712825A (en) Audio processing method and device and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination