WO2016112841A1 - 一种信息处理方法及客户端、计算机存储介质 - Google Patents

一种信息处理方法及客户端、计算机存储介质 Download PDF

Info

Publication number
WO2016112841A1
WO2016112841A1 PCT/CN2016/070683 CN2016070683W WO2016112841A1 WO 2016112841 A1 WO2016112841 A1 WO 2016112841A1 CN 2016070683 W CN2016070683 W CN 2016070683W WO 2016112841 A1 WO2016112841 A1 WO 2016112841A1
Authority
WO
WIPO (PCT)
Prior art keywords
information
text
intercept
client
audio segment
Prior art date
Application number
PCT/CN2016/070683
Other languages
English (en)
French (fr)
Inventor
肖芬
Original Assignee
腾讯科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 腾讯科技(深圳)有限公司 filed Critical 腾讯科技(深圳)有限公司
Priority to MYPI2016702118A priority Critical patent/MY187298A/en
Priority to US15/187,056 priority patent/US9858910B2/en
Publication of WO2016112841A1 publication Critical patent/WO2016112841A1/zh
Priority to US15/824,495 priority patent/US10580394B2/en

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/36Accompaniment arrangements
    • G10H1/361Recording/reproducing of accompaniment for use with an external source, e.g. karaoke systems
    • G10H1/365Recording/reproducing of accompaniment for use with an external source, e.g. karaoke systems the accompaniment information being stored on a host computer and transmitted to a reproducing terminal by means of a network, e.g. public telephone lines
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/061Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for extraction of musical phrases, isolation of musically relevant segments, e.g. musical thumbnail generation, or for temporal structure analysis of a musical piece, e.g. determination of the movement sequence of a musical work
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/091Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for performance evaluation, i.e. judging, grading or scoring the musical qualities or faithfulness of a performance, e.g. with respect to pitch, tempo or other timings of a reference performance
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2220/00Input/output interfacing specifically adapted for electrophonic musical tools or instruments
    • G10H2220/005Non-interactive screen display of musical or status data
    • G10H2220/011Lyrics displays, e.g. for karaoke applications
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2220/00Input/output interfacing specifically adapted for electrophonic musical tools or instruments
    • G10H2220/091Graphical user interface [GUI] specifically adapted for electrophonic musical instruments, e.g. interactive musical displays, musical instrument icons or menus; Details of user interactions therewith
    • G10H2220/101Graphical user interface [GUI] specifically adapted for electrophonic musical instruments, e.g. interactive musical displays, musical instrument icons or menus; Details of user interactions therewith for graphical creation, edition or control of musical data or parameters
    • G10H2220/106Graphical user interface [GUI] specifically adapted for electrophonic musical instruments, e.g. interactive musical displays, musical instrument icons or menus; Details of user interactions therewith for graphical creation, edition or control of musical data or parameters using icons, e.g. selecting, moving or linking icons, on-screen symbols, screen regions or segments representing musical elements or parameters

Definitions

  • the present invention relates to communication technologies, and in particular, to an information processing method, a client, and a computer storage medium.
  • An application scenario of information processing may be a scenario for a commonly used K song APP on a smart terminal.
  • Such an APP installed on the smart terminal may play an audio file downloaded from the server to the user, such as an accompaniment music corresponding to the song.
  • the user's voice information is collected and identified, and the recognition result or the comparison result obtained by further comparing the recognition result with the preset standard reference information is uploaded to the server for sharing and listening to others.
  • the prior art cannot intercept and play part of the content of the audio file, and the larger the amount of information collected, the more complicated the corresponding recognition effect, and the longer the time taken for the recognition;
  • the user's own preferences and sound conditions, such as the vocal tone is relatively high, or relatively low, resulting in a mismatch with the preset standard reference information, thus not getting the best sound performance results, that is, the user may It is also only interested in part of the content of the audio file, and only needs to intercept part of the content of the audio file, but it is impossible to intercept and play part of the content of the audio file by using the prior art.
  • the embodiments of the present invention are intended to provide an information processing method, a client, and a computer storage medium, which at least solve the problems existing in the prior art.
  • the present invention provides an information processing method, which is applied to a terminal, and the method includes:
  • the present invention further provides a client, where the client is applied to a terminal, and the client includes:
  • a downloading unit configured to download the first audio file and the first text file that matches the first audio file
  • the intercepting unit is configured to perform partial truncation of the first audio file to the first audio segment according to the first indication information for identifying the interception start position and the second indication information for identifying the intercept end position;
  • a playing unit configured to play the first audio segment, and synchronously dynamically display a portion of the first text file corresponding to the first audio segment
  • the collecting unit is configured to collect the voice information of the user while playing the first audio segment
  • a synthesizing unit configured to synthesize the first audio segment and the voice information into a first acquisition result.
  • the downloading unit, the intercepting unit, the playing unit, the collecting unit, and the synthesizing unit may adopt a central processing unit (CPU), a digital signal processor (DSP, Digital Singnal Processor), or Field-Programmable Gate Array (FPGA) implementation.
  • CPU central processing unit
  • DSP digital signal processor
  • FPGA Field-Programmable Gate Array
  • Embodiments of the present invention also provide a computer storage medium in which computer executable instructions are stored for executing the above information processing method.
  • An information processing method is applied to a terminal, where the method includes: triggering a first operation; and in response to the first operation, downloading a first audio file and comparing with the first audio file Matching the first text file; performing partial truncation on the first audio file to the first audio segment according to the first indication information for identifying the interception start position and the second indication information for identifying the intercept end position; triggering the second operation Responding to the second operation, playing the first audio segment, and synchronously displaying the text information portion corresponding to the first audio segment in the first text file; while playing the first audio segment, Collecting voice information of the user; synthesizing the first audio segment and the voice information into a first collection result.
  • the information processing method and the client of the embodiment of the present invention can intercept and play part of the downloaded audio file, on the one hand, reduce the amount of collected information, and reduce the processing load of the terminal; on the other hand, the user can
  • the preferences and sound conditions capture the audio content that you are interested in or that you are capable of, thus enhancing the user experience.
  • Embodiment 1 is a schematic flowchart of an implementation process of Embodiment 1 of the present invention.
  • Embodiment 2 is a schematic flowchart of an implementation process of Embodiment 2 of the method according to the present invention
  • Embodiment 3 is a schematic diagram of an application scenario of Embodiment 2 of a method according to the present invention.
  • FIG. 5 is a schematic diagram of an application scenario of Embodiment 3 of the method according to the present invention.
  • FIG. 6 is a schematic flowchart of an implementation process of Embodiment 4 of the method of the present invention.
  • FIG. 7 is a schematic diagram of an application scenario of applying the fourth embodiment of the method of the present invention.
  • FIG. 8 is a schematic structural diagram of a system for implementing an embodiment of the present invention.
  • FIG. 9 is a schematic structural diagram of a terminal that implements an embodiment of the present invention.
  • FIG. 10 is an interaction flowchart of implementing an embodiment of the present invention.
  • FIG. 11 is a schematic structural diagram of a first embodiment of a client according to the present invention.
  • FIG. 12 is a schematic structural diagram of a component of a second embodiment of a client according to the present invention.
  • An information processing method is applied to a terminal in the embodiment of the present invention. As shown in FIG. 1 , the method includes:
  • Step 101 trigger a first operation
  • Step 102 In response to the first operation, downloading the first audio file and the first text file that matches the first audio file;
  • Step 103 Perform partial truncation on the first audio file to the first audio segment according to the first indication information for identifying the location of the interception starting point and the second indication information for identifying the location of the intercepted end point;
  • Step 104 trigger a second operation
  • Step 105 In response to the second operation, playing the first audio segment, and synchronously displaying the text information portion corresponding to the first audio segment in the first text file;
  • Step 106 Collecting voice information of the user while playing the first audio segment
  • Step 107 Synthesize the first audio segment and the voice information into a first collection result.
  • step 102 may be completed before the intercepting operation of step 103, or only the information necessary for the intercepting operation in step 103 may be downloaded, and when the first audio segment is played later in step 105, Use streaming media to play while downloading the audio data you need later.
  • the technical problem that the partial content of the audio file cannot be intercepted and played in the prior art is solved.
  • intercepting part of the audio file for playback and collecting corresponding voice information will greatly reduce the amount of information collected and reduce the processing load of the terminal;
  • users can intercept the audio content that they are interested in or their own ability according to their own preferences and sound conditions, thus improving the user experience.
  • An information processing method of the embodiment of the present invention is applied to a terminal. As shown in FIG. 2, the method includes:
  • Step 201 trigger a first operation
  • Step 202 In response to the first operation, downloading the first audio file and the first text file that matches the first audio file;
  • Step 203 Provide a first interaction object and a second interaction object that support a sliding operation.
  • Step 204 Capture the interception start position and the intercept end position simultaneously or separately by using the first interaction object and the second interaction object;
  • Step 205 Generate first indication information for identifying a location of the interception start point and second indication information for identifying a location of the interception end point;
  • Step 206 intercepting, from the first audio file, a portion between a time point indicated by the first indication information and a time point indicated by the second indication information, to obtain a first audio segment;
  • Step 207 triggering a second operation
  • Step 208 In response to the second operation, playing the first audio segment, and synchronously displaying the text information portion corresponding to the first audio segment in the first text file;
  • Step 209 Collecting voice information of the user while playing the first audio segment
  • Step 210 Synthesize the first audio segment and the voice information into a first collection result.
  • the embodiment provides a specific implementation method for intercepting and playing part of the content of the audio file, wherein the user is provided with two slidable interactive objects.
  • the user can manipulate the two interactive objects by a swipe gesture of the finger on the touch screen to determine the cut start position and the cut end position.
  • the user can separately manipulate two interactive objects by using a single finger according to needs or according to their own operating habits, for example, using a single finger touch operation to manipulate one of the two interactive objects to determine the intercept starting position, at the intercept starting position.
  • a single finger touch operation can also be used to manipulate one of the two interactive objects to determine the interception end point.
  • Position, after determining the interception end position, and then using a single finger to manipulate another interactive object of the two interactive objects to determine the interception starting position the total purpose is to manually manipulate any one of the two interactive objects with one hand, the specific implementation The method is not limited.
  • the user also manipulates two interactive objects simultaneously by using two fingers, for example, using two finger sliding operations on the screen to simultaneously manipulate two interactive objects, selecting a position, and then determining the selection by long press and release operations.
  • the interception starting position and the intercepting end position in the position are always aimed at two fingers interacting with two interactive objects at the same time, and the specific implementation manner is not limited.
  • the text information portion corresponding to the first audio segment may be displayed in a manner different from other text information portions.
  • the third operation can be triggered; in response to the third operation, the text information can be translated up or down, and the interactive object follows the text information up or down.
  • the first collection result may be played back.
  • a fourth operation may be triggered; and in response to the fourth operation, the first collection result may be saved to Local, uploaded to the server or sent to another terminal.
  • a fifth operation may be triggered; and in response to the fifth operation, steps S208-S211 are performed again.
  • a sixth operation may be triggered; and in response to the sixth operation, may be from the first collection result.
  • the intercepted portion obtains the second acquisition result.
  • the first collection result may be superimposed with other portions of the first audio file to obtain a third collection result.
  • the first collection result may be analyzed to obtain an analysis result; the analysis result may be compared with preset standard reference information to obtain a comparison result and the comparison result is displayed.
  • the first collection result may be uploaded to the server together with the comparison result.
  • FIG. 3 is a schematic diagram of an application scenario of the embodiment.
  • the first operation described above is a user's song operation
  • the first audio file is an accompaniment corresponding to the song selected by the user
  • the first text file is a lyric file corresponding to the song selected by the user.
  • the K song app can provide users with a list of recommended tracks or search portals for users to find the songs they want to sing. After the user clicks on the song, the terminal downloads the corresponding accompaniment and lyrics file from the server. The user can choose to sing the entire song or intercept the clip.
  • the K song App displays the lyrics of the entire song and two interactive objects on its interactive interface.
  • the two interactive objects are presented as two labels of "starting point” and "end point” for supporting the sliding operation.
  • the user determines the position of the two labels by the sliding gesture of the finger on the touch screen to intercept the clip that he wants to sing, wherein the lyric portion between the two labels of "starting point” and “end point” is highlighted in a manner different from other lyric parts. , for example, displayed in a different color, displayed in an enlarged font, and the like.
  • the second operation described above is an operation of the user clicking the "record only this segment” button.
  • the terminal audio output unit (such as a speaker or a headphone) starts playing the intercepted audio segment and simultaneously displays and displays on the display screen.
  • the lyrics corresponding to the audio clip.
  • the user sings with reference to the accompaniment and lyrics, and the terminal collects the user's voice through the microphone.
  • the terminal synthesizes the collected user voice and the accompaniment song, so that after the user sings the segment, the first collection result, that is, the user's vocal segment, can be obtained.
  • An information processing method of the embodiment of the present invention is applied to a terminal. As shown in FIG. 4, the method includes:
  • Step 401 trigger a first operation
  • Step 402 In response to the first operation, downloading the first audio file and the first text file that matches the first audio file, where the first text file includes time configuration information and text information.
  • Step 403 Pre-divide the text information in the first text file into a plurality of parts by using time configuration information
  • Step 404 Provide a third interaction object that supports the selected operation for each part of the text information.
  • Step 405 Determine, according to the indication information that the third interaction object is selected, a text information portion corresponding to the third interaction object;
  • Step 406 the beginning of the text information part is taken as the interception starting position, and the end of the text information part is taken as the intercepting end position;
  • Step 407 Generate first indication information for identifying a location of the interception start point and second indication information for identifying a location of the interception end point;
  • Step 408 The portion between the time point indicated by the first indication information and the time point indicated by the second indication information is intercepted from the first audio file, to obtain a first audio segment.
  • Step 409 intercepting the first indication information of the starting point position and the second indication information for identifying the intercepting end position to perform partial truncation of the first audio file to the first audio segment;
  • Step 410 trigger a second operation
  • Step 411 In response to the second operation, playing the first audio segment, and synchronously displaying the text information portion corresponding to the first audio segment in the first text file;
  • Step 412 Collecting voice information of the user while playing the first audio segment
  • Step 413 Synthesize the first audio segment and the voice information into a first collection result.
  • the method of the embodiment also provides a specific implementation method for intercepting and playing part of the audio file.
  • the text information in the first text file is pre-divided into a plurality of parts, and each part of the text information is provided to the user with an interactive object supporting the selection operation.
  • the user can determine the interception start position and the intercept end position by a single tap gesture on the touch screen or a mouse click manipulation of the interactive object by a finger or a stylus pen. In this way, the user can complete the operations required for the interception of the first audio segment by only one click action, so the method of the embodiment is faster than the method embodiment 2.
  • FIG. 5 is a schematic diagram of an application scenario of the embodiment.
  • the lyrics of the entire song are pre-divided into a plurality of parts, and the interactive object is presented as multiple radios supporting the selected operation. Box, each radio button corresponds to a lyric part.
  • the user selects a radio button by tapping a finger on the touch screen to determine a portion of the lyrics to be intercepted, wherein the selected lyric portion is highlighted in a manner different from other lyric portions, for example, displayed in a different color to enlarge the font. Show and so on.
  • the user clicks the "record only this paragraph" button the song is started.
  • An information processing method of the embodiment of the present invention is applied to a terminal. As shown in FIG. 6, the method includes:
  • Step 601 Trigger a first operation
  • Step 602 In response to the first operation, downloading the first audio file and the first text file that matches the first audio file, where the first text file includes time configuration information and text information.
  • Step 603 Pre-divide the text information in the first text file into a plurality of parts by using time configuration information
  • Step 604 Provide a fourth interaction object that supports the selected operation for each part of the text information.
  • Step 605 Determine, according to the indication information that at least two fourth interaction objects are selected, a beginning of a text information portion corresponding to the first one of the two second interaction objects as the interception start position, and a text information portion corresponding to the last one of the at least two fourth interaction objects The end is used as the interception end position;
  • Step 606 Generate first indication information for identifying a location of the interception start point and second indication information for identifying a location of the interception end point;
  • Step 607 The portion between the time point indicated by the first indication information and the time point indicated by the second indication information is intercepted from the first audio file, to obtain a first audio segment.
  • Step 608 triggering a second operation
  • Step 609 In response to the second operation, playing the first audio segment, and synchronously displaying the text information portion corresponding to the first audio segment in the first text file;
  • Step 610 Collect the voice information of the user while playing the first audio segment.
  • Step 611 Synthesize the first audio segment and the voice information into a first collection result.
  • the method of the embodiment also provides a specific implementation method for intercepting and playing part of the audio file.
  • multiple interactive objects can be selected at the same time.
  • the user can manipulate the interactive object by a tap gesture or a mouse click on the touch screen by a finger or a stylus pen, and the interception start position and the intercept end position are determined according to the selected first interactive object and the last interactive object.
  • the user can arbitrarily intercept the audio segment by, for example, two click actions according to his own needs, so that the method of the embodiment can achieve a more satisfactory intercepting effect with respect to the third embodiment of the method.
  • FIG. 7 is a schematic diagram of an application scenario of the embodiment.
  • the lyrics of the entire song are pre-divided into a plurality of parts, and the interactive object is presented as multiple checksums supporting the selected operation. Box, each check box corresponds to a lyric.
  • the user selects at least two check boxes by tapping a finger on the touch screen, the first one
  • the portion between the selected check box and the last selected check box is determined as the portion of the lyrics to be intercepted, and the selected lyric portion is highlighted in a manner different from other lyric portions, for example, displayed in different colors. , display in a magnified font, and more.
  • the song is started.
  • FIG. 8 is a schematic structural diagram of a system for implementing an embodiment of the present invention.
  • the system 800 includes one or more terminals 810 and one or more servers 830 that are connected by a network 820.
  • a K song App is installed in the terminal 810, and the user downloads the accompaniment and the corresponding lyrics from the server 830 by using the App, intercepts a segment from the accompaniment, and performs K song based on the accompaniment piece played by the terminal and the lyrics displayed simultaneously. Upload your own recorded arias and see others’ comments on the aria, listen to and comment on the arias recorded by others, and more.
  • Terminal 810 can be a smart phone, a WiFi headset, a tablet, a laptop, and the like.
  • 9 is a block diagram of a terminal 810 that implements an embodiment of the present invention.
  • the terminal 810 includes a processor 910, a memory 920, a networked device 930, a display device 940, an audio output device 950 (including speakers, headphones, etc.), and an acquisition device 960 (such as a microphone, a headset, etc.).
  • Processor 910 is configured to perform the methods of the above-described embodiments of the present invention in conjunction with other elements.
  • the networking device 930 is configured to enable communication between the terminal 810 and the server 830, such as obtaining a list of recommended accompaniment songs from the server 830 and recommending a vocal list, downloading accompaniment, lyrics, arias uploaded by others, and uploading the recorded arias to On the server, and so on.
  • the memory 920 is configured to store audio files and text files downloaded from the server, voice information collected by the collection device 960, and synthesized acquisition results.
  • the audio output device 950 is configured to play the accompaniment and the acquisition results, and the display device 940 is configured to simultaneously display the lyrics when playing the accompaniment or collecting the results.
  • the collection device 960 is configured to collect voice information of the user.
  • FIG. 10 is an interaction flowchart of implementing an embodiment of the present invention.
  • the interaction process involves the user 1, the terminal 810-1 used by the user 1, the user 2, the terminal 810-2 used by the user 2, and the server 830.
  • the functions implemented by the terminal 810-1 and the terminal 810-2 are distinguished here. It should be understood that these functions can be To be implemented on one terminal, and the entire karaoke system is not limited to two terminals.
  • the interaction process includes the following steps:
  • Step 1001 The user 1 starts an App by performing a tap gesture on the touch screen of the terminal 810-1;
  • Step 1002 The terminal 810-1 starts an App.
  • Step 1003 The server 830 pushes the recommended accompaniment list and the recommended vocal list to the terminal 810-1.
  • Step 1004 The user 1 clicks a song by performing a tap gesture on the touch screen of the terminal 810-1;
  • Step 1005 The terminal 810-1 sends a request for downloading the accompaniment corresponding to the song and the corresponding lyric file to the server;
  • Step 1006 the server 830 returns the accompaniment and the corresponding lyric file
  • Step 1007 the terminal 810-1 displays the lyrics and the start point label and the end point label that are available for the user to operate;
  • Step 1008 The user 1 slides the operation start label and the end point label to determine the start point and the end point;
  • Step 1009 The terminal 810-1 intercepts an audio segment between the start point and the end point from the accompaniment track;
  • Step 1010 The user 1 clicks the record button to send a command to start recording to the terminal 810-1.
  • Step 1011 The terminal 810-1 plays an audio segment, and dynamically displays the lyrics corresponding to the audio segment.
  • Step 1012 User 1 sings
  • Step 1013 The terminal 810-1 collects the singing voice of the user 1, and performs real-time scoring on the singing voice of the user for each sentence, and displays the single sentence score in real time on the display screen;
  • Step 1014 the terminal 810-1 synthesizes the singing voice of the user 1 and the audio segment to obtain the singing segment of the user 1, calculates the overall score of the singing segment according to the real-time score, and displays the overall score on the display screen;
  • Step 1015 The terminal 810-1 plays back the aria of the user 1 for the user 1 to listen to, and provides options for re-recording, uploading, etc. for the user 1 to select;
  • Step 1016 User 1 selects re-recording, and the interaction process jumps to step 1011; if the user selects upload, the next step is continued;
  • Step 1017 The terminal 810-1 uploads the vocal segment of the user 1 together with the overall score to the server 830;
  • Step 1018 The user 2 launches the App by performing a tap gesture on the touch screen of the terminal 810-2;
  • Step 1019 the terminal 810-2 starts the App
  • Step 1020 The server 830 pushes the recommended accompaniment list and the recommended vocal list to the terminal 810-2.
  • Step 1021 The user 2 selects a singer uploaded by the user 1 by performing a tap gesture on the touch screen of the terminal 810-2;
  • Step 1022 The terminal 810-2 sends a request for downloading the vocal segment to the server 830.
  • Step 1023 the server 830 returns the vocal segment and the corresponding lyrics
  • Step 1024 The terminal 810-2 plays the vocal segment and synchronously displays the corresponding lyrics
  • Step 1025 User 2 listens to the aria and evaluates it;
  • Step 1026 The terminal 810-2 uploads the evaluation of the user 2 to the server 830.
  • the present invention provides a client embodiment.
  • the following description of the client embodiment is similar to the above description of the method, and the beneficial effects of the same method are described without further description.
  • the client 810 includes:
  • the download unit 1101 is configured to download the first audio file and the first text file that matches the first audio file;
  • the intercepting unit 1102 is configured to perform partial truncation of the first audio file to the first audio segment according to the first indication information for identifying the intercept start position and the second indication information for identifying the intercept end position;
  • the playing unit 1103 is configured to play the first audio segment, and synchronously dynamically display the text information portion corresponding to the first audio segment in the first text file;
  • the collecting unit 1104 is configured to collect the language of the user while playing the first audio segment Audio information
  • the synthesizing unit 1105 is configured to synthesize the first audio segment and the voice information into a first collection result.
  • the intercepting unit 1102 includes:
  • a capture subunit 1102-1 configured to capture the intercept start location and the intercept endpoint location
  • the indication information generating unit 1102-2 is configured to generate first indication information for identifying the interception start position and second indication information for identifying the intercept end position,
  • the audio segment generating unit 1102-3 is configured to intercept, from the first audio file, a portion between a time point indicated by the first indication information and a time point indicated by the second indication information, to obtain a first audio. Fragment.
  • the capturing subunit 1102-1 is configured to provide a first interactive object and a second interactive object that support a sliding operation, by using the first interactive object and The second interactive object captures the intercept start position and the intercept end position simultaneously or separately.
  • the client 810 is applied to the terminal, except for the download unit 1101, the intercepting unit 1102, the playing unit 1103, the collecting unit 1104, and the synthesizing unit 1105 as shown in FIG.
  • the client further includes a dividing unit 1206 configured to pre-divide the text information in the first text file into a plurality of parts by using time configuration information of the first text file.
  • the capturing sub-unit 1102-1 is configured to provide, for each part of the text information, a third interactive object that supports a selected operation, according to a third interactive object.
  • the selected instruction information identifies a text information portion corresponding to the third interactive object, and the beginning of the text information portion is used as the intercept start position, and the end of the text information portion is used as the intercept end position.
  • the capturing subunit 1102-1 is configured to provide, for each part of the text information, a fourth interactive object that supports a selected operation, according to Determining information that at least two fourth interaction objects are selected, determining a beginning of a text information portion corresponding to the first one of the at least two fourth interaction objects as the interception start position, and the at least two The end of the text information portion corresponding to the last interactive object in the fourth interactive object serves as the intercept end position.
  • the client further includes: a difference flag unit configured to: after capturing the intercept start position and the intercept end position, the first audio The text information portion corresponding to the segment is displayed in a manner different from other text information portions.
  • the client further includes: a translation unit configured to translate the text information up or down, and cause the interactive object to follow the text information upward Or pan down.
  • the client further includes: a playback unit, configured to: after synthesizing the first audio segment and the voice information into a first collection result, The first collection result is described.
  • the client further includes: a storage unit configured to save the first collection result to the local; and an uploading unit configured to: the first collection The result is uploaded to the server.
  • the client further includes: an analyzing unit configured to analyze the first collecting result; and a comparing unit configured to analyze the result and the preset standard reference information The comparison is compared and displayed.
  • the integrated modules described in the embodiments of the present invention may also be stored in a computer readable storage medium if they are implemented in the form of software functional modules and sold or used as separate products. Based on such understanding, those skilled in the art will appreciate that embodiments of the present application can be provided as a method, system, or computer program product. Thus, the present application can take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment in combination of software and hardware. Moreover, the application can take the form of a computer program product embodied on one or more computer-usable storage media containing computer usable program code, including but not limited to a USB flash drive, a mobile hard drive, a read only memory ( ROM, Read-Only Memory), disk storage, CD-ROM, optical storage, etc.
  • the computer program instructions can also be stored in a computer readable memory that can direct a computer or other programmable data processing device to operate in a particular manner, such that the instructions stored in the computer readable memory produce an article of manufacture comprising the instruction device.
  • the apparatus implements the functions specified in one or more blocks of a flow or a flow and/or block diagram of the flowchart.
  • These computer program instructions can also be loaded onto a computer or other programmable data processing device such that a series of operational steps are performed on a computer or other programmable device to produce computer-implemented processing for execution on a computer or other programmable device.
  • the instructions provide steps for implementing the functions specified in one or more of the flow or in a block or blocks of a flow diagram.
  • the embodiment of the present invention further provides a computer storage medium, wherein computer executable instructions are stored, and the computer executable instructions are used to execute the foregoing information processing method of the embodiment of the present invention.
  • the information processing method and the client of the embodiment of the present invention can intercept and play part of the downloaded audio file, on the one hand, reduce the amount of collected information, and reduce the processing load of the terminal; on the other hand, the user can
  • the preferences and sound conditions capture the audio content that you are interested in or that you are capable of, thus enhancing the user experience.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • User Interface Of Digital Computer (AREA)
  • Information Transfer Between Computers (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

本发明公开了一种信息处理方法及客户端、计算机存储介质,其中,所述方法包括:触发第一操作;响应第一操作,下载第一音频文件及与第一音频文件相匹配的第一文字文件;根据用于标识截取起点位置的第一指示信息和用于标识截取终点位置的第二指示信息对第一音频文件进行部分截取得到第一音频片段;触发第二操作;响应第二操作,播放第一音频片段,并同步动态显示第一文字文件中与第一音频片段对应的文字信息部分;在播放第一音频片段的同时,采集用户的语音信息;将第一音频片段与语音信息合成为第一采集结果。

Description

一种信息处理方法及客户端、计算机存储介质 技术领域
本发明涉及通讯技术,尤其涉及一种信息处理方法及客户端、计算机存储介质。
背景技术
本申请发明人在实现本申请实施例技术方案的过程中,至少发现相关技术中存在如下技术问题:
近年来,随着智能终端的智能化,网络技术的高速发展,通过从服务器下载信息用于智能终端上各种应用程序(APP)的信息处理技术越来越多,能满足用户日常生活和工作的需要。
一种信息处理的应用场景可以是针对智能终端上一种常用的K歌类APP的场景,智能终端上安装的这类APP可以向用户播放从服务器下载得到的音频文件,比如歌曲对应的伴奏音乐,采集用户的语音信息并进行识别,将识别结果、或将对识别结果进一步与预设标准参照信息比较后得到的比较结果上传到服务器上以分享给其他人收听和评价。
鉴于这类APP只能够对下载的音频文件全部播放并同时采集用户的语音信息,比如供用户唱完并录制整首歌曲,从而导致以下两方面的问题:
一,采用现有技术无法对音频文件的部分内容进行截取播放,导致采集的信息量越大,对应的识别效果就会越复杂,识别所耗费的时间就会越长;
二、用户出于自身的喜好和声音条件,比如唱的音调比较高,或者比较低,导致与预设标准参照信息不匹配,从而得不到最佳的声音表现结果,也就是说,用户可能也只对其中音频文件的部分内容感兴趣,只需要截取音频文件的部分内容就可以,但是采用现有技术无法实现对音频文件的部分内容进行截取播放。
然而,相关技术中,对于上述问题,尚无有效解决方案。
发明内容
有鉴于此,本发明实施例希望提供一种信息处理方法及客户端、计算机存储介质,至少解决了现有技术存在的问题。
本发明实施例的技术方案是这样实现的:
本发明提供了一种信息处理方法,所述方法应用于终端中,所述方法包括:
触发第一操作;
响应所述第一操作,下载第一音频文件及与所述第一音频文件相匹配的第一文字文件;
根据用于标识截取起点位置的第一指示信息和用于标识截取终点位置的第二指示信息对所述第一音频文件进行部分截取得到第一音频片段;
触发第二操作;
响应所述第二操作,播放所述第一音频片段,并同步动态显示所述第一文字文件中与所述第一音频片段对应的文字信息部分;
在播放所述第一音频片段的同时,采集用户的语音信息;
将所述第一音频片段与所述语音信息合成为第一采集结果。
本发明又提供了一种客户端,所述客户端应用于终端中,所述客户端包括:
下载单元,被配置为下载第一音频文件及与所述第一音频文件相匹配的第一文字文件;
截取单元,被配置为根据用于标识截取起点位置的第一指示信息和用于标识截取终点位置的第二指示信息对所述第一音频文件进行部分截取得到第一音频片段;
播放单元,被配置为播放所述第一音频片段,并同步动态显示所述第一文字文件中与所述第一音频片段对应的文字信息部分;
采集单元,被配置为在播放所述第一音频片段的同时,采集用户的语音信息;
合成单元,被配置为将所述第一音频片段与所述语音信息合成为第一采集 结果。
所述下载单元,所述截取单元,所述播放单元,所述采集单元,所述合成单元可以采用中央处理器(CPU,Central Processing Unit)、数字信号处理器(DSP,Digital Singnal Processor)或可编程逻辑阵列(FPGA,Field-Programmable Gate Array)实现。
本发明实施例还提供一种计算机存储介质,其中存储有计算机可执行指令,该计算机可执行指令用于执行上述信息处理方法。
本发明实施例的一种信息处理方法,所述方法应用于终端中,所述方法包括:触发第一操作;响应所述第一操作,下载第一音频文件及与所述第一音频文件相匹配的第一文字文件;根据用于标识截取起点位置的第一指示信息和用于标识截取终点位置的第二指示信息对所述第一音频文件进行部分截取得到第一音频片段;触发第二操作;响应所述第二操作,播放所述第一音频片段,并同步动态显示所述第一文字文件中与所述第一音频片段对应的文字信息部分;在播放所述第一音频片段的同时,采集用户的语音信息;将所述第一音频片段与所述语音信息合成为第一采集结果。
采用本发明实施例的信息处理方法及客户端,能够对下载的音频文件的部分内容进行截取播放,一方面,减少采集的信息量,降低终端的处理负荷;另一方面,用户可以根据自身的喜好和声音条件截取自己感兴趣的或自己能力所及的音频内容,从而提升用户的使用体验。
附图说明
图1为本发明方法实施例一的一个实现流程示意图;
图2为本发明方法实施例二的一个实现流程示意图;
图3为本发明方法实施例二的一个应用场景的示意图;
图4为本发明方法实施例三的一个实现流程示意图;
图5为本发明方法实施例三的一个应用场景的示意图;
图6为本发明方法实施例四的一个实现流程示意图;
图7为应用本发明方法实施例四的一个应用场景的示意图;
图8为实现本发明实施例的***的结构示意图;
图9为实现本发明实施例的终端的结构示意图;
图10为实现本发明实施例的交互流程图;
图11为本发明客户端实施例一的一个组成结构示意图;
图12为本发明客户端实施例二的一个组成结构示意图。
具体实施方式
下面结合附图对技术方案的实施作进一步的详细描述。
方法实施例一:
本发明实施例的一种信息处理方法,所述方法应用于终端中,如图1所示,所述方法包括:
步骤101、触发第一操作;
步骤102、响应第一操作,下载第一音频文件及与第一音频文件相匹配的第一文字文件;
步骤103、根据用于标识截取起点位置的第一指示信息和用于标识截取终点位置的第二指示信息对第一音频文件进行部分截取得到第一音频片段;
步骤104、触发第二操作;
步骤105、响应第二操作,播放第一音频片段,并同步动态显示第一文字文件中与第一音频片段对应的文字信息部分;
步骤106、在播放第一音频片段的同时,采集用户的语音信息;
步骤107、将第一音频片段与语音信息合成为第一采集结果。
上述方法中的步骤在不矛盾的情况下可以以不同的顺序或并发的方式执行。例如,步骤102中的下载第一音频文件的操作可以步骤103的截取操作之前完成,也可以仅下载步骤103中的截取操作所必需的信息,稍后在步骤105中播放第一音频片段时,采用流媒体的方式,边下载稍后所需的音频数据边进行播放。
采用本实施例的方法,解决了现有技术中无法对音频文件的部分内容进行截取播放的技术问题。一方面,从设备的角度来说,截取音频文件的部分内容进行播放并采集相应的语音信息会大幅度降低采集的信息量,减轻终端的处理负荷;另一方面,从用户的角度来说,用户可以根据自身的喜好和声音条件截取自己感兴趣的或自己能力所及的音频内容,从而提升了用户使用体验。
方法实施例二:
本发明实施例的一种信息处理方法,所述方法应用于终端中,如图2所示,所述方法包括:
步骤201、触发第一操作;
步骤202、响应第一操作,下载第一音频文件及与第一音频文件相匹配的第一文字文件;
步骤203、提供支持进行滑动操作的第一交互对象和第二交互对象;
步骤204、利用第一交互对象和第二交互对象同时或分别捕获截取起点位置和截取终点位置;
步骤205、生成用于标识截取起点位置的第一指示信息和用于标识截取终点位置的第二指示信息;
步骤206、从第一音频文件中截取第一指示信息指示的时间点与第二指示信息指示的时间点之间的部分,得到第一音频片段;
步骤207、触发第二操作;
步骤208、响应第二操作,播放第一音频片段,并同步动态显示第一文字文件中与第一音频片段对应的文字信息部分;
步骤209、在播放第一音频片段的同时,采集用户的语音信息;
步骤210、将第一音频片段与语音信息合成为第一采集结果。
相对于方法实施例一,本实施例提供了一种对音频文件的部分内容进行截取播放的具体实现方法,其中向用户提供了两个可滑动操作的交互对象。例如,在用户使用具有触摸屏的终端的情况下,用户可以通过手指在触摸屏上的滑动手势操纵两个交互对象来确定截取起点位置和截取终点位置。
用户可以根据需要或根据自己的操作习惯利用单个手指分别操纵两个交互对象,比如,利用单个手指的触控操作来操纵两个交互对象中的一个交互对象来确定截取起点位置,在截取起点位置确定后,再利用单个手指操纵两个交互对象中的另一个交互对象来确定截取终点位置;当然,也可以利用单个手指的触控操作来操纵两个交互对象中的一个交互对象来确定截取终点位置,在截取终点位置确定后,再利用单个手指操纵两个交互对象中的另一个交互对象来确定截取起点位置,总归目的是单手分别操纵两个交互对象中的任意一个交互对象,具体实现方式不限。
或者,用户也利用两个手指同时操纵两个交互对象,比如,利用两个手指在屏幕上的滑动操作同时操纵两个交互对象,选定位置后,进而通过长按及释放操作来确定选定位置中的截取起点位置和截取终点位置,总归目的是两个手指同时操纵两个交互对象,具体实现方式不限。
无论单手还是双手的上述具体实现方式都极大地方便了用户的使用。
叠加方案时常用的是:
在本实施例的一个实施方式中,可以在捕获截取起点位置和截取终点位置之后,将与所述第一音频片段对应的文字信息部分以区别于其它文字信息部分的方式进行显示。
在本实施例的一个实施方式中,可以触发第三操作;响应第三操作,可以使文字信息向上或向下平移,并且交互对象跟随文字信息向上或向下平移。
在本实施例的一个实施方式中,在将第一音频片段与语音信息合成为第一采集结果之后,可以回放第一采集结果。
在本实施例的一个实施方式中,在将所述第一音频片段与所述语音信息合成为第一采集结果之后,可以触发第四操作;响应第四操作,可以将第一采集结果保存至本地、上传至服务器或者发送给另一终端。
在本实施例的一个实施方式中,在将所述第一音频片段与所述语音信息合成为第一采集结果之后,可以触发第五操作;响应第五操作,重新进行步骤S208-S211。
在本实施例的一个实施方式中,将所述第一音频片段与所述语音信息合成为第一采集结果之后,可以触发第六操作;响应第六操作,可以从所述第一采集结果中截取部分得到第二采集结果。
在本实施例的一个实施方式中,可以将所述第一采集结果与所述第一音频文件的其他部分进行叠加得到第三采集结果。
在本实施例的一个实施方式中,可以对第一采集结果进行分析,得到一分析结果;可以将分析结果与预设标准参照信息进行比较,得到比较结果并显示该比较结果。
在本实施例的一个实施方式中,可以将第一采集结果与比较结果一起上传至服务器。
在本实施例的一个应用场景中,所描述的方法可以由安装在终端上的K歌App执行。图3为本实施例的一个应用场景的示意图。这里,上述的第一操作为用户的点歌操作,第一音频文件为与用户所选择的歌曲对应的伴奏曲,第一文字文件为与用户所选择的歌曲对应的歌词文件。K歌App可以为用户提供推荐曲目列表或搜索入口,以供用户找到自己想唱的歌曲。在用户点歌之后,该终端从服务器下载对应的伴奏曲和歌词文件。用户可以选择唱整首歌或截取片段,当选择“截取片段”选项卡时,K歌App在其交互界面上显示整首歌的歌词以及两个交互对象。如图3所示,在该应用场景下,这两个交互对象呈现为支持进行滑动操作的“起点”和“终点”两个标签。用户通过手指在触摸屏上的滑动手势确定这两个标签的位置以截取自己想唱的片段,其中“起点”和“终点”两个标签之间的歌词部分以区别于其他歌词部分的方式突出显示,例如以不同的颜色显示、以放大字体显示等等。上述的第二操作为用户点击“只录此段”按钮的操作,当用户点击此按钮时,终端音频输出单元(例如扬声器或耳机)开始播放截取的音频片段,并在显示屏幕上同步显示与该音频片段对应的歌词。用户参考伴奏曲和歌词来唱歌,同时终端通过麦克风采集用户的声音。终端将采集的用户声音与伴奏曲进行合成,从而在用户唱完该片段之后,可以得到第一采集结果,即用户的唱段。
方法实施例三:
本发明实施例的一种信息处理方法,所述方法应用于终端中,如图4所示,所述方法包括:
步骤401、触发第一操作;
步骤402、响应第一操作,下载第一音频文件及与第一音频文件相匹配的第一文字文件,其中第一文字文件包括时间配置信息和文字信息;
步骤403、利用时间配置信息将第一文字文件中的文字信息预先划分成多个部分;
步骤404、针对所述文字信息的每个部分提供一个支持进行选中操作的第三交互对象;
步骤405、根据第三交互对象被选中的指示信息确定与该第三交互对象对应的文字信息部分;
步骤406、将该文字信息部分的开头作为截取起点位置,将该文字信息部分的末尾作为截取终点位置;
步骤407、生成用于标识截取起点位置的第一指示信息和用于标识截取终点位置的第二指示信息;
步骤408、从第一音频文件中截取第一指示信息指示的时间点与第二指示信息指示的时间点之间的部分,得到第一音频片段;
步骤409、截取起点位置的第一指示信息和用于标识截取终点位置的第二指示信息对第一音频文件进行部分截取得到第一音频片段;
步骤410、触发第二操作;
步骤411、响应第二操作,播放第一音频片段,并同步动态显示第一文字文件中与第一音频片段对应的文字信息部分;
步骤412、在播放第一音频片段的同时,采集用户的语音信息;
步骤413、将第一音频片段与语音信息合成为第一采集结果。
与方法实施例二类似的是,相对于方法实施例一,本实施例的方法也提供了一种具体的对音频文件的部分内容进行截取播放的实现方法。相对于方法实 施例一,在本实施例中,第一文字文件中的文字信息被预先划分成多个部分,针对文字信息的每个部分向用户提供了一个支持进行选中操作的交互对象。例如,用户可以通过手指或触屏笔在触摸屏上的单个轻敲手势或者鼠标单击操纵交互对象来确定截取起点位置和截取终点位置。如此一来,用户可以通过仅一个点击动作就完成第一音频片段的截取所需的操作,因此本实施例的方法相对于方法实施例二来说更加快捷。
在方法实施例二中描述的叠加方案同样适用于本实施例,在此不再重复描述。
在本实施例的一个应用场景中,所描述的方法也可以由安装在终端上的K歌App执行。图5为本实施例的一个应用场景的示意图。如图5所示,与图3所示的应用场景不同的是,在该应用场景下,整首歌的歌词被预先划分成多个部分,交互对象呈现为支持进行选中操作的多个单选框,每个单选框对应一个歌词部分。用户通过手指在触摸屏上的轻敲手势选中一个单选框以确定要截取的歌词部分,其中被选中的歌词部分以区别于其他歌词部分的方式突出显示,例如以不同的颜色显示、以放大字体显示等等。同样地,当用户点击“只录此段”按钮时,开始录歌。
方法实施例四:
本发明实施例的一种信息处理方法,所述方法应用于终端中,如图6所示,所述方法包括:
步骤601、触发第一操作;
步骤602、响应第一操作,下载第一音频文件及与第一音频文件相匹配的第一文字文件,其中第一文字文件包括时间配置信息和文字信息;
步骤603、利用时间配置信息将第一文字文件中的文字信息预先划分成多个部分;
步骤604、针对文字信息的每个部分提供一个支持进行选中操作的第四交互对象;
步骤605、根据至少两个第四交互对象被选中的指示信息,确定与所述至 少两个第四交互对象中的第一个交互对象对应的文字信息部分的开头作为所述截取起点位置,与所述至少两个第四交互对象中的最后一个交互对象对应的文字信息部分的末尾作为所述截取终点位置;
步骤606、生成用于标识截取起点位置的第一指示信息和用于标识截取终点位置的第二指示信息;
步骤607、从第一音频文件中截取第一指示信息指示的时间点与第二指示信息指示的时间点之间的部分,得到第一音频片段;
步骤608、触发第二操作;
步骤609、响应第二操作,播放第一音频片段,并同步动态显示第一文字文件中与第一音频片段对应的文字信息部分;
步骤610、在播放第一音频片段的同时,采集用户的语音信息;
步骤611、将第一音频片段与语音信息合成为第一采集结果。
与方法实施例二、三类似的是,相对于方法实施例一,本实施例的方法也提供了一种具体的对音频文件的部分内容进行截取播放的实现方法。相对于方法实施例三,在本实施例中,多个交互对象可以被同时选中。例如,用户可以通过手指或触屏笔在触摸屏上的轻敲手势或者鼠标单击操纵交互对象,截取起点位置和截取终点位置根据被选中的第一个交互对象和最后一个交互对象来确定。如此一来,用户可以根据自己的需要通过例如两次点击动作随意地截取音频片段,因此本实施例的方法相对于方法实施例三来说能够实现更加令人满意的截取效果。
在方法实施例二中描述的叠加方案同样适用于本实施例,在此不再重复描述。
在本实施例的一个应用场景中,所描述的方法也可以由安装在终端上的K歌App执行。图7为本实施例的一个应用场景的示意图。如图7所示,与图3所示的应用场景不同的是,在该应用场景下,整首歌的歌词被预先划分成多个部分,交互对象呈现为支持进行选中操作的多个复选框,每个复选框对应一句歌词。用户通过手指在触摸屏上的轻敲手势选中至少两个复选框,其中第一个 被选中的复选框和最后一个被选中的复选框之间的部分被确定为要截取的歌词部分,被选中的歌词部分以区别于其他歌词部分的方式突出显示,例如以不同的颜色显示、以放大字体显示等等。同样地,当用户点击“只录此段”按钮时,开始录歌。
以一个现实应用场景为例对本发明实施例阐述如下:
图8为实现本发明实施例的***的结构示意图。参照图8,其示出一个用于在终端上进行K歌的***。该***800包括一个或多个终端810以及一个或多个服务器830,这些终端810和服务器830通过网络820连接。终端810中安装有K歌App,用户通过使用该App来从服务器830下载伴奏曲和对应的歌词,从伴奏曲中截取一个片段,基于终端播放的伴奏曲片段并同步显示的歌词进行K歌,上传自己录制的唱段并查看他人对该唱段的评价、收听和评论他人录制的唱段,等等。
终端810可以是智能手机、WiFi耳机、平板电脑、笔记本电脑等等。图9为实现本发明实施例的终端810的框图。该终端810包括处理器910、存储器920、联网设备930、显示设备940、音频输出设备950(包括扬声器、耳机等)以及采集设备960(比如麦克风、耳麦等)。处理器910被配置为结合其他元件执行上述本发明实施例的方法。联网设备930被配置为使能该终端810与服务器830之间的通信,例如从服务器830获取推荐伴奏曲列表以及推荐唱段列表,下载伴奏曲、歌词、他人上传的唱段,将录制的唱段上传到服务器上,等等。存储器920被配置为存储从服务器下载的音频文件和文字文件、通过采集设备960采集的语音信息以及合成得到的采集结果。音频输出设备950被配置为播放伴奏曲以及采集结果,显示设备940被配置为在播放伴奏曲或采集结果时,同步显示歌词。采集设备960被配置为采集用户的语音信息。
下面结合本发明方法实施例二来描述实现该实施例的交互流程。图10为实现本发明实施例的交互流程图。该交互流程涉及用户1、用户1使用的终端810-1,用户2、用户2使用的终端810-2和服务器830。为了便于描述,这里对终端810-1和终端810-2实现的功能进行了区分。应该理解,这些功能均可 以在一个终端上实现,并且整个K歌***不限于两个终端。该交互流程包括以下步骤:
步骤1001、用户1通过在终端810-1的触摸屏上执行轻敲手势来启动App;
步骤1002、终端810-1启动App;
步骤1003、服务器830向终端810-1推送推荐伴奏曲列表和推荐唱段列表;
步骤1004、用户1通过在终端810-1的触摸屏上执行轻敲手势来点歌;
步骤1005、终端810-1向服务器发送下载所点歌曲对应的伴奏曲及对应的歌词文件的请求;
步骤1006、服务器830返回该伴奏曲及对应的歌词文件;
步骤1007、终端810-1显示歌词以及可供用户操作的起点标签和终点标签;
步骤1008、用户1滑动操作起点标签和终点标签以确定起点和终点;
步骤1009、终端810-1从伴奏曲中截取起点和终点之间的音频片段;
步骤1010、用户1点击录制按钮以向终端810-1发送开始录制的命令;
步骤1011、终端810-1播放音频片段,动态显示该与音频片段对应的歌词;
步骤1012、用户1唱歌;
步骤1013、终端810-1采集用户1的歌声,对用户1针对每一句的歌声进行实时打分,并在显示屏幕上实时显示单句得分;
步骤1014、终端810-1将用户1的歌声与音频片段进行合成得到用户1的唱段,根据实时得分计算出唱段的总体得分,并在显示屏幕上显示总体得分;
步骤1015、终端810-1回放用户1的唱段供用户1收听,并提供重新录制、上传等选项供用户1选择;
步骤1016、用户1选择重新录制,该交互流程跳转至步骤1011;用户选择上传,则继续执行下一步;
步骤1017:终端810-1将用户1的唱段与总体得分一起上传至服务器830;
步骤1018:用户2通过在终端810-2的触摸屏上执行轻敲手势来启动App;
步骤1019、终端810-2启动App;
步骤1020、服务器830向终端810-2推送推荐伴奏曲列表和推荐唱段列表;
步骤1021、用户2通过在终端810-2的触摸屏上执行轻敲手势来选择用户1上传的唱段;
步骤1022、终端810-2向服务器830发送下载该唱段的请求;
步骤1023、服务器830返回该唱段及对应的歌词;
步骤1024、终端810-2播放该唱段,并同步显示对应的歌词;
步骤1025:用户2收听该唱段,并进行评价;
步骤1026:终端810-2将用户2的评价上传至服务器830。
这里需要指出的是:以上描述的交互流程仅仅是一个示例,其中的实时打分、计算总体得分、上传以及评价等步骤是可选的;截取音频片段的操作也可以用其它方式来实现,例如以方法实施例三或四所述的方式来实现;以上描述中提及的从推荐伴奏曲列表和推荐唱段列表中选择伴奏曲和唱段的方式,仅仅是一个示例而已,用户可以通过其他手段,例如关键词搜索,来选择自己感兴趣的伴奏曲和唱段。
基于上面方法实施例,本发明提供了客户端实施例。以下客户端实施例的描述,与上述方法描述是类似的,同方法的有益效果描述,不做赘述。对于本发明客户端实施例中未披露的技术细节,请参照本发明方法实施例的描述。
客户端实施例一:
本发明实施例的一种客户端,所述客户端应用于终端中,如图11所示,客户端810包括:
下载单元1101,被配置为下载第一音频文件及与所述第一音频文件相匹配的第一文字文件;
截取单元1102,被配置为根据用于标识截取起点位置的第一指示信息和用于标识截取终点位置的第二指示信息对所述第一音频文件进行部分截取得到第一音频片段;
播放单元1103,被配置为播放所述第一音频片段,并同步动态显示所述第一文字文件中与所述第一音频片段对应的文字信息部分;
采集单元1104,被配置为在播放所述第一音频片段的同时,采集用户的语 音信息;
合成单元1105,被配置为将所述第一音频片段与所述语音信息合成为第一采集结果。
在本实施例的一个具体实现方式中,可选地,所述截取单元1102包括:
捕获子单元1102-1,被配置为捕获所述截取起点位置和所述截取终点位置;
指示信息生成单元1102-2,被配置为生成用于标识所述截取起点位置的第一指示信息和用于标识所述截取终点位置的第二指示信息,
音频片段生成单元1102-3,被配置为从所述第一音频文件中截取所述第一指示信息指示的时间点与所述第二指示信息指示的时间点之间的部分,得到第一音频片段。
在本实施例的一个具体实现方式中,可选地,所述捕获子单元1102-1被配置为提供支持进行滑动操作的第一交互对象和第二交互对象,利用所述第一交互对象和所述第二交互对象同时或分别捕获所述截取起点位置和所述截取终点位置。
客户端实施例二:
本发明实施例的一种客户端,客户端810应用于终端中,除了如图11所示的下载单元1101、截取单元1102、播放单元1103、采集单元1104、合成单元1105之外,如图12所示,所述客户端还包括:划分单元1206,被配置为利用第一文字文件的时间配置信息将所述第一文字文件中的文字信息预先划分成多个部分。
在本实施例的一个具体实现方式中,所述捕获子单元1102-1被配置为针对所述文字信息的每个部分提供一个支持进行选中操作的第三交互对象,根据一个第三交互对象被选中的指示信息确定与该第三交互对象对应的文字信息部分,将该文字信息部分的开头作为所述截取起点位置,将该文字信息部分的末尾作为所述截取终点位置。
在本实施例的另一个具体实现方式中,所述捕获子单元1102-1被配置为针对所述文字信息的每个部分提供一个支持进行选中操作的第四交互对象,根据 至少两个第四交互对象被选中的指示信息,确定与所述至少两个第四交互对象中的第一个交互对象对应的文字信息部分的开头作为所述截取起点位置,与所述至少两个第四交互对象中的最后一个交互对象对应的文字信息部分的末尾作为所述截取终点位置。
在本实施例的一个具体实现方式中,可选地,该客户端进一步包括:区别标志单元,被配置为在捕获所述截取起点位置和所述截取终点位置之后,将与所述第一音频片段对应的文字信息部分以区别于其它文字信息部分的方式进行显示。
在本实施例的一个具体实现方式中,可选地,该客户端进一步包括:平移单元,被配置为使所述文字信息向上或向下平移,并且使所述交互对象跟随所述文字信息向上或向下平移。
在本实施例的一个具体实现方式中,可选地,该客户端进一步包括:回放单元,被配置为在将所述第一音频片段与所述语音信息合成为第一采集结果之后,回放所述第一采集结果。
在本实施例的一个具体实现方式中,可选地,该客户端进一步包括:存储单元,被配置为将所述第一采集结果保存至本地;上传单元,被配置为将所述第一采集结果上传至服务器。
在本实施例的一个具体实现方式中,可选地,该客户端进一步包括:分析单元,被配置为对第一采集结果进行分析;比较单元,被配置为将分析结果与预设标准参照信息比较得到比较结果并显示。
本发明实施例所述集成的模块如果以软件功能模块的形式实现并作为独立的产品销售或使用时,也可以存储在一个计算机可读取存储介质中。基于这样的理解,本领域内的技术人员应明白,本申请的实施例可提供为方法、***、或计算机程序产品。因此,本申请可采用完全硬件实施例、完全软件实施例、或结合软件和硬件方面的实施例的形式。而且,本申请可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质上实施的计算机程序产品的形式,所述存储介质包括但不限于U盘、移动硬盘、只读存储器(ROM, Read-Only Memory)、磁盘存储器、CD-ROM、光学存储器等。
本申请是根据本申请实施例的方法、设备(***)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。
这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能。
这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。
尽管已描述了本申请的优选实施例,但本领域内的技术人员一旦得知了基本创造性概念,则可对这些实施例做出另外的变更和修改。所以,所附权利要求意欲解释为包括优选实施例以及落入本申请范围的所有变更和修改。
相应的,本发明实施例还提供一种计算机存储介质,其中存储有计算机可执行指令,该计算机可执行指令用于执行本发明实施例的上述信息处理方法。
工业实用性
采用本发明实施例的信息处理方法及客户端,能够对下载的音频文件的部分内容进行截取播放,一方面,减少采集的信息量,降低终端的处理负荷;另一方面,用户可以根据自身的喜好和声音条件截取自己感兴趣的或自己能力所及的音频内容,从而提升用户的使用体验。

Claims (21)

  1. 一种信息处理方法,所述方法应用于终端中,所述方法包括:
    触发第一操作;
    响应所述第一操作,下载第一音频文件及与所述第一音频文件相匹配的第一文字文件;
    根据用于标识截取起点位置的第一指示信息和用于标识截取终点位置的第二指示信息对所述第一音频文件进行部分截取得到第一音频片段;
    触发第二操作;
    响应所述第二操作,播放所述第一音频片段,并同步动态显示所述第一文字文件中与所述第一音频片段对应的文字信息部分;
    在播放所述第一音频片段的同时,采集用户的语音信息;
    将所述第一音频片段与所述语音信息合成为第一采集结果。
  2. 根据权利要求1所述的方法,其中,根据用于标识截取起点位置的第一指示信息和用于标识截取终点位置的第二指示信息对所述第一音频文件进行部分截取得到第一音频片段包括:
    捕获所述截取起点位置和所述截取终点位置;
    生成用于标识所述截取起点位置的第一指示信息和用于标识所述截取终点位置的第二指示信息;
    从所述第一音频文件中截取所述第一指示信息指示的时间点与所述第二指示信息指示的时间点之间的部分,得到所述第一音频片段。
  3. 根据权利要求2所述的方法,其中,捕获所述截取起点位置和所述截取终点位置包括:
    提供支持进行滑动操作的第一交互对象和第二交互对象;
    利用所述第一交互对象和所述第二交互对象同时或分别捕获所述截取起点位置和所述截取终点位置。
  4. 根据权利要求2所述的方法,其中,所述第一文字文件包括时间配置信 息和文字信息,所述方法进一步包括:利用所述时间配置信息将所述第一文字文件中的文字信息预先划分成多个部分;
    捕获所述截取起点位置和所述截取终点位置包括:
    针对所述文字信息的每个部分提供一个支持进行选中操作的第三交互对象;
    根据所述第三交互对象被选中的指示信息确定与所述第三交互对象对应的文字信息部分;
    将该文字信息部分的开头作为所述截取起点位置,将该文字信息部分的末尾作为所述截取终点位置。
  5. 根据权利要求2所述的方法,其中,所述第一文字文件包括时间配置信息和文字信息,所述方法进一步包括:利用所述时间配置信息将所述第一文字文件中的文字信息预先划分成多个部分;
    捕获所述截取起点位置和所述截取终点位置包括:
    针对所述文字信息的每个部分提供一个支持进行选中操作的第四交互对象;
    根据所述至少两个第四交互对象被选中的指示信息,确定与所述至少两个第四交互对象中的第一个交互对象对应的文字信息部分的开头作为所述截取起点位置,与所述至少两个第四交互对象中的最后一个交互对象对应的文字信息部分的末尾作为所述截取终点位置。
  6. 根据权利要求3至5中任一项所述的方法,其中,所述方法进一步包括:在捕获所述截取起点位置和所述截取终点位置之后,将与所述第一音频片段对应的文字信息部分以区别于其它文字信息部分的方式进行显示。
  7. 根据权利要求3至5中任一项所述的方法,其中,所述方法进一步包括:
    触发第三操作;
    响应所述第三操作,使所述文字信息向上或向下平移,并且所述交互对象跟随所述文字信息向上或向下平移。
  8. 根据权利要求1所述的方法,其中,所述方法进一步包括:
    在将所述第一音频片段与所述语音信息合成为第一采集结果之后,回放所述第一采集结果。
  9. 根据权利要求1所述的方法,其中,所述方法进一步包括:
    在将所述第一音频片段与所述语音信息合成为第一采集结果之后,触发第四操作;
    响应第四操作,将所述第一采集结果保存至本地、上传至服务器或者发送给另一终端。
  10. 根据权利要求1所述的方法,其中,所述方法进一步包括:
    对所述第一采集结果进行分析,得到一分析结果;
    将所述分析结果与预设标准参照信息进行比较,得到比较结果并显示所述比较结果。
  11. 一种客户端,所述客户端应用于终端中,所述客户端包括:
    下载单元,被配置为下载第一音频文件及与所述第一音频文件相匹配的第一文字文件;
    截取单元,被配置为根据用于标识截取起点位置的第一指示信息和用于标识截取终点位置的第二指示信息对所述第一音频文件进行部分截取得到第一音频片段;
    播放单元,被配置为播放所述第一音频片段,并同步动态显示所述第一文字文件中与所述第一音频片段对应的文字信息部分;
    采集单元,被配置为在播放所述第一音频片段的同时,采集用户的语音信息;
    合成单元,被配置为将所述第一音频片段与所述语音信息合成为第一采集结果。
  12. 根据权利要求1所述的客户端,其中,所述截取单元包括:
    捕获子单元,被配置为捕获所述截取起点位置和所述截取终点位置;
    指示信息生成单元,被配置为生成用于标识所述截取起点位置的第一指示信息和用于标识所述截取终点位置的第二指示信息,
    音频片段生成单元,被配置为从所述第一音频文件中截取所述第一指示信息指示的时间点与所述第二指示信息指示的时间点之间的部分,得到所述第一音频片段。
  13. 根据权利要求12所述的客户端,其中,所述捕获子单元被配置为提供支持进行滑动操作的第一交互对象和第二交互对象,利用所述第一交互对象和所述第二交互对象同时或分别捕获所述截取起点位置和所述截取终点位置。
  14. 根据权利要求12所述的客户端,其中,所述第一文字文件包括时间配置信息和文字信息,所述客户端进一步包括:划分单元,被配置为利用所述时间配置信息将所述第一文字文件中的文字信息预先划分成多个部分;
    所述捕获子单元被配置为针对所述文字信息的每个部分提供一个支持进行选中操作的第三交互对象,根据所述第三交互对象被选中的指示信息确定与所述第三交互对象对应的文字信息部分,将该文字信息部分的开头作为所述截取起点位置,将该文字信息部分的末尾作为所述截取终点位置。
  15. 根据权利要求12所述的客户端,其中,所述第一文字文件包括时间配置信息和文字信息,所述客户端进一步包括:划分单元,被配置为利用时间配置信息将所述第一文字文件中的文字信息预先划分成多个部分,
    所述捕获子单元被配置为针对所述文字信息的每个部分提供一个支持进行选中操作的第四交互对象,根据至少两个第四交互对象被选中的指示信息,确定与所述至少两个第四交互对象中的第一个交互对象对应的文字信息部分的开头作为所述截取起点位置,与所述至少两个第四交互对象中的最后一个交互对象对应的文字信息部分的末尾作为所述截取终点位置。
  16. 根据权利要求13至15中任一项所述的客户端,其中,所述客户端进一步包括:区别标志单元,被配置为在捕获所述截取起点位置和所述截取终点位置之后,将与所述第一音频片段对应的文字信息部分以区别于其它文字信息部分的方式进行显示。
  17. 根据权利要求13至15中任一项所述的客户端,其中,所述客户端进一步包括:
    平移单元,被配置为使所述文字信息向上或向下平移,并且使所述交互对象跟随所述文字信息向上或向下平移。
  18. 根据权利要求11所述的客户端,其中,所述客户端进一步包括:
    回放单元,被配置为在将所述第一音频片段与所述语音信息合成为第一采集结果之后,回放所述第一采集结果。
  19. 根据权利要求11所述的客户端,其中,所述客户端进一步包括:
    存储单元,被配置为将所述第一采集结果保存至本地;
    上传单元,被配置为将所述第一采集结果上传至服务器。
  20. 根据权利要求11所述的客户端,其中,所述客户端进一步包括:
    分析单元,被配置为对所述第一采集结果进行分析,得到一分析结果;
    比较单元,被配置为将所述分析结果与预设标准参照信息进行比较,得到比较结果并显示所述比较结果。
  21. 一种计算机存储介质,其中存储有计算机可执行指令,该计算机可执行指令用于执行所述权利要求1至10任一项所述的方法。
PCT/CN2016/070683 2015-01-12 2016-01-12 一种信息处理方法及客户端、计算机存储介质 WO2016112841A1 (zh)

Priority Applications (3)

Application Number Priority Date Filing Date Title
MYPI2016702118A MY187298A (en) 2015-01-12 2016-01-12 Method, client and computer storage medium for processing information
US15/187,056 US9858910B2 (en) 2015-01-12 2016-06-20 Method, client and computer storage medium for processing information
US15/824,495 US10580394B2 (en) 2015-01-12 2017-11-28 Method, client and computer storage medium for processing information

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201510014200.7 2015-01-12
CN201510014200.7A CN105845158A (zh) 2015-01-12 2015-01-12 一种信息处理方法及客户端

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US15/187,056 Continuation US9858910B2 (en) 2015-01-12 2016-06-20 Method, client and computer storage medium for processing information

Publications (1)

Publication Number Publication Date
WO2016112841A1 true WO2016112841A1 (zh) 2016-07-21

Family

ID=56405247

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2016/070683 WO2016112841A1 (zh) 2015-01-12 2016-01-12 一种信息处理方法及客户端、计算机存储介质

Country Status (4)

Country Link
US (2) US9858910B2 (zh)
CN (1) CN105845158A (zh)
MY (1) MY187298A (zh)
WO (1) WO2016112841A1 (zh)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115119069A (zh) * 2021-03-17 2022-09-27 阿里巴巴新加坡控股有限公司 多媒体内容处理方法、电子设备及计算机存储介质

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105023559A (zh) * 2015-05-27 2015-11-04 腾讯科技(深圳)有限公司 K歌处理方法及***
CN104966527B (zh) * 2015-05-27 2017-04-19 广州酷狗计算机科技有限公司 K歌处理方法、装置以及k歌处理***
CN108428441B (zh) * 2018-02-09 2021-08-06 咪咕音乐有限公司 多媒体文件生成方法、电子设备和存储介质
CN108829318A (zh) * 2018-06-10 2018-11-16 北京酷我科技有限公司 一种截取音频控件的算法
CN109285568A (zh) * 2018-10-12 2019-01-29 北京微播视界科技有限公司 用于生成音频文件的方法和装置
CN115237316A (zh) * 2022-06-06 2022-10-25 华为技术有限公司 一种音轨标记方法及电子设备

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102377914A (zh) * 2011-09-22 2012-03-14 宇龙计算机通信科技(深圳)有限公司 终端和多媒体文件管理方法
JP2012159575A (ja) * 2011-01-31 2012-08-23 Daiichikosho Co Ltd 複数歌唱者による歌唱誘導システム
CN103839565A (zh) * 2012-11-23 2014-06-04 宏达国际电子股份有限公司 音乐歌词显示方法及使用此方法的便携型手持式电子装置
CN104966527A (zh) * 2015-05-27 2015-10-07 腾讯科技(深圳)有限公司 K歌处理方法、装置以及k歌处理***
CN105006234A (zh) * 2015-05-27 2015-10-28 腾讯科技(深圳)有限公司 一种k歌处理方法及装置

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003001477A1 (en) * 2001-06-18 2003-01-03 First International Digital, Inc. Music player with synchronized recording and display
US20060077767A1 (en) * 2004-09-13 2006-04-13 Richard Young Dialog-reading game with background music and sound effects cued to text
CA2606750C (en) 2005-05-11 2015-11-24 Nanolytics, Inc. Method and device for conducting biochemical or chemical reactions at multiple temperatures
US20060277217A1 (en) * 2005-06-01 2006-12-07 Nokia Corporation Method for creating a data file
CN1878217A (zh) * 2006-06-30 2006-12-13 深圳市杰特电信控股有限公司 一种手机的铃声制作方法
CN101131693A (zh) * 2006-08-25 2008-02-27 佛山市顺德区顺达电脑厂有限公司 音乐播放***及其方法
CN101207871B (zh) * 2006-12-22 2011-05-25 联想移动通信科技有限公司 一种移动终端及音乐铃声的播放方法
TWI335528B (en) * 2007-05-15 2011-01-01 Htc Corp A device with multiple functions, and a method for switching the functions and related electronic devices thereof
US20090031883A1 (en) * 2007-07-31 2009-02-05 Christopher Lee Bennetts Networked karaoke system and method
US8929877B2 (en) * 2008-09-12 2015-01-06 Digimarc Corporation Methods and systems for content processing
JP2011215358A (ja) * 2010-03-31 2011-10-27 Sony Corp 情報処理装置、情報処理方法及びプログラム
US9312969B2 (en) * 2010-04-15 2016-04-12 North Eleven Limited Remote server system for combining audio files and for managing combined audio files for downloading by local systems
EP2759126B8 (en) * 2011-09-18 2021-03-31 Touchtunes Music Corporation Digital jukebox device with karaoke and/or photo booth features, and associated methods
CN103345931A (zh) * 2013-06-20 2013-10-09 张家港保税区润桐电子技术研发有限公司 一种音乐播放方法
CN103400593B (zh) * 2013-07-03 2016-02-24 网易(杭州)网络有限公司 一种试听音频的方法及装置
US20140336803A1 (en) * 2014-07-24 2014-11-13 Keely Burke Haverstock Time management and task persistence method, apparatus and system

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012159575A (ja) * 2011-01-31 2012-08-23 Daiichikosho Co Ltd 複数歌唱者による歌唱誘導システム
CN102377914A (zh) * 2011-09-22 2012-03-14 宇龙计算机通信科技(深圳)有限公司 终端和多媒体文件管理方法
CN103839565A (zh) * 2012-11-23 2014-06-04 宏达国际电子股份有限公司 音乐歌词显示方法及使用此方法的便携型手持式电子装置
CN104966527A (zh) * 2015-05-27 2015-10-07 腾讯科技(深圳)有限公司 K歌处理方法、装置以及k歌处理***
CN105006234A (zh) * 2015-05-27 2015-10-28 腾讯科技(深圳)有限公司 一种k歌处理方法及装置

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
3533 MOBILE PHONE WORLD: "How to Intercept the Fragments of Recoding Songs of MASS KARAOKE?", 15 September 2014 (2014-09-15), pages 1 - 2, Retrieved from the Internet <URL:http://www.3533.com/news/15/201409/105524> *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115119069A (zh) * 2021-03-17 2022-09-27 阿里巴巴新加坡控股有限公司 多媒体内容处理方法、电子设备及计算机存储介质

Also Published As

Publication number Publication date
US10580394B2 (en) 2020-03-03
US9858910B2 (en) 2018-01-02
CN105845158A (zh) 2016-08-10
MY187298A (en) 2021-09-20
US20180151163A1 (en) 2018-05-31
US20160300557A1 (en) 2016-10-13

Similar Documents

Publication Publication Date Title
WO2016112841A1 (zh) 一种信息处理方法及客户端、计算机存储介质
US11456017B2 (en) Looping audio-visual file generation based on audio and video analysis
US7973230B2 (en) Methods and systems for providing real-time feedback for karaoke
EP2760014B1 (en) Interactive score curve for adjusting audio parameters of a user&#39;s recording.
CN106468987B (zh) 一种信息处理方法及客户端
US20130006627A1 (en) Method and System for Communicating Between a Sender and a Recipient Via a Personalized Message Including an Audio Clip Extracted from a Pre-Existing Recording
WO2016188211A1 (zh) 音频处理方法、装置及***
US9286943B2 (en) Enhancing karaoke systems utilizing audience sentiment feedback and audio watermarking
JP2012247957A (ja) データ処理装置およびプログラム
JP2014520352A (ja) エンハンスされたメディア記録およびプレイバック
US20190103084A1 (en) Singing voice edit assistant method and singing voice edit assistant device
WO2022160603A1 (zh) 歌曲的推荐方法、装置、电子设备及存储介质
CN106653067B (zh) 一种信息处理方法及终端
CN107731249B (zh) 一种音频文件制作方法及移动终端
JP2023534975A (ja) ユーザーインタラクションに基づく音楽再生方法、装置、機器及び記憶媒体
US20140122606A1 (en) Information processing device, information processing method, and program
CN105373585B (zh) 歌曲收藏方法和装置
US20140142932A1 (en) Method for Producing Audio File and Terminal Device
JP7128222B2 (ja) 映像コンテンツに対する合成音のリアルタイム生成を基盤としたコンテンツ編集支援方法およびシステム
JP7166373B2 (ja) 音声ファイルに対するテキスト変換記録とメモをともに管理する方法、システム、およびコンピュータ読み取り可能な記録媒体
WO2023217003A1 (zh) 音频处理方法、装置、设备及存储介质
JP6255960B2 (ja) 音楽データ生成方法を実現するためのプログラムおよび音楽データ生成装置
KR20180034718A (ko) 마인드맵을 활용한 뮤직 제공 방법 및 이를 실행하는 서버
JP6182493B2 (ja) 音楽再生システム、サーバ、及びプログラム
KR101472034B1 (ko) 라디오 방송 시스템, 라디오 음원 정보 제공 방법 및 라디오 음원 구매 방법

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: IDP00201604567

Country of ref document: ID

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16737066

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205N DATED 07/12/18)

122 Ep: pct application non-entry in european phase

Ref document number: 16737066

Country of ref document: EP

Kind code of ref document: A1