WO2024093525A1 - 信息处理方法和电子设备 - Google Patents

信息处理方法和电子设备 Download PDF

Info

Publication number
WO2024093525A1
WO2024093525A1 PCT/CN2023/117685 CN2023117685W WO2024093525A1 WO 2024093525 A1 WO2024093525 A1 WO 2024093525A1 CN 2023117685 W CN2023117685 W CN 2023117685W WO 2024093525 A1 WO2024093525 A1 WO 2024093525A1
Authority
WO
WIPO (PCT)
Prior art keywords
electronic device
voice
control
information
recording
Prior art date
Application number
PCT/CN2023/117685
Other languages
English (en)
French (fr)
Inventor
全瑞琳
王龙
Original Assignee
荣耀终端有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 荣耀终端有限公司 filed Critical 荣耀终端有限公司
Publication of WO2024093525A1 publication Critical patent/WO2024093525A1/zh

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04847Interaction techniques to control parameter settings, e.g. interaction with sliders or dials
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/166Editing, e.g. inserting or deleting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/60Editing figures and text; Combining figures or text
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72433User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for voice messaging, e.g. dictaphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72436User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for text messaging, e.g. short messaging services [SMS] or e-mails
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72439User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for image or video messaging

Definitions

  • the present application relates to the field of terminal technology, and in particular to an information processing method and an electronic device.
  • the "Memo" application can record and save input text information.
  • users can use electronic devices to record information instead of relying on traditional paper recording methods, which improves the efficiency of information recording.
  • the electronic device can integrate the user's recorded information into a text, and such a text can also be called a note.
  • the information recorded by the user through the electronic device can include text information, picture information, etc.
  • the present application provides an information processing method and an electronic device, which can record voice while recording non-voice information, and convert the recorded voice into text for recording at the same time, and can also turn off the voice-to-text function, thereby improving the efficiency of recording information and the flexibility of voice-to-text conversion.
  • the present application provides an information processing method, which includes: an electronic device displays a first interface, the first interface includes a first area, a second area and a first control in a first state; at a first moment, when the electronic device displays the second control and the first control in the first state, the electronic device records a first voice and displays text obtained by converting the first voice in the first area; the second control is used to indicate that recording is in progress, and the first control in the first state is used to indicate that the recorded voice is converted into text during recording; when the electronic device displays the second control and the first control in the first state, it receives first information input by a user and displays the first information in the second area; in response to an operation on the first control in the first state, when the second control is displayed, the electronic device switches the state of the first control to the second state, stops converting the recorded voice into text, and the first control in the second state is used to indicate pausing the conversion of the recorded voice into text; the electronic device records the first voice, the text obtained by converting the first voice
  • the first area may be area 1 involved in the embodiment
  • the second area may be area 2 involved in the embodiment
  • the first control may be a conversion control involved in the embodiment
  • the conversion control in the first state may be a conversion control in state C
  • the conversion control in the second state may be a conversion control in state D.
  • the second control may be a pause recording control involved in the embodiment
  • the first voice may be the voice recorded for the first time involved in the embodiment.
  • the electronic device can support voice-to-text while recording, and accept other information (such as text, pictures, etc.) input by the user, and display the other information in the second area.
  • other information such as text, pictures, etc.
  • the voice-to-text function does not convert the recorded voice into text information while recording. In this way, when the voice-to-text function is turned off, other information input by the user can be received during recording. Among them, the other information can also be called non-recording information.
  • the efficiency of recording information and the flexibility of voice-to-text are improved.
  • the first area includes a first sub-area and a second sub-area, the first sub-area is used to display the second control; the second sub-area is used to display the text obtained by voice conversion; when the electronic device displays the second control, the method also includes: in response to the operation on the second control, the electronic device replaces the second control with a third control; the third control is used to instruct the electronic device to end voice recording; after the voice recording ends, when the electronic device has not obtained the text obtained by the voice conversion, the first area does not include the second sub-area.
  • the first area can be divided into two sub-areas, wherein the first sub-area can be the area 11 involved in the embodiment, and the second sub-area can be the area 12 involved in the embodiment.
  • the first sub-area can be used to display content related to the recording process, such as controls related to recording, etc.
  • the second sub-area can display speech-to-text content.
  • the second sub-area can be closed and not displayed when the speech recording is paused, so that the user interface can be simplified.
  • the method also includes: when playing the first voice, the electronic device displays a progress bar, which is used to indicate the progress of playing the voice; when the first identifier is displayed at the first position of the progress bar, the electronic device displays the first information in a third state, and displays the second information in the second area in a fourth state; wherein the first identifier is used at the first position to indicate that the electronic device is playing the voice recorded at the first moment; the input time of the second information is a second moment, and the second moment is later than the first moment.
  • the progress bar may be the play progress bar involved in the embodiment
  • the first identifier may be the identifier 1411 involved in the embodiment
  • the first position may be the position corresponding to the currently played audio on the play progress bar
  • the third state is different from the fourth state.
  • the third state may be not adding a mask layer
  • the fourth state may be adding a mask layer.
  • the second information may be the non-recording information involved in the embodiment.
  • the electronic device can achieve audio-text synchronization during the process of playing voice. No mask is added to non-recording information that is recorded in the note earlier than the currently played voice. A mask is added to non-recording information that is recorded in the note later than the currently played voice, so that it is displayed separately to remind the user that the non-recording information with the mask has not been recorded in the note when the currently played voice is recorded in the note.
  • the method when the first mark is displayed at the first position of the progress bar, the method further includes: the electronic device displays text converted from the voice recorded at the first moment in the first area.
  • the electronic device can also achieve audio-text synchronization based on the speech-to-text content in the first area, that is, the text information converted from the voice being played can be displayed in the first area for user preview.
  • the method further includes: in response to an operation on a fourth control, the electronic device displays the second information in the third state.
  • the fourth control may be a control for stopping the note playback animation involved in the embodiment, and may include, for example, a note reply animation setting item 381 and a stop note playback animation control 351.
  • Four controls are used to turn off the audio-text synchronization function, providing users with more options.
  • the method when a first identifier is displayed at a first position of the progress bar, the method further includes: in response to a first operation on a second position in the progress bar, the electronic device displays third information in the second area in a fifth state; wherein the input time of the third information is a third moment, and the voice recorded by the electronic device at the third moment is indicated at the second position.
  • an example of the second position may be the position T involved in the embodiment, and the fifth state may be the highlighting involved in the embodiment.
  • the electronic device can also implement advance preview during voice playback. Long press a certain position on the progress bar, and when the voice corresponding to the position is input, the text input at the same time can be highlighted for user preview.
  • the method also includes: when Q segments of speech are recorded after the first file is established, the electronic device displays Q-1 nodes in the progress bar; the Q-1 nodes divide the progress bar into Q sub-progress bars; the i-th sub-progress bar in the Q sub-progress bars is used to indicate the i-th segment of speech in the Q segments of speech; Q is an integer greater than or equal to 2; after detecting an operation on the fifth control, when the first identifier is in the i-th sub-progress bar, the electronic device plays the i-th segment of speech.
  • the fifth control may be the start play control involved in the embodiment.
  • the electronic device may record multiple segments of voice information (voice) separately, and make the multiple segments of voice information into a long voice, and the multiple segments of voice may be played continuously when the long voice is played.
  • Q segments of voice means that the electronic device has made Q recordings. After a recording starts, the electronic device may collect sound signals. After the recording ends, a segment of voice information may be obtained based on the collected sound signals. That is, the electronic device may record at different times (separate Q time periods) to obtain Q segments of voice. Then, the electronic device may integrate the Q segments of voice to obtain a long voice. In this way, when playing a long voice, the Q segments of voice recorded separately may be played continuously.
  • the method also includes: the electronic device selects L segments of speech from Q segments of speech; L is an integer less than or equal to Q; when the sixth control is in the sixth state, in response to the operation of the seventh control, the electronic device deletes the L segments of speech, and deletes the text converted from the L segments of speech; when the sixth control is not in the sixth state, in response to the operation of the seventh control, the electronic device deletes the L segments of speech, and retains the text converted from the L segments of speech.
  • the sixth control may be the selection control 581b involved in the embodiment
  • the seventh control may be the deletion control 581c involved in the embodiment.
  • the electronic device may provide a function of whether to delete the text converted from the voice when deleting the voice, so that the user can have multiple choices when deleting the voice.
  • the method also includes: when the eighth control is in the eighth state, in response to the operation of the ninth control, the electronic device transmits the content in the second area and the text converted from the Q-segment speech to other devices; when the eighth control is not in the eighth state, in response to the operation of the ninth control, the electronic device transmits the content in the second area to other devices.
  • the eighth control may be a selected control involved in the embodiment (for example, the selected control 521b), and the The nine controls may be sharing controls (such as sharing controls) involved in the embodiments.
  • the electronic device may provide a function for sharing the text converted from the voice in the notes when sharing non-recorded information recorded in the notes, so that the user has multiple choices when sharing information.
  • receiving first input information specifically includes: the electronic device displays a first input tool in the first interface; the electronic device receives the first input information through the first input tool; the method also includes: when displaying the first input tool, also displaying a tenth control; in response to an operation on the tenth control, the electronic device switches the first input tool to a second input tool; when displaying the second input tool, also displaying an eleventh control; the eleventh control is used to trigger switching from the second input tool to the first input tool.
  • the first input tool may be the keyboard involved in the embodiment
  • the tenth control may be the handwriting control involved in the embodiment
  • the second input tool may be the stylus toolbar involved in the embodiment
  • the eleventh control may be the keyboard switching control involved in the embodiment.
  • the electronic device may enable the keyboard to be quickly switched to the stylus toolbar based on the tenth control.
  • the switching from the stylus toolbar to the keyboard may also be implemented based on the eleventh control. In this way, it is convenient for users to use different input tools to input non-recording information.
  • the method further includes: in response to an operation on a twelfth control; the electronic device changes the size of the second sub-area from the first size to the second size.
  • the twelfth control may be an expansion control or a collapse control involved in the embodiment.
  • the electronic device may make the size of the second sub-area adjustable to provide more choices for the user.
  • the present application provides an electronic device, comprising: one or more processors and a memory; the memory is coupled to the one or more processors, the memory is used to store computer program code, the computer program code includes computer instructions, and the one or more processors call the computer instructions to enable the electronic device to execute the method described in the first aspect or any one of the embodiments of the first aspect.
  • the electronic device can support voice-to-text while recording, and accept other information (such as text, pictures, etc.) input by the user, and display the other information in the second area.
  • other information such as text, pictures, etc.
  • the voice-to-text function can also be turned off, and the recorded voice can be not converted into text information while recording.
  • the voice-to-text function when the voice-to-text function is turned off, it is possible to accept other information input by the user while recording.
  • the other information can also be called non-recording information.
  • the efficiency of recording information and the flexibility of voice-to-text are improved.
  • an embodiment of the present application provides a chip system, which is applied to an electronic device, and the chip system includes one or more processors, which are used to call computer instructions so that the electronic device executes the method described in the first aspect or any one of the embodiments of the first aspect.
  • the electronic device can support voice-to-text conversion while recording, and accept other information (such as text, pictures, etc.) input by the user, and display the other information in the second area.
  • the voice-to-text function can be switched on and off. That is, while recording, the voice-to-text function can be turned on to convert the recording into text information, and the converted text information is displayed in the first area. In this way, when the voice-to-text function is turned on, other information input by the user can be accepted while recording.
  • the voice-to-text function can also be turned off, and the recorded voice can be not converted into text information while recording. In this way, when the voice-to-text function is turned off, other information input by the user can be accepted while recording.
  • the other information can also be called non-recording information.
  • the efficiency of recording information and the flexibility of voice-to-text conversion are improved.
  • an embodiment of the present application provides a computer program product comprising instructions, which, when executed on an electronic device, enables the electronic device to execute the method described in the first aspect or any one of the implementations of the first aspect.
  • the electronic device can support voice-to-text while recording, and accept other information (such as text, pictures, etc.) input by the user, and display the other information in the second area.
  • other information such as text, pictures, etc.
  • the voice-to-text function can also be turned off, and the recorded voice can be converted into text information while recording.
  • the voice-to-text function when the voice-to-text function is turned off, it is possible to accept other information input by the user while recording.
  • the other information can also be called non-recording information.
  • the efficiency of recording information and the flexibility of voice-to-text are improved.
  • an embodiment of the present application provides a computer-readable storage medium, comprising instructions, which, when executed on an electronic device, enable the electronic device to execute the method described in the first aspect or any one of the implementations of the first aspect.
  • the electronic device can support voice-to-text while recording, and accept other information (such as text, pictures, etc.) input by the user, and display the other information in the second area.
  • other information such as text, pictures, etc.
  • the voice-to-text function can also be turned off, and the recorded voice can be converted into text information while recording.
  • the voice-to-text function when the voice-to-text function is turned off, it is possible to accept other information input by the user while recording.
  • the other information can also be called non-recording information.
  • the efficiency of recording information and the flexibility of voice-to-text are improved.
  • FIG. 1A and 1B show an exemplary user interface for creating a new note on an electronic device
  • FIG8 shows an exemplary user interface involved in editing speech-to-text content after the first recording is completed
  • 13-17 illustrate a set of exemplary user interfaces involved in viewing information
  • 18-24 are a set of exemplary user interfaces involved in deleting information in an electronic device
  • 25-26, 27A-27B and 28-29 are a set of exemplary user interfaces involved in sharing information by an electronic device
  • FIG30 is a flowchart of an exemplary information processing method in the present application.
  • FIG31 is a schematic diagram of the structure of an electronic device provided in an embodiment of the present application.
  • first and second are used for descriptive purposes only and are not to be understood as suggesting or implying relative importance or implicitly indicating the number of technical features indicated.
  • a feature defined as “first” or “second” may explicitly or implicitly include one or more of the features, and in the description of the embodiments of the present application, unless otherwise specified, "plurality” means two or more.
  • the electronic device can record voice information by recording. While recording, it can also receive other information input by the user, such as one or more of text information (hereinafter referred to as text), picture information (hereinafter referred to as picture), etc.
  • text information hereinafter referred to as text
  • picture information hereinafter referred to as picture
  • an information processing method can be used to generate notes in an electronic device.
  • the information involved here includes but is not limited to voice information, text information or picture information, etc.
  • the information processing involved includes but is not limited to one or more of the following: recording information, viewing information and managing information. Among them, the recorded information includes input information, and processing based on the input information and using the processing results as the content in the notes. For example, converting the input voice information into text information, editing or marking the input text information, etc. Management information may also include deleting information, transmitting (sharing) information, etc.
  • the electronic device can realize one or more of the following functions, but is not limited to the following functions.
  • the electronic device can support voice-to-text while recording, and accept other information (such as text, pictures, etc.) input by the user.
  • other information such as text, pictures, etc.
  • you can also turn off the voice-to-text function and the recorded voice may not be converted into text information while recording. In this way, when the voice-to-text function is turned off, it is possible to accept other information input by the user while recording.
  • Other information may also be referred to as non-recorded information.
  • the electronic device can record multiple segments of voice information (voice) separately, and make the multiple segments of voice information into a long voice, so that the multiple segments of voice can be played continuously when the long voice is played.
  • Q segments of voice means that the electronic device has made Q recordings. After a recording starts, the electronic device can collect sound signals. After the recording ends, a segment of voice information can be obtained based on the collected sound signals. That is, the electronic device can record at different times (Q separate time periods) to obtain Q segments of voice. Then, the electronic device can integrate the Q segments of voice to obtain a long voice. In this way, when playing a long voice, the Q segments of voice recorded separately can be played continuously.
  • Q is an integer greater than or equal to 2.
  • the electronic device can display the playback progress bar corresponding to the long voice.
  • the playback progress bar includes Q-1 nodes, dividing the playback progress bar into Q segments (which can be called Q segment sub-progress bars). Among them, the i-th sub-progress bar in the Q segment sub-progress bar corresponds to the i-th voice.
  • the length of the i-th sub-progress in the playback progress bar is proportional to the ratio of the total duration of the i-th voice segment to the total duration of the long voice: the larger the ratio of the total duration of the i-th voice segment to the total duration of the long voice, the greater the proportion of the i-th voice segment in the long voice, and the longer the length of the i-th sub-progress bar in the playback progress bar.
  • the long speech involved here refers to the speech including more content than any of the multiple voice messages, and can be regarded as a collection of the multiple voice messages. It does not refer to the length of the long speech. The total duration of the long speech is not limited in the embodiment of the present application.
  • Function 1 and function 2 can make electronic devices more convenient to record information.
  • Figures 1A, 1B, and Figures 3-12 please refer to the following description of Figures 1A, 1B, and Figures 3-12, which will not be repeated here.
  • Audio-text synchronization includes: when playing a recording, the non-recorded information corresponding to the played voice is displayed in state A (recorded as information A), and the non-recorded information corresponding to the unplayed voice is displayed in state B (recorded as information B).
  • information A and information B are described by taking the electronic device playing voice A (recording time is time A) as an example:
  • time A may indicate a moment or a time range.
  • information A refers to information inputted at a time earlier than or equal to time A.
  • Information B refers to information inputted at a time later than time A.
  • information A refers to information inputted earlier than the start time of time A, or inputted earlier than the end time of time A.
  • Information B refers to information inputted later than the end time of time A.
  • the time A may also be referred to as the timestamp corresponding to the voice A.
  • the input time of the information A may be referred to as the timestamp corresponding to the information A.
  • the input time of the information B may be referred to as the timestamp corresponding to the information B.
  • Function 3 allows users to view information more conveniently through electronic devices.
  • Function 4 When managing information, the electronic device can delete information, share information, etc.
  • Deleting information includes: choosing to delete part or all of the voice information recorded in a note. For example, you can choose to delete the voice information, and when deleting the voice information, you can also determine whether to delete the text information corresponding to the selected voice information according to user needs.
  • Sharing information includes: transmitting the content to be shared in the note to other terminals, and the sharing method can be selected.
  • the content to be shared includes: other content in the note except voice information, or, other content in the note except voice information and text information converted from the voice information.
  • the user can choose whether to use the text information converted from the voice information as part of the content to be shared through the electronic device.
  • the sharing methods include but are not limited to one or more of the following sharing methods:
  • Sharing method 1 Convert the content to be shared in the note into a picture, and then transfer it to other devices in the form of a picture.
  • Sharing method Convert the content to be shared in the note into a document and then transfer it to other devices in the form of a document.
  • Sharing method 3 Transfer the content to be shared in the note to other terminals in the form of notes.
  • FIG. 1A A set of exemplary user interfaces involved in recording information are described in FIG. 1A , FIG. 1B , and FIG. 3 - FIG. 12 .
  • the content related to the electronic device creating a new note can refer to the following description of Figures 1A and 1B.
  • the first recording and its related content can refer to the following description of Figures 4-7.
  • the content related to the electronic device recording information can refer to the following description of Figure 8.
  • the second recording and its related content can refer to the following description of Figures 9-12.
  • the electronic device in response to an operation (e.g., a click operation) on a note-taking application (e.g., Honor Note 101), the electronic device may open the note-taking application. At this time, the electronic device may display a user interface 11a as shown in (2) of FIG1A .
  • an operation e.g., a click operation
  • a note-taking application e.g., Honor Note 101
  • user interface 11a is an exemplary user interface involved in creating a new note.
  • the user interface 11a may include a control 111 corresponding to viewing all notes.
  • the electronic device may display the user interface 11b shown in (3) of FIG. 1A below.
  • the user interface 11b may include a control 1131, which may trigger the electronic device to display the folder in which the note is saved.
  • the electronic device can display all folders, wherein all folders include folders created by the user through the electronic device (for example, collectively referred to as new folders), and also include default folders in the system.
  • Any non-empty folder may include at least one file, each of which may be referred to as a note. Empty folders do not include files.
  • the new folder may include folder 1, and the default folder may be independent of the new folder.
  • the electronic device may display all notes included in file 1. For example, at this time, the electronic device may display the user interface 11d shown in (1) of FIG. 1B described below.
  • the user interface 11d includes the notes involved in the file 1, including Note A and Note B.
  • the user interfaces 11a and 11c may also include a new control 113, which may be used to trigger a new pen.
  • a new control 113 which may be used to trigger a new pen.
  • the new creation control 113 in the user interface 11c the new note created by the electronic device belongs to the currently opened folder, which belongs to file 1 in this case.
  • the electronic device can create a new note.
  • the electronic device can display the user interface shown in (2) in Figure 1B.
  • the user interface 12 is an initialization interface involved after creating a new note.
  • the electronic device can set the name of the new note.
  • the new note is named Note 1.
  • the aforementioned Note 1 has not recorded any information input by the user.
  • the recording function can be triggered through the recording control.
  • the electronic device can divide the user interface into two areas (area 1 and area 2). Area 1 is used to present content related to recording, and area 1 can also be referred to as the recording area.
  • Area 2 is used to present other information input by the user (such as text, pictures, etc.). The other information can be referred to as non-recording information, and area 2 can also be referred to as an editing area.
  • the recording area (area 1) may further include two sub-areas: a voice recording area (recorded as area 11) and a voice-to-text area (recorded as area 12).
  • area 11 voice recording area
  • area 12 voice-to-text area
  • the voice recording and broadcasting area (area 11) may include functional controls involved in recording and broadcasting (recording or playing) voice, such as a recording bar.
  • the recording bar can be used to trigger functions such as recording voice, pausing recording voice, playing voice, and pausing playing voice.
  • the recording bar has different states, and the recording bar in different states is composed of different components. The state of the recording bar is different, and the function of the recording bar is different. For details, please refer to the description of the relevant content below, which will not be repeated here.
  • the speech-to-text area may include speech-to-text content and may also include a conversion control.
  • the conversion control is a control provided by the electronic device to control the speech-to-text function to be turned on or off.
  • the speech-to-text content includes text converted from the recorded speech.
  • the functional controls involved in recording or playing voice in the aforementioned area 11 may include other controls besides the recording bar, such as conversion controls, etc. This embodiment of the present application does not limit this.
  • the area 11 and the area 12 are displayed in one area (area 1) as an example for description, and in other possible situations, the area 11 and the area 12 can be separated.
  • the editing area (area 2) is used to present other information (non-recording information) input by the user, such as text, pictures, etc.
  • the electronic device can provide an input tool for inputting information.
  • the input tool can include a keyboard and a stylus tool bar.
  • the electronic device may display an input tool in a user interface to display the input tool.
  • a shortcut control for switching to other input tools can be provided.
  • the electronic device when the electronic device is recording, the user can view the recording status through the recording bar in area 11. While recording, the electronic device can also realize voice-to-text conversion, and display the converted text in area 12 for the user to view. While recording, the user can also realize manual input through the input tool, and the input information is displayed in area 2 for the user to view. Manual input includes inputting text, inserting pictures, etc.
  • user interface 13 is an exemplary user interface displayed after the electronic device enters the note.
  • the input tool provided by the electronic device by default is the keyboard.
  • a keyboard 131 may be displayed in user interface 13.
  • the electronic device may also display a note toolbar (e.g., note toolbar 132).
  • the note toolbar can provide an entry for convenient functions (such as a recording function) when the user records information.
  • the note toolbar 132 may include an insert control 132a, a recording control 132b, and a "handwriting" control 132c.
  • the insert control 132a can be used to trigger the input of information such as pictures and videos into the user interface 13 for display.
  • the recording control 132b can be used to trigger the start of the recording function.
  • the "handwriting" control 132c can be used to trigger the switching of the input tool from the keyboard 131 to the stylus toolbar.
  • the note toolbar may also include other controls, for example, it may also include a list creation control 132d and a style setting control 132e.
  • the user interface 13 may also include an editing area 133, which may be used to display non-recording information input by the user.
  • an editing area 133 which may be used to display non-recording information input by the user.
  • the user may input text information into the editing area 133 for display through the keyboard 131, or may insert pictures and other information into the editing area 133 for display through the insert control 132a.
  • the electronic device switches the input tool from the keyboard 131 to the stylus toolbar.
  • the stylus toolbar may be the stylus toolbar 134 displayed in the user interface 13a.
  • the stylus toolbar 134 may include different styluses, such as stylus 134a-stylus 134f.
  • the stylus can be used to mark in the editing area, for example, to add an underline, draw a circle, etc.
  • the user interface 13a may also include a keyboard switching control 135.
  • the keyboard switching control 135 can be used to trigger the electronic device to switch the input tool from the stylus toolbar to the keyboard.
  • the keyboard switch control 135 can be used as a shortcut control in the stylus toolbar 134 .
  • the electronic device can turn on the recording function, and divide the aforementioned area 1 and area 2 in the user interface. At this time, the electronic device can switch the state of the recording control 132b from state A1 to state B1.
  • the recording control 132b in state B1 can indicate that the recording function is turned on, and the recording control 132b in state A1 can indicate that the recording function is not turned on.
  • the state of the recording control 132b in the aforementioned user interface 13 is state A1, and state B1 is another state different from state A1.
  • the recording control 132b in the user interface 13b is grayed out, that is, the grayed-out state is a state B1.
  • the electronic device can turn off the recording function, and update the state of the recording control 132b from state B1 to state A1.
  • the electronic device may divide the user interface into area 1 and area 2, where it can be understood that area 2 is a new editing area.
  • the area 1 can be divided into an area 11 and an area 12 .
  • the area 11 may include a recording bar in state 1 (also referred to as recording bar 1), and the recording bar 1 may include a pause recording control 136a, a recording status bar 136b, and a recording duration 136c corresponding to the recording of the voice, and the recording duration 136c may be the total duration of all the recorded voices.
  • the pause recording control 136a is used to indicate that the current terminal has started recording.
  • the recording status bar 136b can reflect the energy of the sound signal collected by the electronic device. The greater the energy, the greater the change in the recording status bar 136b. The smaller the energy, the smaller the degree of change in the recording status bar 136b.
  • the recording duration 136c can be used to indicate the total duration of all the recorded voices.
  • the area 12 may include a conversion control 137a in state C, and the conversion control 137a in state C may be represented by a solid line.
  • the conversion control 137a in state C may indicate that the electronic device has turned on the speech-to-text function.
  • Area 12 does not display the text information obtained by speech-to-text conversion, and when the speech-to-text function is turned on, the electronic device may also display a prompt message 137b, which may be used to prompt that the electronic device has turned on the speech-to-text function.
  • the prompt message 137b may be a prompt text: "Speech-to-text conversion in progress".
  • the electronic device may process the collected sound signal to obtain speech, and convert the speech, and display the converted text in area 12.
  • the recording duration 136c may not be used to indicate the total duration of all recorded speech, but may be used to indicate the total duration of the segment of speech being recorded.
  • Area 2 is used to present input non-recording information, such as text, pictures, etc.
  • a prompt message 138a may be displayed, and the prompt message 138a may be used to remind the user that recording is in progress and information may be input in area 2 at the same time.
  • the prompt message 138a may be a text prompt: "Recording is in progress, you may write or doodle in the blank space.”
  • the electronic device can also perform voice-to-text conversion while recording, and display the converted text in area 12 for the user to view.
  • the user can also perform manual input through an input tool, and display the input information in area 2 for the user to view.
  • Manual input includes inputting text, inserting pictures, etc.
  • the electronic device can collect sound signals from the user or the environment, and convert the sound signal into speech, the content of which is: "The note includes two parts".
  • the speech can be converted into text and displayed in area 12, and non-voice information input by the user can be received.
  • the non-voice information can be input into area 2 by the user through the keyboard 131 or the tool in the stylus toolbar 134 for display.
  • the non-voice information can include text information input into area 2 and information 139, which is: "Content".
  • the electronic device can also be set to turn off the voice-to-text function. In this way, after starting recording, the recorded voice can be not converted into text information, but non-recording information input by the user can be accepted during recording.
  • the electronic device in response to an operation (e.g., a click operation) on the conversion control 137a (state C), the electronic device can turn off the speech-to-text function, and update the state of the conversion control 137a from state C to state D, which is different from state C.
  • the prompt information 137b is displayed in the area 12, it indicates that the electronic device has not obtained the speech-to-text content.
  • the electronic device in response to an operation (e.g., a click operation) on the conversion control 137 (state C), the electronic device may no longer display the prompt information 137b, but display the prompt information 141.
  • the prompt information 141 is used to notify the user that the speech-to-text function is turned off.
  • user interface 14a is an exemplary user interface involved in turning off the voice-to-text function.
  • a prompt message 141 may be displayed in area 12 to indicate that the electronic device has turned off the voice-to-text function.
  • the prompt content of the prompt message may be: "Voice-to-text has been paused.”
  • the conversion control 137a (state D) may be represented by a dotted line.
  • state D may be represented in other forms besides being represented by a solid line, such as changing color, graying out, etc., which is not limited in the embodiments of the present application.
  • the electronic device may no longer display area 12. At this time, area 1 may only include area 11.
  • the electronic device in response to an operation on the pause recording control 136a, the electronic device may pause recording and display user interface 14b shown in (3) in FIG. 3.
  • State 2 is the state in which the recording function is turned on but the recording is paused.
  • user interface 14b is an exemplary user interface involved after pausing recording.
  • area 1 only includes recording bar 2.
  • the recording bar 2 may include a start recording control 141a, a start playing control 141b, a play progress bar 141c, a play time indicator 141d, and a voice menu control 141e.
  • the start recording control 141a can be used to trigger the electronic device to start recording again.
  • the start playing control 141b can be used to trigger the electronic device to play the recorded voice.
  • the play progress bar 141c can be used to indicate the progress of the currently played voice.
  • the play time indicator 141d is used to indicate the current play time and the total duration of all voices.
  • the voice menu control 141e is used to trigger the electronic device to display and manage all recorded voices, which include all voices recorded in note 1.
  • all the voices recorded in note 1 are Q (an integer greater than or equal to 2) segments of voice
  • the electronic device can integrate the Q segments of voice into a long voice, and the total duration of the Q segments of voice is the total duration of the long voice.
  • the play progress bar 141c may also include Q-1 nodes, dividing the play progress bar 141c into Q segments (which may be referred to as Q segment sub-progress bars).
  • the i-th segment sub-progress bar in the Q segment sub-progress bar corresponds to the i-th segment of speech.
  • the play progress bar 141c can be used to indicate the progress of the current voice playback: the play progress bar 141c can include an identifier 1411, and the identifier 1411 is at position A1 of the play progress bar 141c, indicating that the voice corresponding to the playback position A1 is played.
  • a possible display method includes: after displaying the user interface 14a shown in FIG. 3 (2), the user does not operate the control 136a in the user interface 14a, so the electronic device continues to record and displays the user interface 14c shown in FIG. 3 (4).
  • the starting recording duration corresponding to the i-th segment of recording may be the ending recording duration of the i-1th segment of speech, or the ending recording duration of the i-1th segment of speech + 1.
  • the time displayed corresponding to the recording duration 136c when the electronic device completes recording the Q-th segment of speech is the total duration of the Q segments of speech.
  • the ending recording duration corresponding to the speech recorded by the electronic device is 2s.
  • the starting recording duration corresponding to the i-th segment of recording may start from 0s.
  • the recording duration 136c may not be used to indicate the total duration of all recorded speech. It may be used to indicate the total duration of the segment of speech being recorded.
  • the playback progress bar 141c does not include a node. For example, as shown in the user interface 14b of (3) in FIG3 : Since only the first segment of voice has been recorded at this time, the playback progress bar 141c does not include a node at this time.
  • the recording bar 2 shown in (3) in Figure 3 is only an exemplary illustration.
  • the recording bar 2 may also include more or fewer components, and the shapes of the components may be the same as or different from those shown in the figure.
  • the embodiments of the present application are not limited to this.
  • the electronic device may be triggered to record a new voice, and the area 1 may be divided into area 11 and area 12. At this time, the electronic device may display a user interface as shown in FIG. 3 (4).
  • an operation eg, a click operation
  • user interface 14c is an exemplary user interface involved in recording a new speech by an electronic device.
  • User interface 14c may include area 1, which includes two sub-areas (area 11 and area 12).
  • area 11 may include a recording bar (recording bar 1) in state 1, which is used to prompt the user that the current electronic device is recording a speech.
  • the speech-to-text function of the electronic device is turned off before the operation of the start recording control 141a is detected, then after the operation of the start recording control 141a is detected, the speech-to-text function of the electronic device is still turned off.
  • the conversion control in area 12 is in a dotted state (a state D), indicating that the speech-to-text function is turned off.
  • the electronic device can also display a prompt message 141 to prompt the user that the speech-to-text function is turned off.
  • an operation such as a click operation
  • the electronic device can turn on the speech-to-text function.
  • the size of the speech-to-text area (area 12) can be adjusted, and the size of the area 12 can be switched between size 1 and size 2, wherein size 1 is smaller than size 2.
  • the speech-to-text area supports sliding display.
  • the text obtained by the latest recorded speech conversion can be displayed at the end of all the text in area 12.
  • the maximum number of complete characters that can be displayed in the speech-to-text area is W.
  • W characters are displayed in the speech-to-text area, for every E new characters added, E original characters (already displayed) can be not displayed or cannot be fully displayed in the speech-to-text area.
  • W is an integer greater than or equal to 1
  • E is an integer less than or equal to W.
  • the speech-to-text content here may include the text obtained by speech-to-text conversion and punctuation added by the electronic device.
  • the electronic device can record a voice.
  • the recorded voice content can be: “The first part mainly talks about XXX”.
  • the electronic device can display the text converted from the voice in the voice-to-text area (area 12).
  • the user interface 15a displays a size of 1.
  • An exemplary area 12 is shown.
  • a row of characters can be displayed completely in the area 12, and a row of characters can be displayed incompletely.
  • the electronic device can add a mask to the incompletely displayed row of characters, so that the incompletely displayed row of characters is "looming".
  • a row of characters can always be displayed completely in the area 12, and every time E new characters are added, they can be displayed completely in the area 12, and the front E original characters (already displayed) are slid up one row for incomplete display.
  • the area 12 of size 1 may also include an expansion control 151a and a sliding control 151b.
  • the expansion control 151a may be used to switch the size of the area 12 from size 1 to a larger size 2.
  • the sliding control 151b may be used to indicate that the area 12 supports sliding display.
  • the electronic device may expand the area 12 and change the size to size 2. At this time, the electronic device may display a user interface as shown in (2) in FIG. 4 .
  • user interface 15b is an exemplary user interface after expanding area 12.
  • Area 12 of size 2 may include a collapse control 151c.
  • the collapse control 151c may be used to trigger changing the size of area 12 from size 2 to size 1.
  • the size of area 12 when the size of area 12 is size 2, the size of area 12 can be set to 1/3 of the size of the mobile phone screen, or other sizes, which is not limited in the embodiments of the present application.
  • the electronic device continues to collect sound signals, the content of which is: "The second part mainly talks about XXXX", and then obtains voice through the sound signal, and converts the voice into text (text 1) and displays it in area 12 (size 2). Subsequently, the electronic device can continue to convert the recorded voice into text and display it in area 12, for example, referring to the user interface 15c shown in (3) of FIG. 4 .
  • the electronic device continues to collect a sound signal, the content of which is: "This time, let's explain the first part first.” Then, a voice is obtained through the sound signal, and the voice is converted into text (text 2) and displayed in area 12 (size 2). At this time, text 2 is displayed behind the aforementioned text 1. At this time, after text 2 is displayed in area 12, compared to before text 2 was displayed, a new line of text is added to area 12, and the position of the sliding control 151b can be moved downward. The position of the sliding control 151b can correspond to the position of the newly added characters: the further back the newly added characters are in area 12, the lower the sliding control is.
  • the text displayed in area 12 is the text obtained by converting the voice with a later recording time, and area 12 can support sliding display to display the text obtained by converting the voice with an earlier recording time.
  • the electronic device continues to collect sound signals and converts the sound signals into voice to display text in area 12.
  • the content of the sound signal is: "I will explain in detail later.”
  • the electronic device can display the text obtained by converting the voice with an earlier recording time.
  • the text obtained by converting the voice with an earlier recording time may not be displayed in area 12.
  • the electronic device can display the user interface 15e shown in (2) in Figure 5.
  • the way to display the text converted from the voice recorded earlier in area 12 may be other operations besides operation 1 for area 12 shown in (1) in FIG. 5 , for example, other operations may include: dragging the sliding control 151b upwards. This embodiment of the application is not limited to this.
  • the electronic device can continue to display the latest recorded voice in area 12.
  • the text is obtained by conversion.
  • the user interface displayed by the electronic device can be updated from the user interface 15e to the user interface 15f shown in (3) of FIG. 5 .
  • the electronic device can continue to collect sound signals, obtain speech based on the sound signals, and can continue to display the text converted from the latest recorded speech in area 12.
  • the content of the sound signal here can be: "Content of the second part”.
  • characters may be displayed incompletely in region 12, and a mask may be added to the incompletely displayed characters.
  • a mask may be added to the incompletely displayed characters.
  • only characters that can be displayed completely may be displayed in region 12, and no mask may be added. This may be reflected in region 12 in subsequent figures, and will not be described in detail here.
  • the speech-to-text area may support other functions in addition to supporting sliding display of text.
  • the text in area 12 may be processed, such as editing (including copying, pasting, etc.) and sharing.
  • the text information 152 shown in (1) of FIG. 5 is obtained by copying and pasting the text in area 12.
  • the speech-to-text area can be understood as an editable area, and the text therein can be edited, and the editing involved here includes but is not limited to one or more of the following: copy, cut, select all (optional), insert text, etc.
  • inserting text includes pasting text into area 12 or inputting text through the keyboard and then displaying it in area 12.
  • the electronic device in response to an operation 2 (e.g., a long press operation) on position 1 in area 12, the electronic device can select U characters closest to position 1, and can display a processing toolbar, which can include function controls involved in processing the selected characters. At this time, the electronic device can display a user interface as shown in (2) of FIG6 below.
  • an operation 2 e.g., a long press operation
  • the electronic device can add a check mark 162 to the selected characters, and one end of the check mark 162 includes a control 163, and the other end includes a control 164.
  • the control 163 and the control 164 can be used to trigger the adjustment of the selection area corresponding to the check mark 162, that is, the control 163 and the control 164 can be used to set the number and position of the selected characters.
  • the user interface 16b can also include a processing toolbar 161, and the processing toolbar 161 can include a select all control 161a, a cut control 161b, a copy control 161c, a paste control 161d, and a share control 161e.
  • the select all control 161a can be used to trigger the selection of all characters displayed in the area 12, and display a new toolbar including the cut control 161b, the copy control 161c, the paste control 161d, and the share control 161e.
  • the cut control 161b can be used to trigger the copying of the selected characters, and delete them after copying.
  • the copy control 161c can be used to trigger copying of the selected characters.
  • the paste control 161d can be used to trigger replacing the selected characters with other characters (copied).
  • control 164 In response to an operation on control 164, such as sliding control 164 to the right to position A, the electronic device may display control 164 at position A, and the characters included in the selection mark 162 are all characters between control 163 and control 164 (displayed at position A). At this time, the electronic device may display user interface 16c shown in (3) of FIG. 6 .
  • the characters included in the selected mark 162 are updated to: "Note 1 includes two parts, wherein the first part mainly describes XXXXX.”, which is the text information involved above. 152.
  • the position of the control 163 has not changed, and the position of the control 164 is updated from the position in the user interface 16b to the position shown in the user interface 16c (which can be regarded as a case of position A).
  • the electronic device may copy the selected character 1.
  • the selected character 1 may be subsequently pasted into the area 12 or the area 2.
  • an operation e.g, a click operation
  • the electronic device may copy the selected character 1.
  • the selected character 1 may be subsequently pasted into the area 12 or the area 2.
  • reference may be made to the description of the following content.
  • the electronic device may display the user interface 16e shown in (1) of FIG7 .
  • the user interface 16e includes a paste control 166, which may be triggered to paste the copied characters (e.g., the selected characters 1) into area 2 for display.
  • the electronic device can paste the copied characters (e.g., the selected characters 1) into the area 2.
  • the electronic device can display the user interface 16f as shown in (2) of Figure 7.
  • the text information 152 in the user interface 16f is the pasted characters.
  • the electronic device may turn off the speech-to-text function and change the state of the conversion control 137a from state C (e.g., represented by a solid line) to state D (e.g., represented by a dotted line).
  • state C e.g., represented by a solid line
  • state D e.g., represented by a dotted line
  • the prompt information 141 may no longer be displayed in the area 12.
  • the electronic device may end the first recording after note 1 is created.
  • the electronic device may display area 12 in size 1.
  • Area 12 may also be displayed in size 2, for example, refer to user interface 16h shown in (4) of FIG. 7.
  • the present application embodiment does not limit this.
  • the electronic device in response to an operation (e.g., a click operation) on the pause recording control 136a, the electronic device can end the recording and also end the voice-to-text conversion. That is, regardless of whether the voice-to-text function is available before the recording is paused, the voice-to-text function is not available when the recording is paused.
  • the electronic device in response to an operation (e.g., a click operation) on the pause recording control 136a, the electronic device can also set the state of the conversion control 137a to state D (e.g., indicated by a dotted line) to indicate that the voice-to-text function is not available.
  • the conversion control 137a may not receive the user's operation.
  • the electronic device can process the text in area 12, such as editing (including copying, pasting, etc.) and sharing.
  • the process of editing the text in area 12 can refer to the following description of the relevant content in FIG. 8 .
  • the electronic device can select some or all characters (selected characters 2) in the area 12, add a selection mark 162 to the selected characters, and display a processing toolbar 161 to process the selected characters.
  • the selected characters may include: "The second part mainly describes XXXXX.”
  • the process of selecting the mark 162, the processing toolbar 161, and the electronic device selecting characters can refer to the aforementioned description of the relevant contents of (1) and (2) of FIG6 , which will not be repeated here.
  • the electronic device may
  • the selected character 2 is copied.
  • the selected character 2 can be subsequently pasted into the area 12 or the area 2.
  • the electronic device may display a user interface 17c as shown in (3) of FIG. 8 .
  • the user interface 17c includes a paste control 166, which can be used to trigger pasting of the copied characters (selected characters 2) into area 2 for display.
  • the electronic device can paste the copied characters (selected characters 2) into the area 2.
  • the electronic device can display the user interface 17d as shown in (4) of Figure 8.
  • the text information 171 in the user interface 17d is the pasted characters.
  • the electronic device can perform a second recording.
  • the second recording please refer to the following description of Figs. 9 to 12.
  • the electronic device can make the first recording and the second recording into a long voice.
  • a node 1 can be displayed in the playback progress bar, and the node is used to divide the playback speed bar into two sub-progress bars, the sub-progress bar before node 1 corresponds to the voice obtained by the first recording, and the sub-progress bar after node 1 corresponds to the voice obtained by the second recording.
  • the electronic device can also perform a third recording, a fourth recording, and so on.
  • the number of recordings is recorded here as Q.
  • a node can be added to the playback progress bar.
  • the playback progress bar is divided into Q segments (which can be called Q segment progress bars).
  • the i-th sub-progress bar in the Q segment progress bar corresponds to the i-th segment of voice.
  • Q is an integer greater than or equal to 2.
  • the electronic device can input non-recording information in area 2.
  • the methods of inputting non-recording information include keyboard input and copy and paste.
  • Editing in area 2 can also be performed in other ways, including: inserting pictures into area 2, editing in area 2 through the stylus toolbar (such as graffiti, handwriting text input), etc.
  • the relevant content of editing in area 2 by other methods can be described in the second recording, and the details can be referred to the following description of Figures 9 to 12.
  • user interface 18a is an exemplary user interface involved in the second recording process.
  • the text displayed in area 12 is converted from the voice obtained in the first recording.
  • the electronic device displays the text converted from the voice obtained in the second recording in area 12, it can be displayed in different ways, including but not limited to the following display methods.
  • Display mode 1 The text converted from the speech obtained by the first recording and the text converted from the speech obtained by the second recording are displayed continuously. That is, assuming that after the first recording is completed and before the second recording begins, the last character included in area 12 is displayed in the hth row of all characters in area 12.
  • the electronic device can display the first text converted from the voice obtained by the second recording on the hth row and after the last character. In this case, there can be no space between the first text (obtained in the second recording) and the last character (obtained in the first recording).
  • the electronic device can display the first text converted from the voice obtained by the second recording on the next line of the last character.
  • an exemplary user interface involved in display mode 1 can refer to the following description of (2) in Figure 9.
  • the user interface 18b includes text information 181 and text information 182, wherein the text information 181 is converted from the voice obtained by the first recording.
  • the text information 182 is converted from the voice obtained by the second recording.
  • the content of the text information 182 may be: "The first part starts from the following figure”. At this time, the last character in the text information 181 and the first character in the text information 182 are displayed in the same line and are displayed continuously without a space in between.
  • Display mode 2 Display the text converted from the voice obtained by the first recording and the text converted from the voice obtained by the second recording separately. For example, assuming that after the first recording is completed and before the second recording begins, the last character included in area 12 is displayed on the hth row of all characters in area 12. When the last character is not the last character of the hth row, the electronic device can display the first text converted from the voice obtained by the second recording on the next row of the last character.
  • an exemplary user interface involved in display mode 2 can refer to the following description of (3) in Figure 9.
  • the user interface 18c also includes the aforementioned text information 181 (obtained from the first recording) and text information 182 (obtained from the second recording). At this time, the last character in text information 181 and the first character in text information 182 are displayed on different rows: the first character in text information 182 is displayed on the next row of the last character in text information 181.
  • display mode 2 is that when the last character is the last character of the h-th line, the electronic device can generate a blank line (the blank line may not be used to display any characters), and then display the first text converted from the voice obtained from the second recording on the line below the blank line, or the electronic device can leave two spaces on the line below the h-th line, and then display the first text converted from the voice of the second recording.
  • the electronic device can generate a blank line (the blank line may not be used to display any characters), and then display the first text converted from the voice obtained from the second recording on the line below the blank line, or the electronic device can leave two spaces on the line below the h-th line, and then display the first text converted from the voice of the second recording.
  • the following content takes display mode 1 as an example to describe other contents involved in the embodiment of the present application. For example, the process of inserting a picture in area 2 is described.
  • the electronic device may display at least one insert mode identifier or at least one insertable content identifier. At this time, the electronic device may display the user interface 18d as shown in (1) of Fig. 10.
  • the user interface 18d may include a plurality of insertion method identifiers and a plurality of insertable content identifiers.
  • the insertable content identifiers may include a shooting icon 183a, a table icon 183c, a schedule icon 183d, a card scan icon 183e, a mind map icon 183g, and a recording icon 183h.
  • the insertion method identifier may include a “select from gallery” icon 183b and a “file scan” icon 183f.
  • the shooting icon 183a may be used to trigger opening the camera, taking a picture, and then inserting the taken picture into area 2.
  • the table icon 183c may be used to insert a table into area 2.
  • the schedule icon 183d may be used to insert a schedule identifier into area 2.
  • the “card scan” icon 183e may be used to trigger opening the scanning function, scanning the card, and displaying the scanned content in area 2.
  • the mind map icon 183g may be used to insert a mind map into area 2.
  • the recording icon 183h may be used to insert a recording into area 2.
  • the "Select from Gallery” icon 183b can be used to trigger opening the gallery, and then select a picture or video from the gallery to insert into area 2.
  • the "File Scan” icon 183f can be used to trigger opening the scanning function, scanning the file, and displaying the scanned content in area 2.
  • the insertion mode identifier and the insertable content identifier shown in the user interface 18d are only exemplary, and may actually include other identifiers, or may include more or fewer identifiers than those shown in the figure.
  • the identifier may also be in other forms, such as text, etc. This embodiment of the application does not limit this.
  • the electronic device can open the gallery, select a picture or video from the gallery and insert it into area 2, and display a user interface as shown in (2) in FIG10. At this time, the electronic device can also collect sound signals to obtain speech, and convert the speech into text and display it in area 11.
  • the content of the sound signal can be: "In the picture”.
  • user interface 18e is an exemplary interface after opening the gallery.
  • the electronic device can insert the picture 1 into area 2.
  • the electronic device can display user interface 18f as shown in (3) of FIG10 .
  • picture 1 may be displayed in area 2 to prompt the user that picture 1 has been inserted.
  • the electronic device may also collect sound signals to obtain voice, and convert the voice into text to be displayed in area 11.
  • the content of the sound signal may be: "recorded”.
  • the electronic device may switch the input tool from the keyboard to the stylus toolbar. Editing (e.g., graffiti, handwriting, etc.) may be performed in area 2 through the stylus toolbar.
  • Editing e.g., graffiti, handwriting, etc.
  • Part of the process involved in the second recording and the process of editing (eg, graffiti) in area 2 through the stylus toolbar can refer to the following description of FIG. 11 and FIG. 12 .
  • the electronic device in response to an operation (e.g., a click operation) on the stylus 134b in the stylus toolbar 134, the electronic device can trigger editing (e.g., graffiti) in area 2 by the stylus 134b, and the electronic device can highlight the stylus 134b or change the state of the stylus 134b to indicate that the stylus 134b is selected. At this time, the electronic device can display the user interface shown in (2) of FIG. 11 .
  • editing e.g., graffiti
  • the stylus 134b is highlighted to indicate that the stylus 134b is selected.
  • the electronic device may display a user interface as shown in (3) of Figure 11.
  • the user interface 19c may include an underline 191, which is added in the area 2 using the stylus 134b.
  • the stylus toolbar may include at least one stylus, and different stylus pens may be of different types, for example, marker pen 1, marker pen 2, eraser pen, pencil, etc.
  • different results are obtained.
  • different stylus pens draw different types of lines when editing.
  • the different types of lines may include different colors and shapes.
  • marker pen 1 can draw lines of shape 1 (e.g., straight lines)
  • marker pen 2 can draw lines of shape 2 (e.g., curves), etc.
  • the lines edited by different stylus pens can be adjusted, for example, the color and thickness of the lines can be adjusted.
  • the electronic device may switch the input tool from the stylus toolbar to the keyboard. Subsequently, the electronic device continues to input in area 2 while performing the second recording. Enter non-recording information, and then end the second recording.
  • an operation e.g, a click operation
  • the electronic device may switch the input tool from the stylus toolbar to the keyboard. Subsequently, the electronic device continues to input in area 2 while performing the second recording. Enter non-recording information, and then end the second recording.
  • the user interface 20a includes a picture 2. It is assumed here that picture 2 is displayed after picture 1. For the process of inserting picture 2, reference can be made to the aforementioned description of inserting picture 1, which will not be repeated here.
  • the electronic device can continue to collect sound signals to obtain speech, and the content of the sound signal can be: "In addition, the first part XX" While recording, the electronic device can perform speech-to-text conversion and display the converted text in area 12. Moreover, while recording, the electronic device can also input non-voice information in area 2. For example, in response to an operation (such as a click operation) for editing the text "picture", the electronic device can input the edited text "picture" into area 2 for display. At this time, the electronic device can display a user interface 20b as shown in (2) of FIG. 12 .
  • the electronic device can then end the second recording.
  • the electronic device in response to an operation (such as a click operation) on the pause recording control 136a, the electronic device can end the second recording.
  • the duration of the voice obtained in the second recording is 12 seconds
  • the recording duration is 23 seconds, which can be regarded as the sum of the duration of voice 1 (11 seconds) and the duration of voice 2 (12 seconds).
  • the electronic device can display a user interface as shown in (4) of FIG. 12 .
  • the user interface 20d may include a recording bar (recording bar 2) in state 2.
  • the electronic device may make the voice obtained in the first recording (voice 1) and the voice obtained in the second recording (voice 2) into a long voice (long voice 1).
  • the duration of long voice 1 may be expressed as the duration of voice 1 plus the duration of voice 2.
  • the duration of long voice 1 is 23s, which can be regarded as the sum of the duration of voice 1 (11s) and the duration of voice 2 (12s).
  • the user interface 20d may also include a play progress bar 141c, which may include a node 201 (which may be regarded as a node 1).
  • the node 201 may divide the play progress bar 141c into two sub-progress bars.
  • the sub-progress bar before the node 201 corresponds to the voice (voice 1) obtained by the first recording
  • the sub-progress bar after the node 201 corresponds to the voice (voice 2) obtained by the second recording.
  • the electronic device can save the note 1 recording the information. After saving, the user can view the note 1 through the electronic device.
  • the electronic device in response to an operation (eg, a click operation) on the completion control 401 , the electronic device can save the note 1 .
  • an operation eg, a click operation
  • the electronic device may also display a prompt message to prompt the user that the saving is successful.
  • the user interface 41 may include a prompt message 411, and the prompt message 411 is used to prompt the user that note 1 is saved successfully.
  • FIGS. 14-17 An exemplary set of user interfaces involved in viewing information is described in FIGS. 14-17 .
  • the user Before and after saving Note 1, the user can view the information recorded in Note 1 through the electronic device.
  • the electronic device can play the voice to allow the user to view the recorded content.
  • the electronic device plays the voice, the audio and text can be synchronized.
  • audio-text synchronization may include one or more of the following:
  • Synchronous mode 1 When playing a recording, the non-recorded information corresponding to the played voice in area 2 (editing area) is displayed in state A (recorded as information A), and the non-recorded information corresponding to the unplayed voice in area 2 is displayed in state B (recorded as information B).
  • information A and information B are described by taking the electronic device playing voice A (recording time is time A) as an example: information A refers to information whose input time is earlier than or equal to time A.
  • Information B refers to information whose input time is later than time A. Among them, time A can also be called the timestamp corresponding to voice A.
  • the input time of information A can be called the timestamp corresponding to information A.
  • the input time of information B can be called the timestamp corresponding to information B.
  • state A is different from state B.
  • state B can be to add a mask on information B, or to make information B semi-transparent, so that information B is "looming".
  • the information B has not been entered into the note.
  • State A can be to not add a mask on information A, to indicate that before recording the currently playing voice, the information A has been entered into the note.
  • Synchronous mode 2 When playing a recording, the text converted from the played voice can be displayed in area 12 (voice-to-text area).
  • the audio-text synchronization including the aforementioned synchronization mode 1 and synchronization mode 2 as an example.
  • the audio-text synchronization may only include synchronization mode 1 or synchronization mode 2, and the embodiment of the present application does not limit this.
  • the user interface 30 is an exemplary user interface involved in the case where the recording function is turned on and the recording is paused, and the user interface 30 may include a recording bar (recording bar 2) in state 2.
  • State 2 is the state where the recording function is turned on but the recording is paused.
  • the stylus toolbar 134 shown in (1) of FIG14 can be switched by touching the “handwriting” control 132c.
  • the electronic device can also switch the stylus toolbar 134 to the keyboard through the keyboard switching control 135.
  • the electronic device can start playing the recorded voice (long voice).
  • the electronic device can achieve audio-text synchronization.
  • the electronic device can display a user interface as shown in (2) in FIG. 14 below.
  • the electronic device when the voice corresponding to the 3rd second (voice A1) is played, the electronic device can achieve audio-text synchronization in the following two aspects.
  • text information 311 may be displayed in the speech-to-text area (area 12), and the text information 311 includes text converted when recording the speech A1.
  • the relevant contents involved in the electronic device recording the speech A1 can refer to the aforementioned description of the user interface 15a shown in (1) of FIG. 4, and will not be repeated here.
  • no mask is added to the non-recording information 312a in the editing area (area 2).
  • a mask is added to the non-recording information 312b. It can also be understood that: the area in area 2 displaying the non-recording information 312a has no mask added, and the area displaying the non-recording information 312b has a mask added.
  • the input time of the non-recording information 312a is earlier than or equal to the time of recording the voice A1.
  • the input time of the non-recording information 312b is later than the time of recording the voice A1.
  • the speech-to-text area (area 12) can synchronize the audio and text, taking the area 12 of size 1 as an example.
  • the speech-to-text area (area 12) of size 2 can also synchronize audio and text.
  • the electronic device in response to an operation (such as a click operation) on the expansion control 151a, the electronic device can display a speech-to-text area (area 12) of size 2.
  • the electronic device can display a user interface 32 as shown in (3) in Figure 14.
  • area 12 still includes text information 311. No mask is added to the non-recording information 312a in area 2.
  • a mask is added to the non-recording information other than the non-recording information 312a in area 2.
  • the electronic device can continue to achieve audio-text synchronization when playing voice.
  • the electronic device when the voice corresponding to the 8th second (voice A2) is played, the electronic device can display text information 313 in the voice-to-text area (area 12) of size 2, and the text information 313 includes the text converted when recording voice A2.
  • the relevant content involved in the electronic device when recording voice A2 can refer to the aforementioned description of the user interface 15d shown in (1) in Figure 5, which will not be repeated here.
  • the electronic device can also add a mask to the non-recording information 314b in the editing area (area 2) to distinguish it from the non-recording information 314a. It means that the input time of the non-recording information 314b is later than the time of recording voice A2, and the input time of the non-recording information 314a is earlier than or equal to the time of recording voice A2.
  • the electronic device in response to an operation (e.g., a click operation) on the start play control 141b in the recording strip 2, the electronic device can start playing the recorded voice (long voice). At this time, the electronic device can also replace the start play control 141b in the recording strip 2 with the pause play control 141f to obtain a recording strip (recording strip 3) in state 3, and the play control 141b can be used to trigger the pause play of the voice (long voice).
  • the recording strip 3 can also include other controls, such as a start recording control 141a, a play progress bar 141c, a play time indicator 141d, and a voice menu control 141e.
  • a start recording control 141a a play progress bar 141c
  • a play time indicator 141d a play time indicator 141d
  • voice menu control 141e for the relevant content of the other controls, please refer to the description of the relevant controls in the user interface 14b shown in (3) in the aforementioned Figure 3, which will not be repeated here.
  • the electronic device before the electronic device starts playing the audio, it can also display a prompt message to remind the user that the audio and text can be synchronized when playing the voice.
  • the electronic device can display a prompt box 301, which can be presented in the form of a bubble, pointing to the start play control 141b, and the content in the prompt box 301 can be: "Click to play the recording, and the content you recorded can be displayed when the recording is played.”
  • the content in the prompt box 301 can be used to prompt the user to click the start play control 141b to play the audio, and the audio and text can be synchronized when playing the audio.
  • the prompt box 301 can be displayed after pausing the recording, and the display number is 1.
  • the prompt box 301 will disappear when the user clicks anywhere on the screen.
  • the prompt box 301 can also have other display rules and disappearance rules, which are not limited in the embodiments of the present application.
  • the electronic device can also implement an advance preview.
  • the advance preview may include one or more of the following contents.
  • Preview method 1 When playing the voice recorded at time B, an operation A (such as a continuous long press operation) is detected for position T on the playback progress bar (the voice played at position T corresponds to the voice recorded at time C), and the electronic device can display the text converted from the voice recorded at time C (a certain time after time B) and part of the text converted from the voice recorded before time C in area 12, without affecting the voice recorded at time B. This allows the user to preview the text converted from the voice recorded at time C.
  • time C and time B can indicate a period of time or a moment, which is not limited in the embodiments of the present application.
  • the electronic device can highlight the non-recording information (non-recording information A) input at time C (a time after time B) in area 2, and does not affect the playback of the voice recorded at time B.
  • highlighting includes marking the non-recording information A in red, bolding, etc., one or more, which is not limited in the embodiments of the present application. Highlighting the non-recording information A can also be understood as displaying the non-recording information A in a different state (a state different from state A and state B).
  • operation A and operation B may be the same operation or different operations.
  • an example is used in which operation A and operation B are the same operation, and the electronic device can implement advance preview through preview mode 1 and preview mode 2.
  • operation A and operation B may be different, and the electronic device may implement advance preview through one of preview mode 1 and preview mode 2.
  • the voice currently played by the electronic device is the voice corresponding to the position 341 on the play progress bar 141c in the recording bar 3 (ie, the voice corresponding to the 8th second in the user interface 34).
  • the electronic device can change the content displayed in area 12 and area 2, and at this time, the electronic device can continue to play the voice corresponding to position 341. It is assumed here that the voice corresponding to position 342 is the voice corresponding to the 23rd second, and the recording time of the voice corresponding to the 23rd second is time C. Then the electronic device can display the text obtained by converting the voice recorded at time C in area 12, as well as part of the text obtained by converting the voice recorded before time C. For example, text information 344 can be included.
  • the content involved when the electronic device obtains the text information 344 can refer to the aforementioned description of (1) in Figure 12, which will not be repeated here.
  • the electronic device can also highlight the non-recording information 334 (a non-recording information A) input at time C in area 2.
  • the non-recording information 334 can be displayed in bold.
  • the content of the non-recording information 334 input by the electronic device into area 2 can refer to the aforementioned description of (1) in Figure 12, which will not be repeated here.
  • the electronic device when the electronic device detects that the user ends the long press operation at position 342, the electronic device can update the content included in area 12 and area 2 to the content displayed when the voice corresponding to position 341 is played. For example, referring to user interface 34, in response to the user ending the long press operation at position 342, the electronic device can redisplay the aforementioned user interface 33.
  • the audio-text synchronization function is turned on by default, and the electronic device can turn off the audio-text synchronization function.
  • An exemplary method of turning off the audio-text synchronization function can refer to the following descriptions of (2), (3) and (4) in Figure 15.
  • the electronic device in response to an operation (such as a long press operation) at any position in area 2, can display a control corresponding to turning off the audio-text synchronization function.
  • the control 351 is the control corresponding to turning off the audio-text synchronization function.
  • the control 351 can also include prompt information, which is used to remind the user that the control is used to trigger the turning off of the audio-text synchronization function.
  • the prompt information can be: "Stop the note playback animation", where the note playback animation is the aforementioned audio-text synchronization.
  • the electronic device After turning off the audio-text synchronization function, the electronic device no longer adds a mask in area 2 to distinguish whether the input time of the non-recording information is later than the recording time of the currently playing voice. For example, at this time, the electronic device can display the user interface 36 shown in (4) of FIG. 15 below.
  • the electronic device can also turn off the audio-text synchronization of all notes in other ways.
  • the electronic device can add a setting item 1 for all notes, and the setting item 1 can be used to control whether to turn on the audio-text synchronization.
  • the setting value corresponding to the setting item 1 is off (OFF)
  • the setting value corresponding to the setting item 1 is on (ON), it means that the audio-text synchronization of all notes is turned on, that is, the audio-text synchronization function can be used when entering any note.
  • the relevant content involved in this process please refer to the following description of Figure 16.
  • the user interface 11f may include a control 114, which may be used to trigger the electronic device to display more setting items related to the note.
  • a control 114 which may be used to trigger the electronic device to display more setting items related to the note.
  • the electronic device may display a setting item about the note. For example, at this time, the electronic device may display a user interface as shown in (2) in FIG. 16 below.
  • the user interface 37 may include a setting function item 374, which can be used to trigger the electronic device to set all notes. For example, the audio and text synchronization function for all notes is set (turned off or on).
  • the user interface 37 may also include other setting items, such as a grid view setting item 371, a batch deletion setting item 372, and a sorting method setting item 373.
  • the grid view setting item 371 can be used to set the display mode of all notes.
  • the batch deletion setting item 372 can be used to delete one or more notes at one time.
  • the sorting method setting item 373 can be used to set the sorting method of all notes, such as arranging from front to back according to the size of the notes.
  • the electronic device may display more setting items for all notes.
  • the electronic device may display a user interface as shown in (3) of FIG. 16 described below.
  • the user interface 38 may include a “note playback animation” setting item 381 (equivalent to a setting item 1).
  • the “note playback animation” setting item 381 is enabled by default, that is, the electronic device enables the audio-text synchronization function of all notes by default.
  • the electronic device may disable the audio-text synchronization function of all notes.
  • the user interface 42 includes at least one directory item corresponding to a note.
  • the electronic device can open the corresponding note and display the content recorded in the note. For example, in response to an operation (such as a click operation) on the directory item 421 corresponding to note 1, the electronic device can open note 1 and display the content recorded in note 1.
  • the electronic device when it opens the note 1 that records information, it can still divide the user interface into two areas: the recording area (the aforementioned area 1) and the editing area (the aforementioned area 2). Area 1 is used to present content related to recording, and area 2 is used to display non-recording information. After closing note 1, when you open note 1 again, the content included in area 1 is one of the following two situations:
  • the recording bar 4 may include a start playback control 141b, a playback progress bar 141c, a playback time indicator 141d and a voice menu control 141e.
  • a start playback control 141b For the description of the various controls included in the recording bar 4, reference can be made to the aforementioned description of the various controls included in the recording bar 2, which will not be repeated here.
  • Area 1 may include the recording area (Area 1), but may not include the speech-to-text area (Area 12).
  • Area 1 may include the recording area (Area 1), but may not include the speech-to-text area (Area 12).
  • Area 11 may include Area 1, and Area 11 includes the recording strip 4. At this time, Area 1 does not include the speech-to-text area (Area 12).
  • the display of recording strip 4 here is only an exemplary description.
  • the recording strip can also be in other states, for example, it can be a recording strip (recording strip 2) in the aforementioned state 2. This embodiment of the application does not limit this.
  • case 1 is used as an example for description.
  • the electronic device can start playing the recorded voice (long voice).
  • the electronic device can display the user interface 45 as described in (4) of FIG. 17 below.
  • the user interface 45 may include a recording bar (recording bar 5) in state 5, where the recording function is turned off and the voice is played.
  • the recording bar 5 may include a pause control 141f, a play progress bar 141c, a play time indicator 141d, and a voice menu control 141e.
  • the description of each control included in the recording bar 5 can refer to the description of each control included in the recording bar 3, which will not be repeated here.
  • recording bar 5 here is only an exemplary description.
  • the recording bar can also be in other states, for example, it can be a recording bar (recording bar 3) in the aforementioned state 3. This embodiment of the application is not limited to this.
  • Management information includes deletion information, sharing information, and saving information.
  • Deletion information includes the option to delete part or all of the voice information recorded in Note 1.
  • Sharing information includes transmitting the content to be shared in the note to other terminals, and the sharing method can be selected. The content to be shared can be selected by the user through an electronic device. For example, you can choose to share other information in Note 1 except voice, you can also choose to share other information in Note 1 except voice and text converted from voice, you can also share all the information in Note 1, etc.
  • Figures 18-29 describe the relevant content involved when an electronic device manages information.
  • Figures 18 to 24 are a set of exemplary user interfaces involved in deleting information on an electronic device.
  • Figures 25 to 29 are a set of exemplary user interfaces involved in sharing information on an electronic device.
  • Deleting information includes deleting voice information.
  • the electronic device can also determine whether to delete the text information corresponding to the selected voice information according to user needs.
  • the text information corresponding to the voice information is obtained by converting the voice.
  • Note 1 can record Q voice segments, and the electronic device can choose to delete L voice segments among the Q voice segments, and select Choose whether to delete the text information corresponding to the L segments of speech. Where L is an integer less than or equal to Q.
  • the electronic device may provide a function of deleting all voices.
  • the electronic device may also determine whether to delete text information corresponding to all voices according to user requirements.
  • FIG. 18 depicts a set of exemplary user interfaces involved when an electronic device deletes all speech.
  • the electronic device when voice playback is paused, in response to operation C (such as a long press operation) on the playback progress bar 141c, the electronic device can display a management toolbar.
  • the management toolbar includes at least one control related to managing note 1, and different controls can be used to trigger different processing of note 1.
  • the management toolbar 511 may include a delete control 511a, a save control 511b, and a share control 511c.
  • the delete control 511a can be used to trigger the electronic device to delete all voice information.
  • the save control 511b can be used to trigger the electronic device to save all recorded voices.
  • the share control 511c can be used to trigger the electronic device to share all voice information.
  • the electronic device in response to an operation (e.g., a click operation) on the delete control 511a, the electronic device may display a prompt box A involved in deleting all voices, and the prompt box A may be used to prompt the user that the content to be deleted is all voices, and may also be used to receive a user input to confirm the operation of deleting all voices or cancel the operation of deleting all voices.
  • the prompt box 521 displayed in the user interface 52 may be an exemplary prompt box A.
  • the prompt box 521 may include prompt information 521a, and the prompt information 521a may be used to prompt the user whether to delete all recordings (voices).
  • the prompt box 521 may also include a selection control 521b, which can be used to receive the user's operation of whether to retain the speech-to-text content (the text information corresponding to all recordings).
  • the speech-to-text content is the text information corresponding to all voices.
  • the speech-to-text content is selected by default, and the selection control 521b is in state E, for example, state E is grayed out.
  • state E the state of the selection control 521b is state E, it means that the speech-to-text content is selected to be retained.
  • the electronic device can change the state of the selection control 521b to state F, for example, the state F is not grayed out.
  • state F When the state of the selection control 521b is state F, it means that the speech-to-text content is not selected to be retained. When all voices are deleted, the speech-to-text content will also be deleted.
  • the state of the selection control 521b can be switched between state E and state F to determine whether the speech-to-text content corresponding to all voices is deleted when all voices are deleted.
  • the prompt box 521 may also include a confirm delete control 521c and a cancel control 521d.
  • the confirm delete control 521c may be used to trigger the deletion of all voices in Note 1. If the user chooses to delete (not retain) the voice-to-text content corresponding to all voices, the confirm delete control 521c may also be used to trigger the deletion of the text information (voice-to-text content) corresponding to all voices.
  • the cancel control 521d is used to cancel the deletion of all voices and to close the prompt box 521.
  • the electronic device may also choose to delete part or all of the voice.
  • deleting part or all of the voice it may also be set whether to delete the text information corresponding to the part or all of the voice.
  • the electronic device can delete all the voice-related contents by referring to the relevant descriptions of Figures 19-21.
  • the electronic device may display a display box A including Q segments of voice information.
  • the display box A is used to display information corresponding to the Q segments of recording (e.g., recording duration, etc.).
  • the display box A may also include controls involved in managing the recorded voice.
  • the electronic device may display the user interface shown in (2) of FIG. 19 .
  • the display box 535 included in the user interface 53a can be regarded as an exemplary display box A.
  • the display box 535 can include Q segments of recorded voice information: including the name (recording 001) corresponding to the first recorded voice and its corresponding total duration (11s), and the name (recording 002) corresponding to the second recorded voice and its corresponding total duration (12s).
  • the display box 535 can also include an editing control 535a.
  • the editing control 535a can be used to trigger the display of related interfaces involved in processing (editing) the recorded voice.
  • the electronic device can display more controls involved in managing (editing) the voice in the note. It can also trigger the display of information corresponding to all the voices recorded in Note 1, which can include the total duration of each voice, the name of the voice, etc. At this time, the electronic device can display the user interface shown in (3) in Figure 19.
  • user interface 53b is an exemplary user interface involved in managing (editing) a voice.
  • the user interface 53b may include a sharing control 531, a rename control 532, a delete control 533, and a deselect all control 534.
  • the sharing control 531 may be used to trigger the selected voice to be shared with other devices.
  • the rename control 532 may be used to trigger the name of the voice to be changed.
  • the delete control 533 may be used to trigger the selected voice.
  • the electronic device may be set to select all voices for processing by default, and the deselect all control 534 may be used to trigger the update of all voices from selected to unselected.
  • the electronic device selects all voices by default, for example, the selected control 530a corresponding to recording 001 (the name corresponding to the voice obtained by the first recording) is in state E, for example, a grayed-out state.
  • the grayed-out state of the selected control 530a indicates that the voice obtained by the first recording is the selected voice.
  • the selected control 530b corresponding to recording 002 (the name corresponding to the voice obtained by the second recording) is also in state E, for example, a grayed-out state. It indicates that the voice obtained by the first recording is also the selected voice.
  • the electronic device can display the aforementioned prompt box A involved in deleting all voices, which can be used to prompt the user that the deleted content is all voices, and can also be used to receive the user input of the operation of deleting all voices or canceling the operation of deleting all voices.
  • the delete control 533 and the delete control 511a can be regarded as controls with the same function but different forms.
  • the electronic device can display the user interface 54 shown in (1) in Figure 20.
  • the electronic device When deleting all the voices, the electronic device retains the text information corresponding to all the voices.
  • the relevant content involved please refer to the following description of Figure 20.
  • the user interface 54 is an exemplary user interface involved in managing speech.
  • the user interface 54 may include a prompt box 521, which may be an exemplary prompt box A.
  • a prompt box 521 For the description of the prompt box 521, reference may be made to the content involved in (3) of FIG. 18 , which will not be repeated here.
  • the selected control 521b in the prompt box 521 is in state E, indicating that the speech-to-text content is selected to be retained (not deleted).
  • the electronic device can delete all the voices and retain (not delete) the text information corresponding to all the voices. That is, the electronic device can retain the text information corresponding to all the voices in Note 1, but delete all the voices in Note 1.
  • the way in which the electronic device retains the text information corresponding to all the voices in Note 1 includes but is not limited to one of the following two ways.
  • the electronic device may not display the recording area (area 1), and insert the text information corresponding to all voices into the front of the non-recording information, or into the back of the non-recording information.
  • area 1 For an exemplary user interface involved when text information is inserted in front of non-recorded information, reference may be made to the user interface 55a shown in (2) of FIG. 20 .
  • Retention mode 2 The electronic device may not display the recording area (area 1), and a new area (area 3) supporting sliding display is set in note 1 to display the text information corresponding to all the voices.
  • the user interface 55b may include area 3, and part of the text information corresponding to all the voices may be displayed in area 3.
  • the electronic device may display the other part of the text information corresponding to all the voices.
  • the electronic device When deleting all the voices, the electronic device also deletes the text information corresponding to all the voices. For the related contents involved, please refer to the following description of FIG. 21.
  • user interface 54b is an exemplary user interface involved in managing voice.
  • the user interface 54b may include a prompt box 521, which may be an exemplary prompt box A.
  • a prompt box 521 For the description of the prompt box 521, reference may be made to the content involved in (3) of FIG. 18 , which will not be repeated here.
  • the selected control 521b in the prompt box 521 is in state F, indicating that the voice-to-text content is selected to be deleted.
  • the electronic device can delete all the voices in Note 1, and delete the text information corresponding to all the voices. After the deletion, Note 1 no longer includes the recorded voice and the text information corresponding to the voice, and can include non-recorded information. For example, at this time, the electronic device can display the user interface 55c shown in (2) in Figure 21.
  • the electronic device may change the selected controls corresponding to all voices (e.g., selected controls 530a and selected controls 530b) from state E (e.g., grayed out) to state F (not grayed out) to indicate that all voices are set from selected to unselected.
  • the electronic device may display a user interface as shown in (2) of FIG. 22 below.
  • the description of the user interface 53b reference may be made to the description of (3) of FIG. 19, which will not be repeated here.
  • the selected controls corresponding to all voices in the user interface 56 are not grayed out.
  • the electronic device may gray out the control 530b to indicate that recording 002 (the voice obtained during the second recording) is selected.
  • the electronic device may display the user interface 57 shown in (3) of FIG. 22 described below.
  • the selected voice includes recording 002 (the voice obtained during the second recording) but does not include recording 001 (the voice obtained during the first recording).
  • the electronic device can display a prompt box B involved in deleting part of the voice (the selected voice), and the prompt box B can be used to prompt the user that the content to be deleted is the selected voice, and can also be used to receive the user's input to confirm the deletion of the selected voice or cancel the deletion of the selected voice.
  • the prompt box 581 displayed in the user interface 58 may be an exemplary prompt box B.
  • the prompt box 581 may also include prompt information 581a, and the prompt information 581a may be used to prompt the user whether to delete the selected recording (voice).
  • the prompt box 581 may also include a selection control 581b, which may be used to receive a user's operation of whether to retain the speech-to-text content (select the text information corresponding to the recording).
  • the speech-to-text content is retained in the selected control 581b.
  • the selected control 581b is in state E, for example, state E is grayed out.
  • state E it means that when the selected voice is deleted, the text information corresponding to the selected voice is retained (not deleted).
  • the electronic device can change the state of the selected control 581b to state F, for example, state F is not grayed out.
  • state F When the state of the selected control 581b is state F, it means that when the selected voice is deleted, the text information corresponding to the selected voice will also be deleted.
  • the state of the selected control 581b can be switched between state E and state F to determine whether to delete the speech-to-text content corresponding to the selected voice when the selected voice is deleted.
  • the state of the selected control 581b is not grayed out, indicating that the text information corresponding to the selected voice is deleted when the selected voice is deleted.
  • the selected voice includes recording 002 (the voice obtained during the second recording), but does not include recording 001 (the voice obtained during the first recording).
  • the electronic device can delete the voice obtained during the second recording and the text information corresponding to the voice obtained during the second recording.
  • the voice obtained during the first recording and the text information corresponding to the voice can also be retained in Note 1.
  • the user interface displayed by the electronic device can refer to the aforementioned description of (2) in Figure 21. After deleting L segments of speech but retaining the text information corresponding to the L segments of speech, the user interface displayed by the electronic device can refer to the aforementioned description of (2) in Figure 20 or (3) in Figure 20, which will not be repeated here.
  • the user interface displayed by the electronic device can refer to that shown in (1) in the following Figure 23.
  • the playback progress bar 141c included in the user interface 59a includes Q-L-1 nodes, that is, compared with before the deletion, the number of nodes has been reduced by L, corresponding to the deletion of L segments of speech.
  • the total duration recorded in the playback time indicator 141d is changed from the total duration of Q segments of speech to the total duration of Q-L segments of speech. For example, here it is changed to the total duration of the speech obtained by the first recording: 11s.
  • Area 12 includes text information corresponding to the Q-L segments of speech that have not been deleted, and no longer includes text information corresponding to the L segments of speech that have been deleted.
  • the user interface displayed by the electronic device can refer to that shown in (2) in the following Figure 23.
  • the playback progress bar 141c included in the user interface 59b includes Q-L-1 nodes, that is, compared with before the deletion, the number of nodes has been reduced by L, corresponding to the deletion of L segments of speech.
  • the total duration recorded in the playback time indicator 141d is changed from the total duration of Q segments of speech to the total duration of Q-L segments of speech, for example, here it is changed to the total duration of the speech obtained by the first recording: 11s.
  • Area 12 includes text information corresponding to the Q segments of speech.
  • the electronic device can trigger the display of an interface involved in managing (editing) the segment of voice in other ways.
  • the process can refer to the following description of FIG.
  • the electronic device in response to an operation (eg, a click operation) on the voice menu control 141e, the electronic device may not display the display box A mentioned above. Only the editing control is displayed, and the editing control may be It is used to trigger the display of the interface involved in managing (editing) the speech segment.
  • An exemplary display of the editing control can refer to the editing control 611 included in the user interface 61 shown in (2) of FIG. 24 below.
  • the electronic device can display more controls involved in managing (editing) the speech segment. It can also trigger the display of information corresponding to the speech segment, which may include the total duration of each speech, the name of the speech, etc. At this time, the electronic device can display the user interface shown in (3) in Figure 24.
  • the controls involved in the user interface 62 please refer to the aforementioned description of (2) in Figure 22, which will not be repeated here.
  • Sharing information includes transmitting the content to be shared in the note to other terminals, and the sharing method can be selected.
  • the content to be shared can be selected by the user through the electronic device.
  • the content to be shared may be information other than voice information in the note.
  • the information other than voice information includes speech-to-text content in area 12 and non-recording content in area 2.
  • FIG. 25 describes a set of exemplary user interfaces involved when the content to be shared is information other than voice in a note.
  • user interface 70 is an exemplary user interface involved in processing note 1.
  • the user interface 70 includes at least one functional control for note 1.
  • it may include a sharing control 701a, a favorite control 701b, a delete control 701c, and a more control 701d.
  • the favorite control 701b can be used to trigger the collection of note 1.
  • the delete control 701c can be used to trigger the deletion of note 1.
  • the more control 701d can be used to trigger the display of more functional controls for note 1.
  • the sharing control 701a can be used to trigger the display of related interfaces involved in sharing the content in note 1.
  • the electronic device may display the user interface 71 shown in (2) of FIG. 25.
  • the user interface 71 is an exemplary interface involved in sharing the content in the note 1.
  • the user interface 71 may include a selection box 711, and the selection box 711 may include controls involved in selecting different sharing methods.
  • the selection box 711 may include controls 711a, controls 711b, controls 711c, and controls 711d.
  • the control 711a may be used to trigger the content to be shared in the note to be converted into a picture and then transmitted to other terminals in the form of a picture.
  • the control 711b may be used to trigger the content to be shared in the note to be converted into text and then transmitted to other terminals in the form of a document, and the format of the text includes txt.
  • the control 711d may be used to trigger the content to be shared in the note to be transmitted to other terminals in the form of a note, that is, the format of the note may not be changed when sharing.
  • the control 711c may be used to trigger the conversion of the note into a document and then save it.
  • the process includes: the electronic device may call the function of generating a document with a document editing application to convert the note 1 into a document and then save it. Subsequently, the electronic device may share the document to the terminal, wherein the document may include a word document or a PDF document.
  • the electronic device may display the user interface 72 shown in (3) of the following FIG. 25, which is an exemplary sharing interface involved in converting the content to be shared in the note into a picture for sharing.
  • the user interface 72 may also include icons corresponding to applications with sharing functions, for example, the applications with sharing functions may be application 1, application 2, and other applications shown in the figure. Among them, the icon corresponding to the application with sharing function is used to trigger the transmission of the content to be shared to other devices through the application corresponding to the icon.
  • the user interface 72 may also include a selection control 722.
  • the state of the selection control 722 may be set to state G (e.g., state G is grayed out), and the selection control 722 in state G may be used to indicate that the speech-to-text content in area 12 is to be used as part of the content to be shared.
  • the electronic device may integrate the speech-to-text content in area 12 and the non-recording information in area 2 into a picture for sharing.
  • An exemplary display of the integrated picture may refer to picture 721 shown in the interface 72.
  • the electronic device can integrate the speech-to-text content in area 12 and the non-recording information in area 2 into a picture and transmit it to other devices through an application with a sharing function (such as application 1, application 2 and other applications shown in the figure).
  • a sharing function such as application 1, application 2 and other applications shown in the figure.
  • the electronic device may display the user interface 73 shown in (4) of FIG. 25.
  • the electronic device may set the state of the selected control from state G to state F (e.g., not grayed out), and the selected control 722 in state F may be used to indicate that the speech-to-text content in area 12 may not be included as part of the content to be shared.
  • the electronic device may integrate the non-recorded information in area 2 into a picture for sharing.
  • An exemplary display of the integrated picture may refer to picture 731 shown in interface 73.
  • the electronic device may integrate the non-recording information in area 2 into a picture and transmit it to other devices through an application with a sharing function (such as application 1, application 2, and other applications shown in the figure).
  • a sharing function such as application 1, application 2, and other applications shown in the figure.
  • control 722 to indicate that the speech-to-text content in area 12 is used as part of the content to be shared is used as an example.
  • control 722 can also be selected to indicate that the speech-to-text content in area 12 is not used as part of the content to be shared by default. This embodiment of the application is not limited to this.
  • the content to be shared may be all the content in the note.
  • the entire content may include all voices, voice-to-text content in area 12, and non-recorded content in area 2.
  • FIG. 26 describes a set of exemplary user interfaces involved when the content to be shared may be the entire content in the note.
  • the electronic device may display the user interface 74a shown in (1) of FIG26 .
  • the user interface 74a is another exemplary interface involved in sharing the content in the note 1.
  • the user interface 74a may include a selection box 741, and the selection box 741 may include controls involved in selecting different sharing methods.
  • the selection box 741 may include controls 741a and controls 741b.
  • control 741a can be used to trigger the transmission of the voice in the content to be shared to other terminals in the form of audio, and, after converting other information other than the voice into pictures, it is transmitted to other terminals in the form of pictures.
  • Control 741b can be used to trigger the transmission of the voice in the content to be shared to other terminals in the form of audio, and, after converting other information other than the voice into text, it is transmitted to other terminals in the form of a document, and the format of the text includes txt.
  • Other sharing methods may also be included, such as the aforementioned method of sharing to other devices in the form of notes, which is not limited in the embodiments of the present application.
  • the electronic device may display the user interface 74b shown in (2) of FIG. 26 below.
  • the user interface 74b is an example of converting the content to be shared in the note into pictures and audio for sharing. Sharing interface.
  • the user interface 74b may include a selection control 742a and a selection control 742b.
  • the status of the selection control 742a and the selection control 742b can be set to status J (for example, the status J is grayed out), and the selection control 742a under status J can be used to indicate that the speech-to-text content in area 12 is used as part of the content to be shared.
  • the control 742b under status J is used to indicate that the voice in note 1 is used as part of the content to be shared.
  • the electronic device can transmit the voice in note 1 as audio (for example, audio 743) to other devices.
  • the electronic device can also integrate the speech-to-text content in area 12 and the non-recording information in area 2 into a picture for sharing.
  • An exemplary display of the integrated picture can refer to the picture 744 shown in the user interface 74b.
  • the electronic device may transmit the audio 743 and the picture 744 to other devices through an application with a sharing function (such as application 1, application 2, and other applications shown in the figure).
  • an application with a sharing function such as application 1, application 2, and other applications shown in the figure.
  • the display modes of the audio 743 and the image 744 received by other devices include but are not limited to the following display modes:
  • Display mode 11 The audio 743 and the picture 744 are displayed as two pieces of information.
  • the display mode of the audio 743 includes displaying it in the form of an audio file. After downloading the audio file, the audio 743 can be played.
  • Display mode 12 The audio 743 and the picture 744 are displayed as two pieces of information.
  • the display mode of the audio 743 includes displaying it as an audio message, and in response to an operation (such as a click operation) on the audio message, the audio 743 can be played.
  • Display mode 13 The audio 743 and the picture 744 are displayed as two pieces of information.
  • the display mode of the audio 743 includes displaying it in the form of a voice bar, and two controls can be provided in the voice bar, one of which is used to trigger the audio playback, and the other is used to trigger the voice-to-text conversion.
  • the state of the selected control 742b is state K, for example, the state K may not be grayed out.
  • the electronic device may integrate the speech-to-text content in area 12 and the non-recording information in area 2 into a picture or document for sharing. This process may refer to the aforementioned description of (3) in Figure 25, which will not be repeated here.
  • the states of the selected control 742a and the selected control 742b are both in state K.
  • the electronic device may not analyze the speech-to-text content in area 12, and may integrate the non-recording information in area 2 into a picture or document for sharing. This process may refer to the aforementioned description of (4) in Figure 25, which will not be repeated here.
  • the content to be shared may include the selected voice and may also include text information corresponding to the selected voice.
  • FIG. 27A describes a set of exemplary user interfaces involved when the content to be shared may be a selected voice and text information corresponding to the selected voice.
  • user interface 57 is an exemplary interface involved in editing a voice.
  • the voice selected by the electronic device includes recording 002 (the name corresponding to the voice obtained by the second recording).
  • the electronic device can display a prompt box C involved in sharing the selected voice, and the prompt box C can be used to prompt the user that the content to be shared is the selected voice, and can also be used to receive the user's input to confirm the operation of sharing the selected voice or cancel the operation of sharing the selected voice.
  • the prompt box 751 displayed in the user interface 75 may be an exemplary prompt box C.
  • the prompt box 751 may also include prompt information 751a, and the prompt information 751a may be used to prompt the user whether to share the selected recording (voice).
  • the prompt box 751 may also include a selected control 751b, which can be used to receive an operation of whether the user shares the voice-to-text content (the text information corresponding to the selected voice).
  • the voice-to-text content is selected by default, and the selected control 751b is in state E1, for example, state E1 is grayed out.
  • state E1 is grayed out.
  • the electronic device can change the state of the selected control 751b to state F1, for example, the state F1 is not grayed out.
  • state F1 When the state of the selected control 751b is state F1, it means that when the selected voice is shared, the text information corresponding to the selected voice may not be shared.
  • the state of the selected control 751b can be switched between state E1 and state F1 to determine whether the text information corresponding to the selected voice is shared when the selected voice is shared.
  • the prompt box 751 may also include a confirmation sharing control 751c.
  • the confirmation sharing control 751c can be used to trigger the sharing of the selected voice and the text information corresponding to the selected voice.
  • the electronic device can display the user interface involved in selecting the sharing method. After selecting the sharing method, the electronic device can transmit the selected voice and the text information corresponding to the selected voice to other terminals.
  • the process can refer to the aforementioned description of the sharing methods involved in (1) and (2) in Figure 26, which will not be repeated here.
  • the confirmation sharing control 751c can be used to trigger sharing of the selected voice, but not to share the text information corresponding to the selected voice.
  • the electronic device in response to the operation of the confirmation sharing control 751c, can transmit the selected voice in the form of audio to other devices through applications with sharing functions (e.g., application 1, application 2, and other applications shown in the figure).
  • the aforementioned selection control 751b is optional. In some possible cases, when sharing the selected voice, it may not be involved in whether to share the text information corresponding to the selected voice. By default, the text information corresponding to the selected voice is not shared. The embodiment of the present application does not limit this.
  • the content to be shared may include the selected voice, text information corresponding to the selected voice, and notes to which the voice belongs.
  • the electronic device in response to an operation on the sharing control 531, may not display the aforementioned prompt box A, but may display a prompt box D involved in sharing the selected voice, which may be used to prompt the user that the content to be shared is the selected voice, and may also be used to receive an operation input by the user to confirm sharing the selected voice, or to cancel sharing the selected voice. It may also be used to receive an operation input by the user to confirm sharing the note described by the selected voice, or to cancel sharing the note.
  • the prompt box 761 displayed in the user interface 76 may be an exemplary prompt box D.
  • the prompt box 761 may also include prompt information 751a, which may be used to prompt the user whether to share the selected recording (voice).
  • the prompt box 761 may also include a selection control 751b, which may be used to receive an operation of whether the user shares the voice-to-text content (the text information corresponding to the selected voice).
  • a selection control 751b which may be used to receive an operation of whether the user shares the voice-to-text content (the text information corresponding to the selected voice).
  • the prompt box 761 may also include a check control 7611, which may be used to receive an operation of whether to share the note (note 1) to which the selected voice belongs.
  • note 1 is selected to be shared by default, and the check control is in state E2, for example, state E2 is grayed out.
  • state E2 is grayed out.
  • the electronic The device can change the state of the selected control 7611 to state F2, for example, the state F2 is not grayed out.
  • state F2 When the state of the selected control 7611 is state F2, it means that when the selected voice is shared, the note to which the selected voice belongs may not be shared.
  • the state of the selected control 7611 can be switched between state E2 and state F2 to determine whether to share the note to which the selected voice belongs when the selected voice is shared.
  • the content to be shared may include selected speech-to-text content.
  • FIG. 28 describes a set of exemplary user interfaces involved when the content to be shared may be selected speech-to-text content.
  • the user interface 17a is an exemplary user interface involved in processing the selected speech-to-text content.
  • the selection mark 162 indicates that the selected speech-to-text content is: "The second part mainly talks about XXXXX.”
  • the electronic device can transmit the selected speech-to-text content to other devices.
  • the user interface involved in the embodiment of the present application is an exemplary display, and any user interface can include more or fewer controls.
  • the form of each control can be changed, including but not limited to icons, text, etc.
  • the position of each control can also be changed, and should not constitute a limitation on the embodiment of the present application.
  • the conversion control 137a displayed in area 1 can be displayed in area 11 or other places in addition to being displayed in area 12.
  • the conversion control 137a can be placed in area 11.
  • the conversion control 1 can be placed in area 11.
  • FIG. 30 depicts an exemplary flowchart of the information processing method of the present application.
  • the exemplary process involved in recording information by the electronic device may refer to the following description of steps S101 to S106.
  • the electronic device displays interface 1, which includes area 1, area 2 and a conversion control in state C.
  • the note 1 mentioned in the foregoing content may also be referred to as the first file.
  • the interface 1 may also be referred to as the first interface.
  • the area 1 may be referred to as the first area, and the area 2 may be referred to as the second area.
  • the state C may be referred to as the first state.
  • the conversion control may be referred to as the first control.
  • step S101 the relevant process involved in establishing note 1 can refer to the aforementioned description of Figures 1A and 1B, which will not be repeated here.
  • the conversion control may be the conversion control 137a mentioned above.
  • Area 1 is the area mentioned above for presenting content related to recording.
  • Area 2 is the editing area mentioned above.
  • the interface 1 may be the exemplary user interface mentioned above including area 1, area 2 and the conversion control in state C. For example, it may be the user interface 13b shown in (3) of FIG. 2, the user interface 13c shown in (4) of FIG. 2, etc.
  • the conversion control in state C is used to trigger the conversion of the recorded speech into text.
  • the display control 1 instructs the electronic device to convert the voice 1 into The received text is displayed in area 1, and the input information 1 is received and displayed in area 2.
  • the voice 1 may also be referred to as the first voice
  • the control 1 may also be referred to as the second control
  • the information 1 may also be referred to as the first information.
  • Voice 1 may be the voice recorded for the first time mentioned above.
  • the first time may be a certain time when the electronic device records voice 1.
  • Information 1 may be information input by the electronic device through an input tool, such as text or a picture.
  • the input tool includes a keyboard and a stylus tool bar.
  • Control 1 may be the recording bar (recording bar 1) or the pause recording control 136a in the state 1 mentioned above.
  • the recording control included in interface 1 can trigger recording voice 1.
  • the electronic device can turn on the voice-to-text function.
  • the electronic device can display the text converted from voice 1 in area 1 while recording voice 1, and can input information (e.g., information 1) input by the input tool, and display the information 1 in area 2.
  • the exemplary user interface involved in this process can refer to the user interface 13c and other related interfaces shown in Figure 2 above.
  • the electronic device may update the state of the conversion control to state D, and the conversion control in state D may be used to indicate stopping converting speech into text.
  • the electronic device can still continue to record the voice, and during the voice recording process, information can be input through the input tool.
  • Area 1 is divided into area 11 and area 12, area 11 includes control 1, and area 12 includes text obtained by voice conversion.
  • the electronic device replaces control 1 with control 2, and control 2 is used to instruct the electronic device to end recording voice and is also used to trigger the electronic device to record other voices.
  • the area 11 may be referred to as a first sub-area, and the area 12 may be referred to as a second sub-area.
  • the control 2 may also be referred to as a third control.
  • Area 11 may be the aforementioned voice recording area, and area 12 may be the aforementioned voice-to-text area.
  • control 1 When control 1 is a recording bar (recording bar 1) in state 1, control 2 may be a recording bar (recording bar 2) in state 2.
  • control 1 When control 1 is a pause recording control 136a, control 2 may be a start recording control 141a.
  • step S104 can refer to the aforementioned description of the relevant contents in (2) in FIG. 7 and (3) in FIG. 7 , which will not be repeated here.
  • the electronic device records the voice 1, the text converted from the voice 1, and the information 1 into the note 1.
  • FIG31 is a schematic diagram of the structure of an electronic device provided in an embodiment of the present application.
  • the electronic device may have more or fewer components than those shown in the figure, may combine two or more components, or may have different component configurations.
  • the various components shown in the figure may be implemented in hardware, software, or a combination of hardware and software including one or more signal processing and/or application specific integrated circuits.
  • the electronic device may include: a processor 110, an external memory interface 120, an internal memory 121, a universal serial bus (USB) interface 130, a charging management module 140, a power management module 141, a battery 142, an antenna 1, an antenna 2, a mobile communication module 150, a wireless communication module 160, an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, an earphone interface 170D, a sensor module 180, a button 190, a motor 191, an indicator 192, a camera 193, a display screen 194 and a subscriber identification module (SIM) card interface 195, etc.
  • SIM subscriber identification module
  • the sensor module 180 may include a pressure sensor 180A, a gyroscope sensor 180B, an air pressure sensor 180C, a magnetic sensor 180D, an acceleration sensor 180E, a distance sensor 180F, a proximity light sensor 180G, a fingerprint sensor 180H, a temperature sensor 180J, a touch sensor 180K, an ambient light sensor 180L, a bone conduction sensor 180M, etc.
  • the structures illustrated in the embodiments of the present application do not constitute specific limitations on the electronic device.
  • the electronic device may include more or fewer components than shown in the figure, or combine certain components, or split certain components, or arrange the components differently.
  • the illustrated components may be implemented in hardware, software, or a combination of software and hardware.
  • the processor 110 may include one or more processing units, for example, the processor 110 may include an application processor (application processor, AP), a modem processor, a graphics processor (graphics processing unit, GPU), an image signal processor (image signal processor, ISP), a controller, a memory, a video codec, a digital signal processor (digital signal processor, DSP), a baseband processor, and/or a neural-network processing unit (neural-network processing unit, NPU), etc.
  • different processing units may be independent devices or integrated in one or more processors.
  • the controller can be the nerve center and command center of the electronic device.
  • the controller can generate operation control signals according to the instruction operation code and timing signal to complete the control of fetching and executing instructions.
  • the processor 110 may also be provided with a memory for storing instructions and data.
  • the memory in the processor 110 is a cache memory.
  • the memory may store instructions or data that the processor 110 has just used or cyclically used. If the processor 110 needs to use the instruction or data again, it may be directly called from the memory. This avoids repeated access, reduces the waiting time of the processor 110, and thus improves the efficiency of the system.
  • the processor 110 may include one or more interfaces.
  • the interface may include an inter-integrated circuit (I2C) interface, an inter-integrated circuit sound (I2S) interface, a pulse code modulation (PCM) interface, etc.
  • I2C inter-integrated circuit
  • I2S inter-integrated circuit sound
  • PCM pulse code modulation
  • the interface connection relationship between the modules illustrated in the embodiments of the present application is only a schematic illustration and does not constitute a structural limitation on the electronic device.
  • the electronic device may also adopt different interface connection methods in the above embodiments, or a combination of multiple interface connection methods.
  • the charging management module 140 is used to receive charging input from a charger.
  • the power management module 141 is used to connect the battery 142 , the charging management module 140 and the processor 110 .
  • the wireless communication function of the electronic device can be implemented through antenna 1, antenna 2, mobile communication module 150, wireless communication module 160, modem processor and baseband processor.
  • the antenna 1 of the electronic device is coupled to the mobile communication module 150, and the antenna 2 is coupled to the wireless communication module 160, so that the electronic device can communicate with the network and other devices through wireless communication technology.
  • the wireless communication technology may include global system for mobile communications (GSM), general packet radio service (GPRS), etc.
  • the electronic device realizes the display function through a GPU, a display screen 194, and an application processor.
  • the display screen 194 is used to display images, videos, etc.
  • the display screen 194 includes a display panel.
  • the electronic device may include 1 or N display screens 194, where N is a positive integer greater than 1.
  • the electronic device can realize the shooting function through ISP, camera 193, video codec, GPU, display screen 194 and application processor.
  • the ISP is used to process data fed back by the camera 193 .
  • the camera 193 is used to capture still images or videos.
  • Digital signal processors are used to process digital signals. In addition to processing digital image signals, they can also process other digital signals. For example, when an electronic device selects a frequency point, a digital signal processor is used to perform Fourier transform on the frequency point energy.
  • Video codecs are used to compress or decompress digital video.
  • NPU is a neural-network (NN) computing processor. It can quickly process input information by drawing on the structure of biological neural networks, such as the transmission mode between neurons in the human brain, and can also continuously self-learn.
  • NN neural-network
  • the internal memory 121 may include one or more random access memories (RAM) and one or more non-volatile memories (NVM).
  • RAM random access memories
  • NVM non-volatile memories
  • the external memory interface 120 can be used to connect to an external non-volatile memory to expand the storage capacity of the electronic device.
  • the external non-volatile memory communicates with the processor 110 through the external memory interface 120 to implement a data storage function. For example, files such as music and videos are stored in the external non-volatile memory.
  • the electronic device can implement audio functions such as music playing and recording through the audio module 170, the speaker 170A, the receiver 170B, the microphone 170C, the headphone jack 170D, and the application processor.
  • the audio module 170 is used to convert digital audio information into analog audio signal output, and is also used to convert analog audio input into digital audio signals.
  • the speaker 170A also called a "speaker" is used to convert an audio electrical signal into a sound signal.
  • the electronic device can listen to music, recordings, or listen to hands-free calls through the speaker 170A.
  • the receiver 170B also called a “handset”, is used to convert audio electrical signals into sound signals.
  • Microphone 170C also called “microphone” or “microphone”, is used to convert sound signals into electrical signals.
  • the earphone jack 170D is used to connect a wired earphone.
  • Buttons 190 include a power button, a volume button, and the like.
  • Motor 191 can generate vibration prompts.
  • the indicator 192 may be an indicator light, which may be used to indicate the charging status, the change in the power level, or may be used to indicate a message. Missed calls, notifications, and more.
  • the SIM card interface 195 is used to connect a SIM card.
  • the processor 110 can call the computer instructions stored in the internal memory 121 to enable the electronic device to execute the information processing method in the embodiment of the present application.
  • the term "when" may be interpreted to mean “if" or “after" or “in response to determining" or “in response to detecting", depending on the context.
  • the phrases “upon determining" or “if (the stated condition or event) is detected” may be interpreted to mean “if determining" or “in response to determining" or “upon detecting (the stated condition or event)” or “in response to detecting (the stated condition or event)", depending on the context.
  • the computer program product includes one or more computer instructions.
  • the computer can be a general-purpose computer, a special-purpose computer, a computer network, or other programmable devices.
  • the computer instructions can be stored in a computer-readable storage medium, or transmitted from one computer-readable storage medium to another computer-readable storage medium.
  • the computer instructions can be transmitted from a website site, computer, server or data center by wired (e.g., coaxial cable, optical fiber, digital subscriber line) or wireless (e.g., infrared, wireless, microwave, etc.) mode to another website site, computer, server or data center.
  • the computer-readable storage medium can be any available medium that a computer can access or a data storage device such as a server or data center that contains one or more available media integration.
  • the available medium can be a magnetic medium, (e.g., a floppy disk, a hard disk, a tape), an optical medium (e.g., a DVD), or a semiconductor medium (e.g., a solid-state hard disk), etc.
  • the processes can be completed by a computer program to instruct the relevant hardware, and the program can be stored in a computer-readable storage medium.
  • the program When the program is executed, it can include the processes of the above-mentioned method embodiments.
  • the aforementioned storage medium includes: ROM or random access memory RAM, magnetic disk or optical disk and other media that can store program codes.

Landscapes

  • Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Physics & Mathematics (AREA)
  • Signal Processing (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Business, Economics & Management (AREA)
  • Business, Economics & Management (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Multimedia (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Acoustics & Sound (AREA)
  • General Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

本申请实施例提供了一种信息处理方法和电子设备。该方法可以用于电子设备生成笔记,在笔记中记录信息,并对记录的信息进行处理。这里涉及的信息包括但不限于语音信息、文字信息或者图片信息等。涉及的信息处理包括但不限于以下内容中的一个或者多个:记录信息、查看信息以及管理信息。其中,记录信息包括输入信息,以及,基于所输入信息进行处理并将处理结果作为笔记中的内容。例如,将输入的语音信息转换为文字信息、对输入的文字信息进行编辑或者标记等。管理信息又可以包括删除信息、传输(分享)信息等。

Description

信息处理方法和电子设备
本申请要求于2022年11月4日提交中国专利局、申请号为202211379551.4、申请名称为“信息处理方法和电子设备”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请涉及终端技术领域,尤其涉及信息处理方法和电子设备。
背景技术
随着电子设备功能的逐渐完善,越来越多的电子设备可以记录信息,并将该信息进行保存。例如,“备忘录”应用中就可以记录并保存输入的文字信息等。这样,使得用户利用电子设备实现信息记录,不再依靠传统的纸质记录方式,提高了信息记录的效率。
电子设备可以将用户的记录信息整合到一个文本中,这样的文本也可以被称为笔记。现阶段,用户通过电子设备记录的信息可以包括文字信息、图片信息等。
发明内容
本申请提供了一种信息处理方法和电子设备,可以在记录非语音信息时,同时进行录音记录语音,且同时将记录的语音转换成文字进行记录,也可以关闭语音转文本功能,提高了记录信息的效率,以及语音转文字的灵活性。
第一方面,本申请提供了一种信息处理方法,该方法包括:电子设备显示第一界面,该第一界面包括第一区域、第二区域以及第一状态的第一控件;在第一时刻,该电子设备显示第二控件和第一状态的第一控件时,该电子设备录制第一语音,将该第一语音转换得到的文字显示在该第一区域;第二控件用于指示正在录音,第一状态的第一控件用于指示在录音时将录制的语音转化为文本;该电子设备显示第二控件和第一状态的第一控件时,接收用户输入的第一信息,并将该第一信息显示在该第二区域中;响应于针对该第一状态的第一控件的操作,在显示第二控件的情况下,该电子设备将该第一控件的状态切换为第二状态,停止将录制的语音转换为文字,该第二状态的第一控件用于指示暂停将录制的语音转化为文本;该电子设备将该第一语音、该第一语音转换得到的文字以及该第一信息记录到第一文件中。
上述实施例中,第一区域可以是实施例中涉及的区域1,第二区域可以是实施例中涉及的区域2,第一控件可以是实施例中涉及的转换控件,第一状态的转换控件可以为状态C的转换控件,第二状态的转换控件可以为状态D的转换控件。第二控件可以为实施例中涉及的暂停录制控件,第一语音可以为实施例中涉及的第一次录制的语音。
电子设备可以在录音的同时支持语音转文字,以及接受用户输入的其他信息(例如文字、图片等),并且将该其他信息显示在第二区域中。并且,可以在开启语音转文字功能、关闭语音转文字功能之间进行切换。即,录音的同时,可以通过开启语音转文字功能以实现将录音转换为文字信息,并将转换得到的文字信息显示在第一区域中。这样,在开启语音转文字功能的情况下,可以实现在录音的同时接受用户输入的其他信息。还可以关闭语 音转文字功能,在录音的同时可以不将该录制的语音转换为文字信息。这样,在关闭语音转文字功能的情况下,可以实现在录音时接受用户输入的其他信息。其中,该其他信息也可以被称为非录音信息。提高了记录信息的效率,以及语音转文字的灵活性。
结合第一方面,在一些实施例中,该第一区域中包括第一子区域以及第二子区域,第一子区域用于显示该第二控件;该第二子区域用于显示语音转换得到的文本;在该电子设备显示第二控件时,该方法还包括:响应于针对该第二控件的操作,该电子设备将该第二控件替换成第三控件;该第三控件用于指示该电子设备结束录制语音;结束录制语音后,在该电子设备未获取到该语音转换得到的文字时,该第一区域不包括该第二子区域。
上述实施例中,第一区域可以被划分为两个子区域,其中,第一子区域可以为实施例中涉及的区域11,第二子区域可以为实施例中涉及的区域12。第一子区域可以用于展示与录音过程相关的内容,例如与录音相关的控件等。第二子区域中可以显示语音转文本内容。
如果第二子区域中没有显示语音转换后的文字,则可以在暂停录制语音的情况下,将该第二子区域关闭,不显示该第二子区域,这样,可以简化用户界面。
结合第一方面,在一些实施例中,该方法还包括:在播放该第一语音的情况下,该电子设备显示进度条,该进度条用于指示播放语音的进度;在该进度条的第一位置显示第一标识的情况下,该电子设备用第三状态显示该第一信息,且,以第四状态显示该第二区域中的第二信息;其中,该第一标识在该第一位置用于指示该电子设备在播放该第一时刻录制的语音;该第二信息的输入时间为第二时刻,该第二时刻晚于该第一时刻。
上述实施例中,进度条可以为实施例中涉及的播放进度条,第一标识可以为实施例中涉及的标识1411,第一位置可以为当前播放的音频在播放进度条上对应的位置,第三状态与第四状态不同。例如,在实施例中,第三状态可以为不添加蒙层,第四状态可以为添加蒙层。第二信息可以为实施例中涉及的非录音信息。
电子设备可以在播放语音的过程中,可以实现音文同步。对于比当前播放的语音被记录到笔记中的时间更早的非录音信息不添加蒙层。对于比当前播放的语音被记录到笔记中的时间更晚的非录音信息添加蒙层,使得其被区别显示,以提示用户该添加了蒙层的非录音信息在播放当前播放的语音被记录到笔记时,该非录音信息还没有被记录到笔记中。
结合第一方面,在一些实施例中,在该进度条的第一位置显示第一标识的情况下,该方法还包括:该电子设备在该第一区域中显示该第一时刻录制的语音转换得到的文字。
上述实施例中,电子设备还可以在第一区域中基于语音转文本内容实现音文同步,即第一区域中可以显示正在播放的语音转换的文字信息,便于用户预览。
结合第一方面,在一些实施例中,以第四状态显示第二区域中的第二信息之后,该方法还包括:响应于针对第四控件的操作,该电子设备以该第三状态显示该第二信息。
上述实施例中,该第四控件可以为实施例中涉及的停止笔记回放动画的控件,例如可以包括:笔记回复动画设置项381以及停止笔记回放动画控件351。电子设备可以通过第 四控件实现关闭音文同步功能,给用户提供更多的选项。
结合第一方面,在一些实施例中,在该进度条的第一位置显示第一标识的情况下,该方法还包括:响应于针对该进度条中第二位置的第一操作,该电子设备以第五状态显示第二区域中的第三信息;其中,该第三信息的输入时间为第三时刻,且,在该第二位置指示该电子设备在该第三时刻录制的语音。
上述实施例中,该第二位置一个示例可以为实施例中涉及的位置T,第五状态可以为实施例中涉及突出显示。电子设备在播放语音的过程中,还可以实现提前预览。长按进度条上的某一位置,该位置对应的语音输入时,同时被输入的文字可以被突出显示,便于用户预览。
结合第一方面,在一些实施例中,该方法还包括:在建立该第一文件后录制了Q段语音的情况下,该电子设备在该进度条中显示Q-1个节点;该Q-1个节点将该进度条划分为Q段子进度条;该Q段子进度条中的第i段子进度条用于指示该Q段语音中的第i段语音;该Q为大于等于2的整数;在检测到针对第五控件的操作之后,在该第一标识处于该第i段子进度条时,该电子设备播放该第i段语音。
上述实施例中,第五控件可以为实施例中涉及的开始播放控件。电子设备可以将多段语音信息(语音)分开录制,并将该多段语音信息制作成一段长语音,在播放该长语音时即可实现连续播放该多段语音。Q段语音表示电子设备进行了Q次录音,一次录音开始之后,电子设备可以采集声音信号,结束该次录音之后,即可基于采集的声音信号得到一段语音信息。即,电子设备可以在不同时间(分开的Q个时间段)录音,得到Q段语音。然后,电子设备可以将该Q段语音进行整合,得到一段长语音。这样,在播放长语音时,可以实现将分开录制的Q段语音进行连续播放。
结合第一方面,在一些实施例中,该方法还包括:该电子设备选中Q段语音中L段语音;该L为小于等于Q的整数;在第六控件处于第六状态的情况下,响应于针对第七控件的操作,该电子设备删除该L段语音,且,删除该L段语音转换得到的文字;在该第六控件不处于第六状态的情况下,响应于针对该第七控件的操作,该电子设备删除该L段语音,且,保留该L段语音转换得到的文字。
上述实施例中,第六控件可以为实施例中涉及的选中控件581b,第七控件可以为实施例至涉及的删除控件581c。电子设备可以提供在删除语音时,是否删除该语音转换得到的文字的功能,使得用户在删除语音时,可以有多种选择。
结合第一方面,在一些实施例中,该方法还包括:在该第八控件处于第八状态的情况下,响应于针对第九控件的操作,该电子设备将该第二区域中内容以及该Q段语音转换得到的文字传输至其他设备;在该第八控件不处于第八状态的情况下,响应于针对该第九控件的操作,该电子设备将该第二区域中内容传输至其他设备。
上述实施例中,第八控件可以为实施例中涉及的选中控件(例如选中控件521b),第 九控件可以为实施例中涉及的分享控件(例如分享控件)。电子设备可以提供在分享笔记中记录的非录音信息时,是否将笔记中的语音转换得到的文字进行分享功能,使得用户在分享信息时,可以有多种选择。
结合第一方面,在一些实施例中,接收输入的第一信息,具体包括:该电子设备在该第一界面中显示第一输入工具;该电子设备通过该第一输入工具接收输入的第一信息;该方法还包括:在显示该第一输入工具时,还显示第十控件;响应于针对该第十控件的操作,该电子设备将该第一输入工具切换为第二输入工具;在显示该第二输入工具时,还显示第十一控件;该第十一控件用于触发从第二输入工具切换至第一输入工具。
上述实施例中,第一输入工具可以是实施例中涉及的键盘,第十控件可以为实施例中涉及的手写控件,第二输入工具可以为实施例中涉及的手写笔工具栏,第十一控件可以为实施例中涉及的键盘切换控件。电子设备可以基于第十控件使得键盘快速切换成为手写笔工具栏。也可以基于第十一控件实现手写笔工具栏到键盘的切换。这样,便于用户使用不同的输入工具输入非录音信息。
结合第一方面,在一些实施例中,该方法还包括:响应于针对第十二控件的操作;该电子设备将该第二子区域的大小从第一尺寸变更为第二尺寸。
上述实施例中,第十二控件可以为实施例中涉及的展开控件或者收起控件。电子设备可以使得第二子区域的大小可调,为用户提供更多选择性。
第二方面,本申请提供了一种电子设备,该电子设备包括:一个或多个处理器和存储器;该存储器与该一个或多个处理器耦合,该存储器用于存储计算机程序代码,该计算机程序代码包括计算机指令,该一个或多个处理器调用该计算机指令以使得该电子设备执行如第一方面或第一方面的任意一种实施方式所描述的方法。
上述实施例中,电子设备可以在录音的同时支持语音转文字,以及接受用户输入的其他信息(例如文字、图片等),并且将该其他信息显示在第二区域中。并且,可以在开启语音转文字功能、关闭语音转文字功能之间进行切换。即,录音的同时,可以通过开启语音转文字功能以实现将录音转换为文字信息,并将转换得到的文字信息显示在第一区域中。这样,在开启语音转文字功能的情况下,可以实现在录音的同时接受用户输入的其他信息。还可以关闭语音转文字功能,在录音的同时可以不将该录制的语音转换为文字信息。这样,在关闭语音转文字功能的情况下,可以实现在录音时接受用户输入的其他信息。其中,该其他信息也可以被称为非录音信息。提高了记录信息的效率,以及语音转文字的灵活性。
第三方面,本申请实施例提供了一种芯片***,该芯片***应用于电子设备,该芯片***包括一个或多个处理器,该处理器用于调用计算机指令以使得该电子设备执行如第一方面或第一方面的任意一种实施方式所描述的方法。
上述实施例中,电子设备可以在录音的同时支持语音转文字,以及接受用户输入的其他信息(例如文字、图片等),并且将该其他信息显示在第二区域中。并且,可以在开启 语音转文字功能、关闭语音转文字功能之间进行切换。即,录音的同时,可以通过开启语音转文字功能以实现将录音转换为文字信息,并将转换得到的文字信息显示在第一区域中。这样,在开启语音转文字功能的情况下,可以实现在录音的同时接受用户输入的其他信息。还可以关闭语音转文字功能,在录音的同时可以不将该录制的语音转换为文字信息。这样,在关闭语音转文字功能的情况下,可以实现在录音时接受用户输入的其他信息。其中,该其他信息也可以被称为非录音信息。提高了记录信息的效率,以及语音转文字的灵活性。
第四方面,本申请实施例提供了一种包含指令的计算机程序产品,当该计算机程序产品在电子设备上运行时,使得该电子设备执行如第一方面或第一方面的任意一种实施方式所描述的方法。
上述实施例中,电子设备可以在录音的同时支持语音转文字,以及接受用户输入的其他信息(例如文字、图片等),并且将该其他信息显示在第二区域中。并且,可以在开启语音转文字功能、关闭语音转文字功能之间进行切换。即,录音的同时,可以通过开启语音转文字功能以实现将录音转换为文字信息,并将转换得到的文字信息显示在第一区域中。这样,在开启语音转文字功能的情况下,可以实现在录音的同时接受用户输入的其他信息。还可以关闭语音转文字功能,在录音的同时可以不将该录制的语音转换为文字信息。这样,在关闭语音转文字功能的情况下,可以实现在录音时接受用户输入的其他信息。其中,该其他信息也可以被称为非录音信息。提高了记录信息的效率,以及语音转文字的灵活性。
第五方面,本申请实施例提供了一种计算机可读存储介质,包括指令,当该指令在电子设备上运行时,使得该电子设备执行如第一方面或第一方面的任意一种实施方式所描述的方法。
上述实施例中,电子设备可以在录音的同时支持语音转文字,以及接受用户输入的其他信息(例如文字、图片等),并且将该其他信息显示在第二区域中。并且,可以在开启语音转文字功能、关闭语音转文字功能之间进行切换。即,录音的同时,可以通过开启语音转文字功能以实现将录音转换为文字信息,并将转换得到的文字信息显示在第一区域中。这样,在开启语音转文字功能的情况下,可以实现在录音的同时接受用户输入的其他信息。还可以关闭语音转文字功能,在录音的同时可以不将该录制的语音转换为文字信息。这样,在关闭语音转文字功能的情况下,可以实现在录音时接受用户输入的其他信息。其中,该其他信息也可以被称为非录音信息。提高了记录信息的效率,以及语音转文字的灵活性。
附图说明
图1A以及图1B示出了对电子设备新建一个笔记的示例性用户界面;
图2-图7示出了第一次录音及其涉及的示例性用户界面;
图8示出了第一次录音结束后对语音转文字内容进行编辑处理时涉及的示例性用户界面;
图9-图12示出了第二次录音及其涉及的示例性用户界面;
图13-图17示出了查看信息涉及的一组示例性用户界面;
图18-图24为电子设备删除信息时涉及的一组示例性用户界面;
图25-图26和图27A-图27B以及图28-图29为电子设备分享信息时涉及的一组示例性用户界面;
图30中描述了本申请中信息处理方法的一个示例性流程图;
图31是本申请实施例提供的电子设备的结构示意图。
具体实施方式
本申请以下实施例中所使用的术语只是为了描述特定实施例的目的,而并非旨在作为对本申请的限制。如在本申请的说明书和所附权利要求书中所使用的那样,单数表达形式“一个”、“一种”、“所述”、“上述”、“该”和“这一”旨在也包括复数表达形式,除非其上下文中明确地有相反指示。还应当理解,本申请中使用的术语“和/或”是指并包含一个或多个所列出项目的任何或所有可能组合。
以下,术语“第一”、“第二”仅用于描述目的,而不能理解为暗示或暗示相对重要性或者隐含指明所指示的技术特征的数量。由此,限定有“第一”、“第二”的特征可以明示或者隐含地包括一个或者更多个该特征,在本申请实施例的描述中,除非另有说明,“多个”的含义是两个或两个以上。
在一种方案中,电子设备可以通过录音的方式记录语音信息。在进行录音的同时还可以接收用户输入的其他信息,例如文字信息(以下内容中可以称为文字)、图片信息(以下内容中可以称为图片)等中的一个或者多个。
但是在该种记录信息的方式中,电子设备在录音的同时支持输入其他信息的情况下,不支持同时将录音转换成对应的文字信息。
本申请实施例中,提供了一种信息处理方法。该信息处理方法可以用于电子设备生成笔记。这里涉及的信息包括但不限于语音信息、文字信息或者图片信息等。涉及的信息处理包括但不限于以下内容中的一个或者多个:记录信息、查看信息以及管理信息。其中,记录信息包括输入信息,以及,基于所输入信息进行处理并将处理结果作为笔记中的内容。例如,将输入的语音信息转换为文字信息、对输入的文字信息进行编辑或者标记等。管理信息又可以包括删除信息、传输(分享)信息等。
实施该信息处理方法,电子设备可以实现以下功能中的一个或者多个,且不限于以下功能。
功能1.在记录信息时,电子设备可以在录音的同时支持语音转文字,以及接受用户输入的其他信息(例如文字、图片等)。并且,可以在开启语音转文字功能、关闭语音转文字功能之间进行切换。即,录音的同时,可以通过开启语音转文字功能以实现将录音转换为文字信息。这样,在开启语音转文字功能的情况下,可以实现在录音的同时接受用户输入的其他信息。还可以关闭语音转文字功能,在录音的同时可以不将该录制的语音转换为文字信息。这样,在关闭语音转文字功能的情况下,可以实现在录音时接受用户输入的其 他信息。其中,该其他信息也可以被称为非录音信息。
功能2.在记录信息时,电子设备可以将多段语音信息(语音)分开录制,并将该多段语音信息制作成一段长语音,在播放该长语音时即可实现连续播放该多段语音。Q段语音表示电子设备进行了Q次录音,一次录音开始之后,电子设备可以采集声音信号,结束该次录音之后,即可基于采集的声音信号得到一段语音信息。即,电子设备可以在不同时间(分开的Q个时间段)录音,得到Q段语音。然后,电子设备可以将该Q段语音进行整合,得到一段长语音。这样,在播放长语音时,可以实现将分开录制的Q段语音进行连续播放。其中,Q为大于等于2的整数。
在一些可能的情况下,第Q次录音停止之后,电子设备可以显示该段长语音对应的播放进度条。该播放进度条中包括Q-1个节点,将播进度条划分为Q段(可以被称为Q段子进度条)。其中,Q段子进度条中的第i段子进度条对应第i段语音。该第i段子进度在播放进度条中的长度与第i段语音总时长与长语音总时长的比值成正比:第i段语音总时长与长语音总时长的比值越大,表示第i段语音在长语音中的比重大,则第i段子进度条在播放进度条中的长度越长。
应该理解的是,这里涉及的长语音是指该语音比多段语音信息中的任何一段都包括更多的内容,可以看作是该多段语音信息的集合。而不是指该长语音的时间长。对于该长语音的总时长,本申请实施例不作限定。
还应该理解的是,在记录信息生成笔记的过程中,除了功能1、功能2以外,还包括其他功能。功能1、功能2可以使得电子设备更便捷的记录信息。对于记录信息这个场景下的相关内容可以参考下述对图1A、图1B以及图3-图12的描述,此处暂不赘述。
功能3.在查看信息时,电子设备在播放录音的情况下,可以实现音文同步。音文同步包括:在播放录音时,以状态A显示已播放语音对应的非录音信息(记为信息A),并且,以状态B显示未播放语音对应的非录音信息(记为信息B)。这里以电子设备在播放语音A(录制时间为时间A)为例对信息A以及信息B进行描述:
应该理解的是,该时间A可以指示一个时刻,也可以指示一个时间范围。
在时间A指示一个时刻的情况下,信息A是指输入时间早于或者等于时间A的信息。信息B是指输入时间晚于时间A的信息。
在时间A指示一个时间范围的情况下,信息A是指输入时间早于时间A的起始时间,或者,输入时间早于时间A的终止时间的信息。信息B是指输入时间晚于时间A的终止时间的信息。
其中,时间A也可以被称为语音A对应的时间戳。信息A的输入时间可以被称为信息A对应的时间戳。信息B的输入时间可以被称为信息B对应的时间戳。
应该理解的是,在查看信息时,除了功能3以外,还包括其他功能。功能3可以使得用户通过电子设备更便捷的查看信息。对于查看信息这个场景下的相关内容可以参考下述对图14-图17的描述,此处暂不赘述。
功能4.在管理信息时,电子设备可以实现删除信息、分享信息等。其中,删除信息包括:选择删除一个笔记中记录的部分或者全部语音信息。例如,可以选择删除语音信息,在删除语音信息时,还可以根据用户需求确定是否删除所选中语音信息对应的文字信息。 分享信息包括:将笔记中的待分享内容传输至其他终端,且可以选择分享方式。在一些可能的情况下,该待分享内容包括:笔记中除语音信息以外的其他内容,或者,包括笔记中除语音信息以及该语音信息所转换的文字信息以外的其他内容。且,用户可以通过电子设备选择是否将语音信息所转换的文字信息作为待分享内容中的一部分。
其中,分享方式包括但不限于以下分享方式中的一个或者多个:
分享方式1.将笔记中的待分享内容转换为图片之后,以图片的形式传输至其他终端。
分享方式2.将笔记中的待分享内容转换为文档之后,以文档的形式传输至其他终端。
分享方式3.将笔记中的待分享内容仍然以笔记的形式传输至其他终端。
关于管理信息这个场景下的相关内容可以参考下述对图18-图29的描述,此处暂不赘述。
首先对记录信息涉及的相关内容进行描述。
图1A、图1B以及图3-图12中描述了记录信息涉及的一组示例性用户界面。
这里以在一个笔记中分别录制两段语音,且在两次录音过程中,同时进行语音转文本以及输入其他信息为例进行描述。
其中,电子设备新建一个笔记的相关内容可以参加下述对图1A以及图1B的描述。第一次录音及其涉及的相关内容可以参考下述对图4-图7的描述。第一次录音结束之后,电子设备记录信息的相关内容,可以参考下述对图8的描述。第二次录音及其涉及的相关内容可以参加下述对图9-图12的描述。
下面基于图1A以及图1B对电子设备新建一个笔记的相关内容进行示例性描述。
参考图1A中(1)所示的用户界面10,响应于针对笔记类应用(例如荣耀笔记101)的操作(例如点击操作),电子设备可以开启该笔记类应用。此时,电子设备可以显示如图1A中(2)所示的用户界面11a。
如图1A中(2)所示,用户界面11a为新建笔记时涉及的一个示例性用户界面。该用户界面11a中可以包括查看全部笔记所对应的控件111。响应于针对该控件111的操作,电子设备可以显示下述图1A中(3)所示的用户界面11b。
该用户界面11b中,可以包括控件1131,该控件1131可以触发电子设备显示保存笔记涉及的文件夹。
响应于针对控件1131的操作,电子设备可以显示全部文件夹,其中,全部文件夹中包括用户通过电子设备建立的文件夹(例如可以统称为新建文件夹),还包括***中的默认文件夹。任一非空文件夹中可以包括至少一个文件,其中的每一个文件都可以被称为一个笔记。空文件夹中不包括文件。例如,参考图1A中(4)所示的用户界面11c所示,新建文件夹中可以包括文件夹1,该默认文件夹可以独立于新建文件夹。响应于针对文件夹1对应的图标112的操作,电子设备可以显示文件1中包括的全部笔记。例如,此时电子设备可以显示下述图1B中(1)所示的用户界面11d。
如图1B中(1)所示,用户界面11d中包括文件1中涉及的笔记,包括笔记A以及笔记B。
用户界面11a、11c中还可以包括新建控件113,该新建控件113可以用于触发新建笔 记。响应于用户在用户界面11c中针对新建控件113的操作(例如点击操作),电子设备新建的笔记属于当前打开的文件夹,此处属于文件1。
这里应该理解的时,在图1A中(2)示出的用户界面11a中,由于现在没有打开文件夹,而是显示的全部笔记,则响应于用户在用户界面11a中针对新建控件113的操作(例如点击操作),电子设备新建的笔记不属于用户通过电子设备建立的文件夹,属于***中包括的默认文件夹。后续,用户可以将该新建的笔记设置到其他文件夹中。
响应于针对新建控件113的操作,电子设备可以新建笔记。此时,电子设备可以显示图1B中(2)所示的用户界面。
如图1B中(2)所示,用户界面12为新建笔记之后涉及的一个初始化界面。响应于针对标题输入控件121的操作,电子设备可以设置该新建笔记的名称。例如,此处将该新建的笔记命名为笔记1。
应该理解的是,图1A以及图1B中示出的新建笔记的方式是一种示例性说明。实际应用中,新建笔记的过程可以涉及更多或者更少的界面以及控件,不应该构成对本申请实施例的限定。
下面基于图2-图7对第一次录音及其涉及的相关内容进行示例性描述。
前述涉及的笔记1中还没有记录用户输入的信息。电子设备进入笔记1之后,可以通过录音控件触发开启录音功能。在录音功能开启之后,电子设备可在用户界面中划分出两个区域(区域1以及区域2)。区域1用于呈现与录音相关的内容,区域1也可以被称为录音区域。区域2中用于呈现用户输入的其他信息(例如文字、图片等)。该其他信息可以被称为非录音信息,区域2也可以被称为编辑区域。
其中,录音区域(区域1)中还可以包括两个子区域:语音录播区域(记为区域11)以及语音转文本区域(记为区域12)。关于区域11以及区域12的相关描述如下。
语音录播区域(区域11)中可以包括录播(录制或者播放)语音时涉及的功能控件,例如录音条。该录音条可以用于触发录制语音、暂停录制语音、播放语音、暂停播放语音等功能。该录音条具有不同的状态,不同状态下的录音条由不同组件组成。录音条的状态不同,则录音条的功能不同,具体可以参考下文中相关内容的描述,此处暂不赘述。
语音转本文区域(区域12)中可以包括语音转文本内容,还可以包括转换控件。该转换控件为电子设备提供的控制语音转文本功能开启或者关闭的控件。其中语音转本内容包括录制的语音转换得到的文字。
应该理解的是,前述区域11中包括录播(录制或者播放)语音时涉及的功能控件除了录音条以外,还可以包括其他控件,例如包括转换控件等。本申请实施例对此不作限定。
应该理解的是,这里以区域11以及区域12显示在一个区域(区域1)中为例进行说明,在其他可能的情况下,该区域11以及区域12可以分开。
编辑区域(区域2)中用于呈现用户输入的其他信息(非录音信息),例如文字、图片等。电子设备可以提供输入工具用于输入信息。其中,输入工具可以包括键盘以及手写笔工具栏。
在一些可能的情况下,电子设备在一个用户界面中可以显示一个输入工具,显示该输 入工具时,可以提供切换到其他输入工具的快捷控件。
这样,一些可能的情况下,电子设备在录音时,用户可以在区域11中通过录音条查看录音的情况。在录音的同时,电子设备还可以实现语音转文本,将转换得到的文字显示在区域12中供用户查看。在录音的同时,用户还可以通过输入工具实现手动输入,并将输入的信息显示在区域2中供用户查看。手动输入包括输入文字、***图片等。
关于区域1、区域2以及不同输入工具的相关内容可以参考下述对图2的示例性描述。
如图2中(1)所示,用户界面13为电子设备进入笔记后显示的一个示例性用户界面。此时,电子设备默认提供的输入工具为键盘。例如,用户界面13中可以显示键盘131。显示键盘时,电子设备还可以显示笔记工具栏(例如笔记工具栏132)。
笔记工具栏可以在用户记录信息时提供便捷功能(例如录音功能)入口。笔记工具栏132中可以包括***控件132a、录音控件132b、“手写”控件132c。其中,***控件132a可以用于触发将图片、视频等信息输入用户界面13中进行显示。录音控件132b可以用于触发开启录音功能。“手写”控件132c可以用于触发将输入工具从键盘131切换到手写笔工具栏。该笔记工具栏中除了可以包括***控件132a、录音控件132b以及“手写”控件132c以外,还可以包括其他的控件,例如还可以包括清单建立控件132d以及样式设置控件132e。
用户界面13中还可以包括编辑区域133,该编辑区域133中可以用于显示用户输入的非录音信息。例如,用户可以通过键盘131将文字信息输入到该编辑区域133中进行显示,还可以通过***控件132a***图片等信息输入到编辑区域133中进行显示。
响应于针对“手写”控件132c的操作(例如点击操作),电子设备将输入工具从键盘131切换到手写笔工具栏。如图2中(2)所示,该手写笔工具栏可以如用户界面13a中显示的手写笔工具栏134。该手写笔工具栏134中可以包括不同的手写笔,例如手写笔134a-手写笔134f。手写笔可以用于在编辑区域中做标记,例如,添加下划线,画圈等。该用户界面13a中还可以包括键盘切换控件135。该键盘切换控件135可以用于触发电子设备将输入工具从手写笔工具栏切换到键盘。
在一些可能的情况下,该键盘切换控件135可以作为手写笔工具栏134中的一个快捷控件。
响应于针对状态A1的录音控件132b的操作(例如点击操作),电子设备可以开启录音功能,并且,在用户界面中划分出前述涉及的区域1以及区域2。此时,电子设备可以将录音控件132b的状态从状态A1切换至状态B1。状态B1的录音控件132b可以表示开启了录音功能,状态A1的录音控件132b可以表示录音功能未开启。前述用户界面13中录音控件132b的状态即为状态A1,状态B1是与状态A1不同的另外一种状态。如图2中(3)所示,用户界面13b中录音控件132b被置灰,即置灰状态为一种状态B1。响应于针对状态B1的录音控件132b的操作(例如点击操作),电子设备可以关闭录音功能,且将录音控件132b的状态从状态B1更新至状态A1。
参考图2中(3)所示的用户界面13b,在录音控件132b的状态为状态B1时,电子设备可以在用户界面中划分出区域1以及区域2,这里可以理解为区域2为新的编辑区域。
其中,区域1可被划分为区域11以及区域12。
该区域11中可以包括状态1下的录音条(也可以被称为录音条1),该录音条1中可以包括暂停录音控件136a、录音状态条136b、录制语音时对应的录制时长136c,该录制时长136c可以为录制的全部语音的总时长。暂停录音控件136a用于表示当前终端已经开始录音。录音状态条136b可以反映电子设备采集到的声音信号的能量大小,能量越大,则该录音状态条136b可以产生较大程度的变化。能量越小,则该录音状态条136b产生的变化程度越小。录制时长136c可以用于表示录制的全部语音的总时长。
该区域12中可以包括状态C下的转换控件137a,该状态C下的转换控件137a可以用实线表示。状态C下的转换控件137a可以表示电子设备开启了语音转文本功能。区域12未显示语音转文本得到的文字信息,且,语音转文本功能开启的情况下,电子设备还可以显示提示信息137b,该提示信息137b可以用于提示电子设备开启了语音转文本功能。例如,该提示信息137b可以为提示文字:“正在语音转文本”。在语音转文本功能开启的情况下,电子设备可以将采集得到的声音信号进行处理得到语音,并且,对该语音进行转换,将转换得到的文字显示在区域12中。
在另一些可能的情况下,录制时长136c可以不用于表示录制的全部语音的总时长。可以用于表示正在录制的这段语音的总时长。
区域2中用于呈现输入的非录音信息,例如文字、图片等。在区域2中未显示非录音信息时,可以显示提示信息138a,该提示信息138a可以用于提示用户:正在录音且同时可以在区域2输入信息。该提示信息138a可以为文字提示:“正在录音,您可以在空白处书写内容或者涂鸦”。
在开始录音,且,语音转本文功能开启的情况下,电子设备在录音的同时,还可以实现语音转文本,将转换得到的文字显示在区域12中供用户查看。在录音的同时,用户还可以通过输入工具实现手动输入,并将输入的信息显示在区域2中供用户查看。手动输入包括输入文字、***图片等。如图2中(4)的用户界面13c所示,开始录音之后,电子设备可以采集用户或者环境中的声音信号,并且将该声音信号转化为语音,该声音信号的内容为:“笔记中包括两个部分”。同时,可以将该语音转换成文字显示在区域12中,以及,接收用户输入的非语音信息。例如,该非语音信息可以是用户通过键盘131或手写笔工具栏134中的工具输入到区域2中进行显示的,该非语音信息可以包括已输入到区域2中的文字信息和信息139,该信息139为:“内容”。
电子设备还可以设置关闭语音转文本功能。这样,在开始录音之后,可以不将录制的语音转换为文字信息,但是可以实现在录音时接受用户输入的非录音信息。
关闭语音转文本功能时涉及的内容可以参考下述对图3的示例性描述。
如图3中(1)的用户界面13b所示,响应于针对转换控件137a(状态C)的操作(例如点击操作),电子设备可以关闭语音转本文功能,且,将转换控件137a的状态从状态C更新为状态D,状态D不同于状态C。在区域12中显示提示信息137b的情况下,表示电子设备没有获取到语音转文本的内容。此时,响应于针对转换控件137(状态C)的操作(例如点击操作),电子设备可以不再显示提示信息137b,而是显示提示信息141。该提示信息141用于通知用户关闭了语音转文本功能。
如图3中的(2)所示,用户界面14a为关闭语音转本文功能时涉及的一个示例性用户界面。区域12中可以显示提示信息141用于表示电子设备关闭了语音转文本功能。例如,该提示信息的提示内容可以为:“语音转文本已暂停”。此时,转换控件137a(状态D)可以用虚线表示。在一些可能的情况下,状态D除了可以用实线表示以外,还可以表示为其他的形式,例如改变颜色、置灰等,本申请实施例对此不作限定。
在一些实施例中,录音功能开启后,但是在暂停录音的情况下,如果区域12中没有包括语音转文本得到的文字,则电子设备可以不再显示区域12。则此时,区域1中可以只包括区域11。例如,在用户界面14a中,响应于针对暂停录音控件136a的操作,电子设备可以暂停录音,同时显示图3中(3)所示的用户界面14b。其中,状态2即为录音功能开启后,但是在暂停录音的状态。
如图3中(3)所示,用户界面14b为暂停录音之后涉及的一个示例性用户界面。此时,区域1中只包括录音条2。在一些可能的情况下,该录音条2中可以包括开始录音控件141a、开始播放控件141b、播放进度条141c、播放时间指示符141d以及语音菜单控件141e。其中,开始录音控件141a可以用于触发电子设备再次开始录音。开始播放控件141b可以用于触发电子设备播放已录制的语音。播放进度条141c可以用于指示当前播放语音的进度。播放时间指示符141d用于指示当前播放时间以及全部语音的总时长。语音菜单控件141e用于触发电子设备显示以及管理录制的全部语音,该全部语音包括在笔记1中记录的全部语音。在笔记1中记录的全部语音为Q(大于等于2的整数)段语音时,电子设备可以将该Q段语音整合成一段长语音,该Q段语音的总时长即为该长语音的总时长。这时,播放进度条141c中还可以包括Q-1个节点,将播进度条划141c分为Q段(可以被称为Q段子进度条)。其中,Q段子进度条中的第i段子进度条对应第i段语音。对于节点及其相关内容的描述可以参考下述对图12中(4)的描述。此处暂不赘述。
应该理解的是,播放进度条141c可以用于指示当前播放语音的进度:该播放进度条141c中可以包括标识1411,该标识1411在播放进度条141c的位置A1,则表示播放位置A1所对应的语音。
这里应该理解的是,这里描述在图3中(2)示出的用户界面14a中响应于针对暂停录音控件136a的操作,即可暂停录制语音,显示图3中(3)所示的用户界面14b的目的在于表示录音过程可以暂停,暂停时如果区域12中未显示语音转换后的文字即可关闭区域12。实际,在本申请的实施例中此处并未结束第一次录制语音的过程。一种可能的显示方式包括:在显示图3中(2)示出的用户界面14a之后,用户未操作用户界面14a中的控件136a,因此,电子设备继续录音,显示图3中(4)示出的用户界面14c。
在一些可能的情况下,电子设备在录制Q段语音中的第i段语音时,该第i段录音对应的起始录制时长可以为第i-1段语音的结束录制时长,或者,第i-1段语音的结束录制时长+1。则电子设备完成录制第Q段语音时录制时长136c对应显示的时间即为Q段语音的总时长。例如,参考图3中(3)以及图3中(4)所示,如图3中(3)的用户界面14b所示,电子设备录制的语音对应的结束录制时长为2s。点击开始录音控件141a录制下一段语音时,图3中(4)示出的录制时长136c中指示起始录制时长为3s,也可以为2s。本申请实施例对此不作限定。
在另一些可能的情况下,电子设备在录制Q段语音中的第i段语音时,该第i段录音对应的起始录制时长可以从第0s开始。此时,录制时长136c可以不用于表示录制的全部语音的总时长。可以用于表示正在录制的这段语音的总时长。
应该理解的是,在笔记1中记录的全部语音为一段语音时,则该播放进度条141c中不包括节点。例如,如图3中(3)的用户界面14b所示:由于此时只完成了第一段语音的录制,因此,此时播放进度条141c中不包括节点。
还应该理解的是,图3中(3)所示的录音条2只是示例性说明,实际应用中该录音条2还可以包括更多或者更少的组件,各组件的形态可以与图中示出的相同,也可以不同,本申请实施例对此不作限定。
响应于针对开始录音控件141a的操作(例如点击操作),可以触发电子设备录制一段新语音,且将在区域1中划分区域11以及区域12。此时,电子设备可以显示如图3中(4)所示的用户界面。
如图3中(4)所示,用户界面14c为电子设备录制一段新语音时涉及的示例性用户界面。用户界面14c中可以包括区域1,该区域1中包括两个子区域(区域11以及区域12)。此时,区域11中可以包括状态1下的录音条(录音条1),用于提示用户当前电子设备正在录制语音。
在一些可能的情况下,如果在检测到针对开始录音控件141a操作之前,电子设备的语音转文本功能是关闭的,则在检测到针对开始录音控件141a操作之后,电子设备的语音转文本功能仍然是关闭的。例如,用户界面14c中,在电子设备开始录制一段新语音时,区域12中的转换控件为虚线状态(一种状态D),则表示语音转文本功能是关闭的。如果区域12中不包括语音转换得到的文字,则电子设备还可以显示提示信息141以用于提示用户关闭了语音转文本功能。响应于针对状态D下的转换控件137a的操作(例如点击操作),电子设备可以开启语音转文本功能。
在一些可能的情况下,语音转本文区域(区域12)的大小是可以调节的,该区域12的尺寸可以在尺寸1以及尺寸2之间进行切换,其中,尺寸1小于尺寸2。且,该语音转文本区域是支持滑动显示的。最新录制的语音转换得到的文字可以显示在区域12中全部文字的最后。该语音转文本区域中最大程度能显示的完整字符为W个,在语音转文本区域中显示了W个字符的情况下,每增加E个新字符就可以有E个原字符(已经显示的)不显示在或者不能完全显示在语音转文本区域中。其中,W为大于等于1的整数,E为小于等于W的整数。语音转文本区域的大小为尺寸1以及尺寸2时W的取值不同,为尺寸2是可以显示更多的语音转文本内容。这里的语音转文本内容可以包括语音转本文得到的文字以及电子设备添加的标点符号。
关于语音转文本区域(区域12)的大小变化以及文字显示的相关内容可以参考下述对图4以及图5的描述。
如图4中(1)所示,在用户界面15a中,电子设备可以录制语音,例如此时录制的语音内容可以为:“第一部分主要讲述了XXX”。并且,电子设备可以将该语音转换得到的文字显示在语音转文本区域(区域12)中。此时,用户界面15a显示了大小为尺寸1的一 个示例性区域12。该区域12中可以完整的显示一排字符,以及不完整的显示一排字符。在一些可能的情况下,电子设备可以在不完整显示的一排字符上添加一个蒙层,使得该不完整显示的一排字符“若隐若现”。此时,区域12中始终可以完整显示一排字符,每增加E个新字符,可以完整的将其显示在该区域12中,且,将最靠前的E个原字符(已经显示的)向上滑动一排进行不完整显示。
大小为尺寸1的区域12中还可以包括展开控件151a以及滑动控件151b。展开控件151a可以用于将区域12的尺寸从尺寸1切换到更大的尺寸2。滑动控件151b可以用于指示该区域12是支持滑动显示的。
响应于针对展开控件151a的操作(例如点击操作),电子设备可以将区域12展开,尺寸变更为尺寸2。此时,电子设备可以显示如图4中(2)所示的用户界面。
如图4中(2)所示,用户界面15b为展开区域12之后的示例性用户界面。大小为尺寸2的区域12中可以包括收起控件151c。该收起控件151c可以用于触发将区域12的大小从尺寸2变更为尺寸1。
在一些可能的情况下,在区域12的大小为尺寸2的情况下,该区域12的大小可以设置为手机屏幕的1/3的大小,或者其他大小,本申请实施例对此不作限定。
如图4中(2)所示,电子设备继续采集声音信号,该声音信号的内容为:“第二部分主要讲述了XXXX”,然后,通过该声音信号得到语音,并将该语音转换为文字(文字1)显示在区域12(大小为尺寸2)中。后续,电子设备可以继续将录制的语音转换为文字显示在区域12中,例如,可以参考图4中(3)示出的用户界面15c。
如图4中(3)所示,电子设备继续采集声音信号,该声音信号的内容为:“本次,先讲解第一部分的”,然后,通过该声音信号得到语音,并将该语音转换为文字(文字2)显示在区域12(大小为尺寸2)中。此时,文字2显示在前述涉及的文字1中的后面。此时,区域12中显示文字2后,相比于未显示文字2之前,区域12中包括的文字新增加一行,则滑动控件151b的位置可以向下移动。该滑动控件151b的位置可以对应新增加字符的位置:新增加的字符在区域12中越靠后,则滑动控件越靠下。
在一些可能的情况下,区域12中显示的文字为录制时间靠后的语音转换得到文字,区域12可以支持滑动显示,将录制时间靠前的语音转换得到的文字进行显示。如图5中(1)所示,电子设备继续采集声音信号并将该声音信号得到语音转换成为文字显示在区域12中。例如,该声音信号的内容为:“后续,再详细讲述”。响应于在区域12中的操作1(例如向下滑动操作),电子设备可以显示时间靠前的语音转换得到文字。在检测到操作1之前,该时间靠前的语音转换得到的文字可以不显示在区域12中。例如,此时,电子设备可以显示图5中(2)所示的用户界面15e。
这里应该理解的是,将区域12中录制时间靠前的语音转换得到的文字进行显示的方式,除了图5中(1)示出的针对区域12的操作1,还可以是其他操作,例如还其他操作可以包括:向上拖动滑动控件151b。本申请实施例对此不作限定。
如图5中(2)、图5中(3)所示,检测到操作1之后,在预设时间(例如1s)内没有再次检测到针对区域12的操作,则电子设备可以在区域12中继续显示最新录制的语音 转换得到文字。此时电子设备显示的用户界面可以从用户界面15e更新为图5中(3)所示的用户界面15f。
如图5中(3)所示,电子设备可以继续采集声音信号,基于该声音信号得到语音,电子设备可以在区域12中继续显示最新录制的语音转换得到文字。例如,这里声音信号的内容可以为:“第二部分的内容”。
应该理解的是,前述内容中,区域12中可以对字符进行不完整显示,且对不完整显示的字符添加蒙层。在一些可能的情况下,区域12中可以只显示能完整显示的字符,也可以不添加蒙层,该情况可以在后续附图中的区域12中进行体现,此处不再赘述。
在一些可能的情况下,语音转文本区域(区域12)中除了支持滑动显示文字以外,还可以支持其他的功能。例如可以对区域12中的文字进行处理,例如编辑处理(包括复制、粘贴等)以及分享等。如图5中(1)示出的文字信息152即是将区域12中的文字进行复制、粘贴之后得到的。
录音过程中,对于区域12中的文字进行编辑处理的过程可以参考下述对图6以及图7中相关内容的描述。
这里可以将语音转文本区域(区域12)理解为是一个可编辑区域,可以对其中的文字进行编辑处理,这里涉及的编辑包括但不限于以下内容中的一个或者多个:复制、剪切、全部选择(权选)、***文字等。其中,***文字包括粘贴文字至区域12或者通过键盘输入文字然后在区域12中进行显示。
如图6中(1)所示的用户界面16a,响应于针对区域12中位置1的操作2(例如长按操作),电子设备可以选中最靠近位置1的U个文字,并且,可以显示处理工具栏,该处理工具栏中可以包括对被选中文字进行处理时涉及的功能控件。此时,电子设备可以显示如下述图6中(2)所示的用户界面。
如图6中(2)所示,在用户界面16b中,6个文字(“笔记1中包括”)被选中,电子设备可以为被选中的字符添加选中标记162,该选中标记162的一端包括控件163,以及另一端包括控件164。其中,控件163以及控件164可以用于触发调节选中标记162对应的选择区域,即控件163以及控件164可以用于设置被选中字符的数量以及位置。该用户界面16b中还可以包括处理工具栏161,该处理工具栏161中可以包括全选控件161a、剪切控件161b、复制控件161c、粘贴控件161d以及分享控件161e。其中,全选控件161a可以用于触发将区域12中显示的字符全部选中,并显示包括剪切控件161b、复制控件161c、粘贴控件161d以及分享控件161e的新的工具栏。剪切控件161b可以用于触发将被选中的字符进行复制,且,在复制之后进行删除。复制控件161c可以用于触发将被选中的字符进行复制。粘贴控件161d可以用于触发利用其它字符(已复制)替换被选中的字符。
响应于针对控件164的操作,例如向右滑动该控件164到位置A的操作,电子设备可以将控件164显示在位置A,则选中标记162中包括的字符为控件163到控件164(显示在位置A处)之间的全部字符。此时,电子设备可以显示图6中(3)所示的用户界面16c。
如图6中(3)所示,选中标记162中包括的字符(被选中的字符1)更新为:“笔记1中包括两个部分,其中,第一部分主要讲述了XXXXXX。”,即为前述涉及的文字信息 152。此时,控件163的位置没有发生变化,控件164的位置从用户界面16b中的位置更新为用户界面16c中示出的位置(可以看作位置A的一种情况)。
响应于针对处理工具栏161中复制控件161c的操作(例如点击操作),电子设备可以对被选中的字符1进行复制。后续可以将该被选中的字符1粘贴到区域12或者区域2中。例如,可以参考对下述内容的描述。
如图6中(4)所示的用户界面16d,响应于针对区域2中光标165所在位置的操作(例如长按操作),电子设备可以显示如图7中(1)所示的用户界面16e。用户界面16e中包括粘贴控件166,该粘贴控件166可以应用触发将复制的字符(例如被选中的字符1)粘贴到区域2中进行显示。
响应于针对粘贴控件166的操作,电子设备可以将复制的字符(例如被选中的字符1)粘贴到区域2中。此时,电子设备可以显示如图7中(2)所示的用户界面16f。例如,用户界面16f中的文字信息152即为粘贴的字符。
应该理解的是,在第一次录音的过程中,响应于针对状态C下的转换控件137a的操作(例如点击操作),电子设备可以关闭语音转文本功能,同时将转换控件137a的状态从状态C(例如实线表示)变更为状态D(例如虚线表示)。如用户界面16f所示,如果区域12中包括语音转文本得到的文字,则该区域12中可以不再显示提示信息141。
如图7中(2)所示,响应于针对暂停录音控件136a的操作(例如点击操作),电子设备可以结束建立笔记1之后的第一次录音。
结束一次录音之后,电子设备可以以尺寸1显示区域12。例如,参加图7中(3)所示的用户界面16g。也可以以尺寸2显示区域12,例如参考图7中(4)所示的用户界面16h。本申请实施例对此不作限定。
应该理解的是,响应于针对暂停录音控件136a的操作(例如点击操作),电子设备可以结束录音,还可以结束语音转文字。也就是说,不论在暂停录音之前,语音转文字的功能是否可用,在暂停录音的情况下,语音转文字功能不可用。此时,响应于针对暂停录音控件136a的操作(例如点击操作),电子设备还可以将转换控件137a的状态设置为状态D(例如虚线表示),以指示语音转功能不可用。暂停录音的情况下,该转换控件137a可以不接收用户的操作。
在一些可能的情况下,在暂停录音之后,电子设备可以对区域12中的文字进行处理,例如编辑处理(包括复制、粘贴等)以及分享等。
暂停录音的情况下,对于区域12中的文字进行编辑处理的过程可以参考下述对图8中相关内容的描述。
参考图8中(1)所示的用户界面17a,电子设备可以在区域12中选中部分或者全部字符(被选中的字符2),为被选中的字符添加选中标记162,并且,显示处理工具栏161对被选中的字符进行处理。例如,被选中的字符可以包括:“第二部分主要讲述了XXXXXX。”。关于选中标记162、处理工具栏161以及电子设备选中字符的过程可以参考前述对图6中(1)以及图6中(2)相关内容的描述,此处不再赘述。
响应于针对处理工具栏161中复制控件161c的操作(例如点击操作),电子设备可以 对被选中的字符2进行复制。后续可以将该被选中的字符2粘贴到区域12或者区域2中。例如,可以参考对下述内容的描述。
参考图8中(2)所示的用户界面17b,响应于针对区域2中光标165所在位置的操作(例如长按操作),电子设备可以显示如图8中(3)所示的用户界面17c。
如图8中(3)所示,用户界面17c中包括粘贴控件166,该粘贴控件166可以用于触发将复制的字符(被选中的字符2)粘贴到区域2中进行显示。
响应于针对粘贴控件166的操作,电子设备可以将复制的字符(被选中的字符2)粘贴到区域2中。此时,电子设备可以显示如图8中(4)所示的用户界面17d。例如,用户界面17d中的文字信息171即为粘贴的字符。
如图8中(4)所示,在第一次录音完成之后,响应于针对开始录音控件141a的操作,电子设备可以进行第二次录音。关于第二次录音的相关描述可以参考下述对图9-图12的描述。
然后,电子设备可以将第一次录音与该第二次录音制作成一段长语音。在停止录制第二次录音之后,可以在播放进度条中显示一个节点1,利用该节点将播放进速条划分为两段子进度条,节点1之前的子进度条对应第一次录音得到的语音,节点1之后的子进度条对应第二次录音得到的语音。
应该理解的是,这里是以第一次录音以及第二次录音得到长语音为例进行讲解。实际情况下,电子设备还可以进行第三次录音、第四次录音等。这里将录音次数记为Q。第Q次录音完成之后,播放进度条中可以增加一个节点。第Q次录音完成之后,播放进度条中共计Q-1个节点,将播进度条划分为Q段(可以被称为Q段子进度条)。Q段子进度条中的第i段子进度条对应第i段语音。其中,Q为大于等于2的整数。关于后续第三次录音、第四次录音等录音的过程可以参考前述对第一次录音以及后续对第二次录音的描述,此处不再赘述。
还应该理解的是,在录音过程中以及录音暂停后,电子设备都可以在区域2中输入非录音信息。输入非录音信息的方式除了包括前述涉及的通过键盘输入以及复制粘贴以外。还可以通过其他方式在区域2中进行编辑,包括:***图片到区域2中、通过手写笔工具栏在区域2中进行编辑(例如涂鸦、手写输入文字)等。通过其他方式在区域2中进行编辑的相关内容可以在第二次录音时进行描述,具体可以参考下述对图9-图12的相关描述。
第二次录音涉及的部分过程以及在区域2中***图片的过程可以参考下述对图9以及图10的描述。
如图9中(1)所示,用户界面18a为第二次录音过程中涉及的一个示例性用户界面。此时,区域12中显示的文字为第一次录音得到的语音所转换的。电子设备在区域12中显示第二次录音得到的语音转换得到的文字时,可以采取不同的方式进行显示,包括但不限于以下显示方式。
显示方式1:将第一次录音得到的语音所转换得到的文字与第二次录音得到的语音所转换得到的文字进行连续显示。即,假设第一次录音完成之后,第二次录音开始之前,区域12中包括的最后一个字符显示在区域12中全部字符的第h行。在该最后一个字符不是 第h行的最后一个字符时,电子设备可以将第二次录音得到的语音所转换得到的第一个文字显示在第h行,且显示在该最后一个字符之后。在这样的情况下,该第一个文字(第二次录音时得到的)与该最后一个字符(第一次录音时得到的)之间可以没有空格。在该最后一个字符是第h行的最后一个字符时,电子设备可以将第二次录音得到的语音所转换得到的第一个文字显示在该最后一个字符的下一行。其中,显示方式1涉及的一个示例性用户界面可以参考下述对图9中(2)的描述。
如图9中(2)所示,用户界面18b中包括文字信息181以及文字信息182,其中文字信息181为第一次录音得到的语音所转换的。文字信息182为第二次录音得到的语音所转换的。例如,该文字信息182的内容可以为:“第一部分从下图开始讲解”。此时,文字信息181中的最后一个字符与文字信息182中的第一个字符显示在同一行,且连续显示,中间没有空格。
显示方式2:将第一次录音得到的语音所转换得到的文字与第二次录音得到的语音所转换得到的文字分开进行显示。例如,假设第一次录音完成之后,第二次录音开始之前,区域12中包括的最后一个字符显示在区域12中全部字符的第h行。在该最后一个字符不是第h行的最后一个字符时,电子设备可以将第二次录音得到的语音所转换得到的第一个文字显示在该最后一个字符的下一行。其中,显示方式2涉及的一个示例性用户界面可以参考下述对图9中(3)的描述。如图9中(3)所示,用户界面18c中也包括前述涉及的文字信息181(第一次录音得到的)以及文字信息182(第二次录音得到的)。此时,文字信息181中的最后一个字符与文字信息182中的第一个字符显示在不同行:文字信息182中的第一个字符显示在文字信息181中的最后一个字符的下一行。
显示方式2的另外一个可能的实现方式,在该最后一个字符是第h行的最后一个字符时,电子设备可以生成一个空白行(该空白行可以不用于显示任何字符),然后将第二次录音得到的语音所转换得到的第一个文字显示在该空白行的下一行,或者,电子设备也可以在第h行的下一行空两格,然后显示对第二次录音的语音转换得到的第一个文字。
后续内容以显示方式1为例对本申请实施例涉及的其他内容进行描述。例如,对在区域2中***图片的过程进行描述。
如图9中(2)所示,响应于针对***控件132a的操作,电子设备可以显示至少一个***方式标识或者至少一种可***内容标识。此时,电子设备可以显示如图10中(1)所示出的用户界面18d。
如图10中(1)所示,用户界面18d中可以包括多种***方式标识以及多种可***内容标识。例如,可***内容标识中可以包括拍摄图标183a、表格图标183c、日程图标183d、卡证扫描图标183e、导图图标183g以及录音图标183h。***方式标识中可以包括“从图库中选择”图标183b、“文件扫描”图标183f。其中,拍摄图标183a可以用于触发打开相机,拍摄图片,然后将拍摄的图片***到区域2中。表格图标183c可以用于向区域2中***表格。日程图标183d可以用于向区域2中***日程标识。“卡证扫描”图标183e可以用于触发打开扫描功能,对卡证进行扫描,将得到的扫描内容显示在区域2中。导图图标183g可以用于向区域2中***思维导图。录音图标183h可以用于向区域2中***录制 完成的语音。“从图库中选择”图标183b可以用于触发打开图库,然后从图库中选择图片或视频***到区域2中。“文件扫描”图标183f可以用于触发打开扫描功能,对文件进行扫描,将得到的扫描内容显示在区域2中。
应该理解的是,用户界面18d中示出的***方式标识以及可***内容标识只是示例性说明,实际可以包括其他的标识,还可以包括比图中更多或者更少的标识。标识的形态除了可以是图标,还可以是其他的形态,例如文字等。本申请实施例对此不作限定。
响应于针对“从图库中选择”图标183b的操作(例如点击操作),电子设备可以打开图库,从图库中选择图片或视频***到区域2中,显示如图10中(2)示出的用户界面。此时,电子设备同时还可以采集声音信号得到语音,并将该语音转换为文字显示在区域11中。例如,该声音信号的内容可以为:“图中”。
如图10中(2)所示,用户界面18e为打开图库之后的一个示例性界面。响应于针对图片1的操作(例如点击操作),电子设备可以将该图片1***到区域2中。此时电子设备可以显示如图10中(3)所示的用户界面18f。
如图10中(3)所示,区域2中可以显示图片1以提示用户已完成***图片1。此时,电子设备同时还可以采集声音信号得到语音,并将该语音转换为文字显示在区域11中。例如,该声音信号的内容可以为:“记录了”。
响应于针对“手写”控件132c的操作(例如点击操作),电子设备可以将输入工具从键盘切换到手写笔工具栏。后续可以通过手写笔工具栏在区域2中进行编辑(例如涂鸦、手写文字等)。
第二次录音涉及的部分过程以及通过手写笔工具栏在区域2中进行编辑(例如涂鸦)的过程可以参考下述对图11以及图12的描述。
参考图11中(1)所示的用户界面19a,响应于针对手写笔工具栏134中手写笔134b的操作(例如点击操作),电子设备可以触发通过手写笔134b在区域2中进行编辑(例如涂鸦),且,电子设备可以突出显示手写笔134b或者改变手写笔134b的状态以指示该手写笔134b被选中。此时,电子设备可以显示图11中(2)所示的用户界面。
参考图11中(2)所示的用户界面19b,手写笔134b突出显示以指示该手写笔134b被选中。响应于在区域2中利用手写笔134b进行涂鸦的操作,电子设备可以显示如图11中(3)所示的用户界面。
如图11中(3)所示,用户界面19c中可以包括下划线191,该下划线191为利用手写笔134b在区域2中添加的。
应该理解的是,手写笔工具栏中可以包括至少一个手写笔,不同手写笔的类型不同,例如,可以包括记号笔1、记号笔2、擦除笔、铅笔等,通过不同手写笔在区域2中进行编辑(例如涂鸦)时,得到的结果不同。也可以理解为不同手写笔进行编辑时,绘制的线条类型不同。其中,线条的类型不同可以包括颜色、形状不同。例如通过记号笔1可以绘制形状1的线条(例如直线),通过记号笔2可以绘制形状2的线条(例如曲线)等。不同手写笔所编辑出的线条可以调节,例如线条的颜色、粗细等可以调节。
响应于针对键盘切换控件135的操作(例如点击操作),电子设备可以将输入工具从手写笔工具栏切换到键盘。后续,电子设备在进行第二次录音的同时,继续在区域2中输 入非录音信息,然后,结束第二次录音。
如图12中(1)所示,用户界面20a中包括图片2,这里假设图片2显示在图片1之后,对于***图片2的过程可以参考前述对***图片1的描述,此处不再赘述。此时,电子设备可以继续采集声音信号以得到语音,该声音信号的内容可以为:“另外,第一部分XX…”。在录音的同时,电子设备可以进行语音转文本并且将转换得到的文字显示在区域12中。并且,在录音的同时,电子设备还可以在区域2中输入非语音信息。例如,响应于针对编辑文字“图片”的操作(例如点击操作),电子设备可以将该编辑文字“图片”输入到区域2中进行显示。此时,电子设备可以显示如图12中(2)所示出的用户界面20b。
后续电子设备可以结束第二次录音。参考图12中(3)所示出的用户界面20c,响应于针对暂停录音控件136a的操作(例如点击操作),电子设备可以结束第二次录音。假设此时,第二次录音得到的语音时长为12s,则录制时长体现为23s,可以看作语音1的时长(11s)与语音2的时长(12s)之和。在第二次录音结束之后,电子设备可以显示如图12中(4)所示的用户界面。
如图12中(4)所示,用户界面20d中可以包括状态2下的录音条(录音条2)。第二次录音结束之后,电子设备可以将第一次录音得到的语音(语音1)以及第二次录音得到的语音(语音2)制作成一段长语音(长语音1)。这里,长语音1的时长可以表示为语音1加语音2的时长。例如,如用户界面20d所示,播放时间指示符141d中可以看出,长语音1的时长为23s,可以看作语音1的时长(11s)与语音2的时长(12s)之和。
用户界面20d中还可以包括播放进度条141c,该播放进度条141c中可以包括节点201(可以看作一种节点1)。该节点201可以将该播放进度条141c划分为两个子进度条。节点201之前的子进度条对应第一次录音得到的语音(语音1),节点201之后的子进度条对应第二次录音得到的语音(语音2)。
后续,电子设备可以将记录了信息的笔记1进行保存。在保存之后,用户可以通过电子设备对该笔记1进行查看。
其中,电子设备对笔记1进行保存的内容可以参考下述描述。
参考图12中(4)示出的用户界面20d,响应于针对完成控件401的操作(例如点击操作),电子设备可以保存笔记1。
在一些可能的情况下,在完成保存笔记1之后,电子设备还可以显示提示信息以提示用户保存成功。例如,如图13所示,用户界面41中可以包括提示信息411,该提示信息411用于提示用户笔记1保存成功。
应该理解的是,前述内容是以电子设备在建立笔记1之后录制了两次语音为例进行说明的。实际情况中,电子设备在建立笔记1之后,还可以录制更多或者更少的语音,本申请实施例对此不作限定。在笔记1中记录信息之后,用户还可以通过电子设备查看笔记1中记录的信息。对于查看信息时涉及的内容可以参考下述对图14-图17的描述。
下面对查看信息涉及的相关内容进行描述。
图14-图17中描述了查看信息涉及的一组示例性用户界面。
在保存笔记1之前,以及保存笔记1之后,用户都可以通过电子设备查看笔记1中记录的信息。查看笔记时,电子设备可以通过播放语音使得用户查看录音内容。电子设备在播放语音时,可以实现音文同步。
在一些可能的情况下,音文同步可以包括下述内容中的一个或者多个:
同步方式1.在播放录音时,以状态A显示区域2(编辑区域)中已播放语音对应的非录音信息(记为信息A),并且,以状态B显示区域2中未播放语音对应的非录音信息(记为信息B)。这里以电子设备在播放语音A(录制时间为时间A)为例对信息A以及信息B进行描述:信息A是指输入时间早于或者等于时间A的信息。信息B是指输入时间晚于时间A的信息。其中,时间A也可以被称为语音A对应的时间戳。信息A的输入时间可以被称为信息A对应的时间戳。信息B的输入时间可以被称为信息B对应的时间戳。其中,状态A与状态B不同。例如,状态B可以为在信息B上添加蒙层,或者使得信息B半透明显示,使得该信息B“若隐若现”。以指示在录制当前播放的语音之前,该信息B还没有输入到笔记中。状态A可以为不在信息A上添加蒙层,以指示在录制当前播放的语音之前,该信息A已经输入到笔记中了。
同步方式2.在播放录音时,区域12(语音转文本区域)中可以显示已播放语音转换得到的文字。
图14-图17中以音文同步包括前述涉及的同步方式1以及同步方式2为例进行说明。实际应用中音文同步可以只包括同步方式1或者同步方式2本申请实施例对此不作限定。
下面基于图14-图16对音文同步涉及的相关内容进行示例性描述。
如图14中(1)所示,用户界面30为录音功能开启后,暂停录音的情况下涉及的一个示例性用户界面,该用户界面30中可以包括状态2下的录音条(录音条2)。其中,状态2即为录音功能开启后,但是在暂停录音的状态。
应该理解的是,图14中(1)显示的手写笔工具栏134可以是通过控触“手写”控件132c切换得到的。电子设备还可以通过键盘切换控件135将手写笔工具栏134切换至键盘。
响应于针对录音条2中开始播放控件141b的操作(例如点击操作),电子设备可以开始播放录制得到的语音(长语音)。在播放语音的过程中,电子设备可以实现音文同步。例如,播放到第3s时,电子设备可以实现音文同步。此时,电子设备可以显示如下述图14中(2)所示的用户界面。
参考图14中(2)所示的用户界面31,播放到第3s对应的语音(语音A1)时,电子设备实现音文同步可以体现在以下两个方面。
第一方面,语音转文本区域(区域12)中可以显示文字信息311,该文字信息311中包括录制语音A1时转换得到的文字。电子设备在录制语音A1(第3s对应的语音)时涉及的相关内容可以参考前述对图4中(1)所示出的用户界面15a的描述,此处不再赘述。
第二方面,编辑区域(区域2)中的非录音信息312a上没有添加蒙层。非录音信息312b上添加了蒙层。也可以理解为:区域2中显示非录音信息312a的区域没有添加蒙层,显示非录音信息312b的区域中添加了蒙层。其中,非录音信息312a的输入时间早于或者等于录制语音A1的时间。非录音信息312b的输入时间晚于录制语音A1的时间。
前述描述语音转文本区域(区域12)可以音文同步,是以尺寸1的区域12为例进行 说明的,大小为尺寸2的语音转文本区域(区域12)中也可以音文同步。例如,响应于针对展开控件151a的操作(例如点击操作),电子设备可以显示大小为尺寸2的语音转文本区域(区域12),此时,电子设备可以显示如图14中(3)所示的用户界面32。用户界面32中,区域12中仍然包括文字信息311。区域2中的非录音信息312a上没有添加蒙层。区域2中的除非录音信息312a以外的其他非录音信息上添加了蒙层。
后续,电子设备可以继续在播放语音时实现音文同步。如图14中(4)的用户界面33所示,播放到第8s对应的语音(语音A2)时,电子设备可以在大小为尺寸2的语音转文本区域(区域12)中显示文字信息313,该文字信息313中包括录制语音A2时转换得到的文字。电子设备在录制语音A2(第8s对应的语音)时涉及的相关内容可以参考前述对图5中(1)所示出的用户界面15d的描述,此处不再赘述。电子设备还可以在编辑区域(区域2)中的非录音信息314b上添加蒙层以区别于非录音信息314a。表示该非录音信息314b的输入时间晚于录制语音A2的时间,非录音信息314a的输入时间早于或者等于录制语音A2的时间。
应该理解的是,响应于针对录音条2中开始播放控件141b的操作(例如点击操作)电子设备可以开始播放录制得到的语音(长语音)。此时,电子设备还可以将录音条2中的开始播放控件141b替换为暂停播放控件141f得到状态3下的录音条(录音条3),该播放控件141b可以用于触发暂停播放语音(长语音)。该录音条3中还可以包括其他控件,例如开始录音控件141a、播放进度条141c、播放时间指示符141d以及语音菜单控件141e。关于该其他控件的相关内容可以参考前述图3中(3)所示的用户界面14b中相关控件的描述,此处不再赘述。
在一些可能的情况下,电子设备在开始播放音频之前,还可以显示提示信息以提示用户播放语音时可以实现音文同步。例如,参加前述图14中的用户界面30,电子设备可以显示提示框301,该提示框301可以用气泡的形式呈现,指向开始播放控件141b,提示框301中的内容可以为:“点击播放录音,您记录的内容可在录音播放时展示”。该提示框301中的内容可以用于提示用户点击开始播放控件141b即可播放音频,播放音频时可以实现音文同步。该提示框301可以在暂停录制之后显示,显示次数为1次。用户点击屏幕任一位置该提示框301即可消失,该提示框301也可以有其他的显示规则以及消失规则,本申请实施例对此不作限定。
在一些可能的情况下,在查看笔记时,电子设备还可以实现提前预览。提前预览可以包括下述内容的一个或者多个。
预览方式1:在播放时间B录制的语音时,检测到针对播放进度条上位置T(位置T对应播放的语音为在时间C录制的语音)的操作A(例如持续性长按操作),电子设备可以在区域12中显示时间C(时间B之后的某一时间)录制的语音转换得到的文字,以及时间C之前录制的语音转换的得到的部分文字,并且不影响播放时间B录制的语音。以使得用户可以预览时间C录制的语音转换得到的文字。其中,时间C以及时间B可以指示一段时间,也可以指示一个时刻,本申请实施例对此不作限定。
预览方式2:在播放时间B录制的语音时,检测到针对播放进度条上位置T(位置T 对应播放的语音为在时间C录制的语音)的操作B(例如双击之后持续性长按操作),电子设备可以在区域2中突出显示时间C(时间B之后的某一时间)输入的非录音信息(非录音信息A),并且不影响播放时间B录制的语音。其中,突出显示包括将非录音信息A进行标红、加粗等中的一个或者多个,本申请实施例对此不作限定。突出显示非录音信息A也可以理解为以不同状态(不同于状态A以及状态B的状态)显示该非录音信息A。
应该理解的是,操作A与操作B可以是相同的操作也可以是不同的操作。这里以操作A以及操作B是相同的操作,且电子设备可以通过预览方式1以及预览方式2实现提前预览为例进行说明。实际应用中,操作A与操作B可以不同,电子设备可以通过预览方式1或者预览方式2中的一个实现提前预览。
如图15中(1)所示,电子设备当前播放的语音为录音条3中播放进度条141c上位置341对应播放的语音(即用户界面34中第8s对应的语音)。
响应于针对录音条3中播放进度条141c上位置342(相当于一种位置T)的操作(例如持续性长按操作),电子设备可以改变区域12以及区域2中显示的内容,且,此时,电子设备可以继续播放位置341对应播放的语音。这里假设位置342对应播放的语音为第23s对应的语音,该第23s对应的语音的录制时间为时间C。则电子设备可以在区域12中显示时间C录制的语音转换得到的文字,以及,以及时间C之前录制的语音转换的得到的部分文字。例如,可以包括文字信息344。其中,电子设备得到文字信息344时涉及的内容可以参考前述对图12中(1)的描述,此处不再赘述。电子设备还可以在区域2中突出显示时间C输入的非录音信息334(一种非录音信息A)。例如,可以将该非录音信息334进行加粗显示。电子设备输入非录音信息334到区域2中的内容可以参考前述对图12中(1)的描述,此处不再赘述。
应该理解的是,电子设备在检测到用户结束长按位置342的操作时,可以将区域12以及区域2中包括的内容更新为:播放位置341对应的语音时所显示的内容。例如,参考用户界面34,响应于用户结束长按位置342的操作,电子设备可以重新显示前述涉及的用户界面33。
在一些可能的情况下,电子设备打开笔记1之后,在播放该笔记1中的语音时,默认情况下开启音文同步功能,电子设备可以关闭音文同步功能。关闭音文同步功能的一种示例性方法可以参考下述对图15中(2)、图15中(3)以及图15中(4)所示出的描述。
参考图15中(2)所示的用户界面33a,响应于针对区域2中任一位置的操作(例如长按操作),电子设备可以显示关闭音文同步功能所对应的控件。例如,参考图15中(3)所示的用户界面35,控件351即为关闭音文同步功能所对应的控件。该控件351中还可以包括提示信息,该提示信息用于提醒用户该控件用于触发关闭音文同步功能,例如该提示信息可以为:“停止笔记回放动画”,这里的笔记回放动画即为前述涉及的音文同步。关闭音文同步功能之后,电子设备在区域2中不再添加蒙层区分非录音信息的输入时间是否晚于当前播放语音的录制时间。例如,此时,电子设备可以显示如下述图15中(4)所示的用户界面36。
应该理解的是,前述涉及的关闭音文同步功能的方式可以看作关闭了笔记1的音文同 步。也可以看作关闭了全部笔记的音文同步。本申请实施例对此不作限定。
在另一些可能的情况下,电子设备还可以通过其他的方式关闭全部笔记的音文同步。例如,电子设备可以在添加一个针对全部笔记的设置项1,该设置项1可以用于控制是否开启音文同步,在该设置项1对应的设置值为关闭(OFF)时,表示关闭了全部笔记的音文同步,即进入任一笔记都无法使用音文同步功能。在该设置项1对应的设置值为开启(ON)时,表示开启了全部笔记的音文同步,即进入任一笔记都可以使用音文同步功能。该过程涉及的相关内容可以参考下述对图16的描述。
如图16中(1)所示,用户界面11f中可以包括控件114,该控件114可以用于触发电子设备显示更多关于笔记的设置项。关于该用户界面11f的描述可以参考前述对图1A中(2)对用户界面11a的描述,此处不再赘述。
响应于针对该控件114的操作,电子设备可以显示字少一个关于笔记的设置项。例如,此时,电子设备可以显示如下述图16中(2)所示的用户界面。
如图16中(2)所示,用户界面37中可以包括设置功能项374,该设置功能项374可以用于触发电子设备对全部笔记进行设置。例如,对针对全部笔记的以音文同步功能进行设置(关闭或者开启)。该用户界面37中还可以包括其他设置项,例如还可以包括宫格视图设置项371、批量删除设置项372以及排序方式设置项373。其中,宫格视图设置项371可以用于设置全部笔记的显示方式。批量删除设置项372可以用于一次性删除一个或者多个笔记。排序方式设置项373可以用于设置全部笔记的排序方式,例如按照笔记的大小从前向后排列等。
响应于针对设置功能项374的操作(例如点击操作),电子设备可以显示更多针对全部笔记的设置项。此时,电子设备可以显示如下述图16中(3)所示的用户界面。
如图16中(3)所示,用户界面38中可以包括“笔记回放动画”设置项381(相当于一种设置项1)。该“笔记回放动画”设置项381默认开启,即电子设备默认开启全部笔记的音文同步功能。响应于针对关闭“笔记回放动画”设置项381的操作,电子设备可以关闭全部笔记的音文同步功能。
电子设备在保存笔记1之后,用户可以查看笔记1中记录的信息,还可以继续在笔记1中记录信息。该过程可以参考下述对图17的描述。
如图17中(1)所示,用户界面42中包括至少一个笔记对应的目录项。用户触发一个笔记对应的目录项,电子设备可以打开对应的笔记,显示该笔记中记录的内容。例如,响应于针对笔记1对应的目录项421的操作(例如点击操作),电子设备可以打开笔记1,显示该笔记1中记录的内容。
在一些可能的情况下,电子设备打开记录了信息的笔记1时,仍然可以在用户界面中划分出两个区域:录音区域(前述涉及的区域1)、编辑区域(前述涉及的区域2)。区域1用于呈现与录音相关的内容,区域2用于显示非录音信息。关闭笔记1之后,再次打开笔记1时,区域1中包括的内容为以下两种情况中的一个:
情况1.关闭笔记1之后,再次打开笔记1时,区域1仍然被划分为语音录播区域(前 述涉及的区域11)以及语音转文本区域(前述涉及的区域12)。例如,参考图17中(2)所示的用户界面43,用户界面43中,区域11中包括录音条。关闭笔记1之后,再次打开笔记1的情况下,默认录音功能处于关闭状态。此时区域11中显示的录音条可以被称为第4状态下的录音条(录音条4)。该状态4为录音功能处于关闭,且没有播放语音时的状态。区域12中可以显示语音转文本的内容。该录音条4中可以包括开始播放控件141b、播放进度条141c、播放时间指示符141d以及语音菜单控件141e。关于录音条4中包括的各控件的描述可以参考前述对录音条2中包括的各控件的描述,此处不再赘述。
情况2.关闭笔记1之后,再次打开笔记1时,区域1中可以包括录音区域(区域1),但是不包括语音转文本区域(区域12)。例如,参考图17中(3)所示的用户界面44,用户界面44中,区域11中可以包括区域1,该区域11中包括录音条4。此时,区域1中不包括语音转文本区域(区域12)。
应该理解的是,这里显示录音条4只是一种示例性说明。在其他的情况下,录音条还可以为其他的状态,例如可以为前述涉及的状态2下的录音条(录音条2)。本申请实施例对此不作限定。
这里以情况1为例进行说明。响应于针对开始播放控件141b的操作(例如点击操作),电子设备可以开始播放录制的语音(长语音)。此时,电子设备可以显示如下述图17中(4)所述的用户界面45。
用户界面45中可以包括状态5下的录音条(录音条5),该状态5为录音功能处于关闭,且播放语音时的状态。该录音条5中可以包括暂停播放控件141f、播放进度条141c、播放时间指示符141d以及语音菜单控件141e。关于录音条5中包括的各控件的描述可以参考前述对录音条3中包括的各控件的描述,此处不再赘述。
应该理解的是,这里显示录音条5只是一种示例性说明。在其他的情况下,录音条还可以为其他的状态,例如可以为前述涉及的状态3下的录音条(录音条3)。本申请实施例对此不作限定。
下面对管理信息涉及的相关内容进行描述。
管理信息包括删除信息、分享信息、保存信息等。其中,删除信息包括选择删除笔记1中记录的部分或者全部语音信息。分享信息包括将笔记中的待分享内容传输至其他终端,且可以选择分享方式。该待分享内容可以是用户通过电子设备选择的,例如可以选择分享笔记1中除语音以外的其他信息,还可以选择分享笔记1中除语音以及语音转换得到的文字以外的其他信息,还可以分享笔记1中的全部信息等。
图18-图29描述了电子设备管理信息时涉及的相关内容。
其中,图18-图24为电子设备删除信息时涉及的一组示例性用户界面。图25-图29为电子设备分享信息时涉及的一组示例性用户界面。
下面基于图18-图24描述电子设备删除信息时涉及的相关内容。
删除信息包括删除语音信息,在删除语音信息时,电子设备还可以根据用户需求确定是否删除所选择语音信息对应的文字信息。语音信息对应的文字信息是通过该语音转换得到的。笔记1中可以记录Q段语音,电子设备可以选择删除Q段语音中的L段语音,且选 择是否删除该L段语音对应的文字信息。其中L为小于等于Q的整数。
在一些可能的情况下,电子设备可以提供删除全部语音的功能。在删除全部语音时,电子设备还可以根据用户需求确定是否删除全部语音对应的文字信息。
图18中描述了电子设备删除全部语音时涉及的一组示例性用户界面。
如图18中(1)所示,在暂停语音播放的情况下,响应于针对播放进度条141c的操作C(例如长按操作),电子设备可以显示一个管理工具栏。该管理工具栏中包括至少一个管理笔记1涉及的控件,不同控件可以用于触发对笔记1实施不同处理。例如,如图18中(2)所示,管理工具栏511中可以包括删除控件511a、保存控件511b以及分享控件511c。其中,删除控件511a可以用于触发电子设备删除全部语音信息。保存控件511b可以用于触发电子设备保存录制的全部语音。分享控件511c可以用于触发电子设备分享全部语音信息。
参考图18中(2)所示的用户界面51,响应于针对删除控件511a的操作(例如点击操作),电子设备可以显示一个删除全部语音时涉及的提示框A,该提示框A可以用于提示用户待删除的内容为全部语音,还可以用于接收用户输入的确认删除该全部语音的操作或者取消删除该全部语音的操作。如图18中(3)所示,用户界面52中显示的提示框521可以为一种示例性提示框A。
该提示框521中可以包括提示信息521a,该提示信息521a可以用于提示用户是否删除全部录音(语音)。
该提示框521中还可以包括选择控件521b,该选择控件521b可以用于接收用户是否保留语音转文本内容(全部录音对应的文字信息)的操作。此时,该语音转文本内容即为全部语音对应的文字信息。在一些可能的情况下,默认选中保留语音转文本内容,此时该选择控件521b处于状态E,例如状态E为置灰。选择控件521b的状态为状态E时表示选择保留语音转文本内容,响应于针对状态E的选择控件521b的操作,电子设备可以更改该选择控件521b的状态为状态F,例如该状态F为不进行置灰。选择控件521b的状态为状态F时,表示未选择保留语音转文本内容,则删除全部语音时,会将该语音转文本内容也删除。选择控件521b的状态可以在状态E以及状态F之间进行切换,以确定在删除全部语音时是否删除该全部语音对应的语音转文本内容。
该提示框521中还可以包括确认删除控件521c以及取消控件521d。其中,确认删除控件521c可以用于触发删除笔记1中的全部语音。在用户选择删除(不保留)该全部语音对应的语音转文本内容的情况下,确认删除控件521c还可以用于触发删除全部语音对应的文字信息(语音文本内容)。取消控件521d用于取消删除全部语音,且用于关闭提示框521。
在另一些可能的情况下,电子设备还可以选择删除部分或者全部语音。在删除部分或者全部语音时,还可以设置是否删除该部分或者全部语音对应的文字信息。
其中,电子设备删除全部语音涉及的内容可以参考对图19-图21的相关描述。
参考图19中(1)中示出的用户界面50,响应于针对语音菜单控件141e的操作(例如点击操作),电子设备可以显示一个包括Q段语音信息的显示框A。该显示框A用于显示Q段录音对应的信息(例如录制时长等)。该显示框A中还可以包括对该录制的语音进行管理时涉及的控件。此时电子设备可以显示图19中(2)所示的用户界面。
如图19中(2)所示,用户界面53a中包括的显示框535可以看作一种示例性显示框A。该显示框535中可以包括录制的Q段语音信息:包括第一次录制的语音对应的名称(录音001)及其对应的总时长(11s),以及第二次录制的语音对应的名称(录音002)及其对应的总时长(12s)。该显示框535中还可以包括编辑控件535a。该编辑控件535a可以用于触发显示对录制的语音进行处理(编辑)时涉及的相关界面。
响应于针对编辑控件535a的操作,电子设备可以显示更多对笔记中的语音进行管理(编辑)时所涉及的控件。还可以触发显示笔记1中录制的全部语音对应的信息,该信息可以包括各语音的总时长、语音的名称等。此时,电子设备可以显示图19中(3)所示的用户界面。
如图19中(3)所示,用户界面53b为对语音进行管理(编辑)时涉及的一个示例性用户界面。该用户界面53b中可以包括分享控件531、重命名控件532、删除控件533以及取消全选控件534。该分享控件531可以用于触发将选中的语音分享至其他设备。重命名控件532可以用于触发对语音的名称进行更改。删除控件533可以用于触发选中的语音。电子设备可以设置默认情况下选中全部语音进行处理,取消全选控件534可以用于触发将全部语音从选中更新为未选中。该用户界面53b中,电子设备默认选中全部语音,例如录音001(第一次录音得到的语音对应的名称)对应的选中控件530a处于状态E,例如置灰状态。选中控件530a被置灰则表示第一次录音得到的语音为被选中的语音。同理,录音002(第二次录音得到的语音对应的名称)对应的选中控件530b也处于状态E,例如置灰状态。则表示第一次录音得到的语音也为被选中的语音。
在选中语音为全部语音的情况下,响应于针对删除控件533的操作(例如点击操作),电子设备可以显示前述涉及的删除全部语音时涉及的提示框A,该提示框A可以用于提示用户删除的内容为全部语音,还可以用于接收用户输入的删除该全部语音的操作或者取消删除该全部语音的操作。其中,在全部语音都被选中的情况下,删除控件533以及删除控件511a可以看作功能相同,但是形态不同的控件。此时,电子设备可以显示图20中(1)所示的用户界面54。
在删除全部语音时,电子设备保留该全部语音对应的文字信息时涉及的相关内容,可以参考下述对图20的描述。
如图20中(1)所示,用户界面54为对语音进行管理时涉及的一个示例性用户界面。该用户界面54中可以包括提示框521,该提示框521可以为一种示例性提示框A。关于该提示框521的相关描述可以参考前述图18中(3)涉及的内容,此处不再赘述。此时提示框521中的选中控件521b为状态E,表示选择保留(不删除)语音转文本内容。
在选中全部语音,且选择保留全部语音对应的文字信息(语音本文内容)的情况下,响应于针对确认删除控件521c的操作,电子设备可以删除全部语音,且保留(不删除)全部语音对应的文字信息。即,电子设备可以将全部语音对应的文字信息保留在笔记1中,但是删除笔记1中的全部语音。电子设备将全部语音对应的文字信息保留在笔记1的方式包括但不限于以下两种方式中的一种。
保留方式1:电子设备可以不显示录音区域(区域1),且,将全部语音对应的文字信息***到非录音信息的前面,或者,***到非录音信息的后面。其中,将全部语音对应的 文字信息***到非录音信息的前面时涉及的示例性用户界面可以参考图20中(2)示出的用户界面55a。
保留方式2:电子设备可以不显示录音区域(区域1),且,在笔记1中设置一个支持滑动显示的新区域(区域3)以显示全部语音对应的文字信息。如图20中(3)所示,用户界面55b中可以包括区域3,该区域3中可以显示全部语音对应的文字信息中的一部分。响应于针对该区域3的操作(例如滑动操作),电子设备可以显示全部语音对应的文字信息中的其他部分。
在删除全部语音时,电子设备也一并删除该全部语音对应的文字信息时涉及的相关内容,可以参考下述对图21的描述。
如图21中(1)所示,用户界面54b为对语音进行管理时涉及的一个示例性用户界面。该用户界面54b中可以包括提示框521,该提示框521可以为一种示例性提示框A。关于该提示框521的相关描述可以参考前述图18中(3)涉及的内容,此处不再赘述。此时提示框521中的选中控件521b为状态F,表示选择删除语音转文本内容。
响应于针对确认删除控件521c的操作,电子设备可以删除笔记1中的全部语音,以及删除该全部语音对应的文字信息。删除之后,笔记1中不再包括录制的语音以及该语音对应的文字信息,可以包括非录音信息。例如,此时,电子设备可以显示图21中(2)所示的用户界面55c。
电子设备删除部分语音涉及的内容可以参考对图22-图23的相关描述。
参考图22中(1)所示的用户界面53b,响应于针对取消全选控件534的操作(例如点击操作),电子设备可以将全部语音对应的选中控件(例如选中控件530a以及选中控件530b)从状态E(例如置灰)更改为状态F(不置灰),以表示将全部语音从选中设置为未选中。此时电子设备可以显示如下述图22中(2)所示的用户界面。其中,关于用户界面53b的相关描述可以参考前述对图19中(3)的描述,此处不再赘述。
如图22中(2)所示,用户界面56中全部语音对应的选中控件(例如选中控件530a以及选中控件530b)未置灰。响应于针对未置灰的选中控件530b的操作(例如点击操作),电子设备可以将控件530b置灰以表示选中了录音002(第二次录音时得到的语音)。此时,电子设备可以显示下述图22中(3)所示的用户界面57。
如图22中(3)所示,被选中的语音包括录音002(第二次录音时得到的语音),不包括录音001(第一次录音时得到的语音)。在部分语音被选中的情况下,响应于针对删除控件533的操作,电子设备可以显示一个删除部分语音(选中语音)时涉及的提示框B,该提示框B可以用于提示用户待删除的内容为选中语音,还可以用于接收用户输入的确认删除该选中语音的操作或者取消删除该选中语音的操作。
如图22中(4)所示,用户界面58中显示的提示框581可以为一种示例性提示框B。该提示框581中还可以包括提示信息581a,该提示信息581a可以用于提示用户是否删除选中录音(语音)。
该提示框581中还可以包括选中控件581b,该选中控件581b可以用于接收用户是否保留语音转文本内容(选中录音对应的文字信息)的操作。在一些可能的情况下,默认选 中保留语音转文本内容,此时该选中控件581b处于状态E,例如状态E为置灰。选中控件581b的状态为状态E时表示在删除选中语音时,保留(不删除)该选中语音对应的文字信息,响应于针对选中控件581b的操作,电子设备可以更改该选中控件581b的状态为状态F,例如该状态F为不进行置灰。选中控件581b的状态为状态F时,表示删除选中语音时,会将该选中语音对应的文字信息也删除。选中控件581b的状态可以在状态E以及状态F之间进行切换,以确定在删除选中语音时是否删除该选中语音对应的语音转文本内容。
用户界面58中,选中控件581b的状态为未置灰,表示在删除选中语音时一并删除选中语音对应的文字信息。此时,被选中语音包括录音002(第二次录音时得到的语音),不包括录音001(第一次录音时得到的语音)。响应于针对确认删除控件581c的操作,电子设备可以删除第二次录音得到的语音以及删除该第二次录音得到的语音对应的文字信息。第一次录音得到的语音及其对应的文字信息还可以保留在笔记1中。
在笔记1中记录了Q段语音,且删除L段语音的情况下,在L等于Q时,删除L段语音且删除该L段语音对应的文字信息之后,电子设备显示的用户界面可以参考前述对图21中(2)的描述。删除L段语音但是保留了该L段语音对应的文字信息之后,电子设备显示的用户界面可以参考前述对图20中(2)或者图20中(3)的描述,此处不再赘述。
在笔记1中记录了Q段语音,且删除L段语音的情况下,在L小于Q时,删除L段语音且删除该L段语音对应的文字信息之后,电子设备显示的用户界面可以参考下述图23中(1)所示。
如图23中(1)所示,用户界面59a中包括的播放进度条141c中包括Q-L-1个节点,即相比于删除之前,节点减少了L个,对应删除了L段语音。这里Q=2,L=1,则播放进度条141c不再包括节点,表示此时该笔记中只记录了一个语音。播放时间指示符141d中记录的总时长从Q段语音的总时长变更为Q-L段语音的总时长,例如这里变更为第一次录音得到的语音的总时长:11s。区域12中包括未被删除的Q-L段语音对应的文字信息,不再包括被删除的L段语音对应的文字信息。
在笔记1中记录了Q段语音,且删除L段语音的情况下,在L小于Q时,删除L段语音且不删除(保留)该L段语音对应的文字信息之后,电子设备显示的用户界面可以参考下述图23中(2)所示。
如图23中(2)所示,用户界面59b中包括的播放进度条141c中包括Q-L-1个节点,即相比于删除之前,节点减少了L个,对应删除了L段语音。这里Q=2,L=1,则播放进度条141c不再包括节点,表示此时该笔记中只记录了一个语音。播放时间指示符141d中记录的总时长从Q段语音的总时长变更为Q-L段语音的总时长,例如这里变更为第一次录音得到的语音的总时长:11s。区域12中包括Q段语音对应的文字信息。
在一些可能的情况下,在笔记1中录制的语音只包括一段语音时,电子设备可以用其他的方式触发显示对该段语音进行管理(编辑)时涉及的界面。该过程可以参考下述对图24的描述。
参考图24中(1)所示的用户界面60,响应于针对语音菜单控件141e的操作(例如点击操作),电子设备可以不显示前述涉及的显示框A。只显示编辑控件,该编辑控件即可 用于触发显示对该段语音进行管理(编辑)时涉及的界面。该编辑控件的一种示例性展示可以参考下述图24中(2)所示的用户界面61中包括的编辑控件611。
响应于针对编辑控件611的操作,电子设备可以显示更多对该段语音进行管理(编辑)时所涉及的控件。还可以触发显示该段语音对应的信息,该信息可以包括各语音的总时长、语音的名称等。此时,电子设备可以显示图24中(3)所示的用户界面。关于该用户界面62中涉及的控件的相关描述可以参考前述对图22中(2)的描述,此处不再赘述。
下面基于图25-图28描述电子设备分享信息时涉及的相关内容。
分享信息包括将笔记中的待分享内容传输至其他终端,且可以选择分享方式。该待分享内容可以是用户通过电子设备选择的。
在一些可能的情况下,该待分享内容可以为笔记中除语音信息以外的信息。例如,除语音信息以外的信息中包括区域12中的语音转文本内容,还可以包括区域2中的非录音内容。
示例性的,图25中描述了待分享内容为笔记中除语音以外的信息时涉及的一组示例性用户界面。
如图25中(1)所示,用户界面70为对笔记1进行处理时涉及的一个示例性用户界面。该用户界面70包括至少一个针对笔记1的功能控件。例如,可以包括分享控件701a、收藏控件701b、删除控件701c以及更多控件701d。其中,收藏控件701b可以用于触发对笔记1进行收藏。删除控件701c可以用于触发删除笔记1。更多控件701d可以用于触发显示更多针对笔记1的功能控件。分享控件701a可以用于触发显示对笔记1中的内容进行分享时所涉及的相关界面。
响应于针对分享控件701a的操作(例如点击操作),电子设备可以显示图25中(2)所示的用户界面71。该用户界面71为对笔记1中的内容进行分享时所涉及的一个示例性界面。该用户界面71中可以包括选择框711,该选择框711中可以包括选择不同分享方式所涉及的控件。例如,该选择框711中可以包括控件711a、控件711b、控件711c以及控件711d。其中,控件711a可以用于触发将笔记中的待分享内容转换为图片之后,以图片的形式传输至其他终端。控件711b可以用于触发将笔记中的待分享内容转换为文本之后,以文档的形式传输至其他终端,该文本的格式包括txt。控件711d可以用于触发将笔记中的待分享内容仍然以笔记的形式传输至其他终端,即分享时笔记的格式可以不改变。控件711c可以用于触发将笔记转换为文档之后进行保存,该过程包括:电子设备可以调用与文档编辑类应用生成文档的功能将笔记1转换为文档之后进行保存。后续,电子设备可以将该文档分享至终端。其中,文档可以包括word文档或者PDF文档。
这里以将待分享内容转换为图片进行分享为例进行说明。
响应于针对控件711a的操作,电子设备可以显示下述图25中(3)示出的用户界面72,该用户界面72为将笔记中的待分享内容转换为图片进行分享时涉及的一个示例性分享界面。该用户界面72中还可以包括具有分享功能的应用对应的图标,例如具有分享功能的应用可以为图中示出的应用1、应用2以及其他应用等。其中,有分享功能的应用对应的图标用于触发通过该图标对应的应用将待分享内容传输至其他设备。
该用户界面72中还可以包括选中控件722。默认情况下,可以设置该选中控件722的状态为状态G(例如该状态G为置灰),状态G下的选中控件722可以用于指示将区域12中的语音转文本内容作为待分享内容中的一部分。此时,电子设备可以将区域12中的语音转文本内容以及区域2中的非录音信息整合成图片进行分享。该整合后的图片的一种示例性展示可以参考界面72中示出的图片721。
后续,电子设备可以通过具有分享功能的应用(例如图中示出的应用1、应用2以及其他应用等)将区域12中的语音转文本内容以及区域2中的非录音信息整合成图片传输至其他设备。
响应于针对选中控件722的操作,电子设备可以显示图25中(4)所示的用户界面73。用户界面73中,电子设备可以将选中控件的状态从状态G设置为状态F(例如不置灰),状态F下的选中控件722可以用于指示可以不将区域12中的语音转文本内容作为待分享内容中的一部分。此时,电子设备可以将区域2中的非录音信息整合成图片进行分享。该整合后的图片的一种示例性展示可以参考界面73中示出的图片731。
后续,电子设备可以通过具有分享功能的应用(例如图中示出的应用1、应用2以及其他应用等)将区域2中的非录音信息整合成图片传输至其他设备。
应该理解的是,这里是以默认状态下,选中控件722指示将区域12中的语音转文本内容作为待分享内容中的一部分为例进行说明的。实际使用中,还可以使得选中控件722默认指示不将区域12中的语音转文本内容作为待分享内容中的一部分。本申请实施例对此不作限定。
在一些可能的情况下,该待分享内容可以为笔记中的全部内容。例如,该全部内容中可以包括全部语音、可以包括区域12中的语音转文本内容,还可以包括区域2中的非录音内容。
示例性的,图26中描述了待分享内容可以为笔记中全部内容时涉及的一组示例性用户界面。
在该情况下,参考前述图25中(1)所示的用户界面70,响应于针对分享控件701a的操作,电子设备可以显示图26中(1)所示的用户界面74a。该用户界面74a为对笔记1中的内容进行分享时所涉及的另一个示例性界面。
如图26中(1)所示,用户界面74a中可以包括选择框741,该选择框741中可以包括选择不同分享方式所涉及的控件。例如,该选择框741中可以包括控件741a、控件741b。其中,控件741a可以用于触发将待分享内容中的语音以音频的形式传输至其他终端,且,将出语音以外的其他信息转换为图片后,以图片的形式传输至其他终端。控件741b可以用于触发将待分享内容中的语音以音频的形式传输至其他终端,且,将出语音以外的其他信息转换为文本之后,以文档的形式传输至其他终端,该文本的格式包括txt。还可以包括其他的分享方式,例如前述涉及的以笔记的形式分享至其他设备的方式,本申请实施例对此不作限定。
响应于针对控件741a的操作,电子设备可以显示下述图26中(2)示出的用户界面74b。该用户界面74b为将笔记中的待分享内容转换为图片以及音频进行分享时涉及的一个示例 性分享界面。该用户界面74b中可以包括选中控件742a以及选中控件742b。默认情况下,可以设置该选中控件742a以及选中控件742b的状态为状态J(例如该状态J为置灰),状态J下的选中控件742a可以用于指示将区域12中的语音转文本内容作为待分享内容中的一部分。状态J下的控件742b用于指示将笔记1中的语音作为待分享内容中的一部分。此时,电子设备而可以将笔记1中的语音作为音频(例如音频743)传输至其他设备。电子设备还可以将区域12中的语音转文本内容以及区域2中的非录音信息整合成图片进行分享。该整合后的图片的一种示例性展示可以参考用户界面74b中示出的图片744。
后续,电子设备可以通过具有分享功能的应用(例如图中示出的应用1、应用2以及其他应用等)将音频743以及图片744传输至其他设备。
其他设备接受到该音频743以及图片744的显示方式包括但不限于以下显示方式:
显示方式11:将音频743以及图片744分成两条信息进行显示。音频743的显示方式包括以音频文件的形式进行显示,下载该音频文件之后,即可播放该音频743。
显示方式12:将音频743以及图片744分成两条信息进行显示。音频743的显示方式包括以音频消息进行显示,响应于针对该音频消息的操作(例如点击操作),即可播放该音频743。
显示方式13:将音频743以及图片744分成两条信息进行显示。音频743的显示方式包括以语音条的形式进行显示,该语音条中可以提供两个控件,其中一个控件用于触发播放音频,另一个控件用于触发语音转文本。
这里应该理解的是,在选中控件742b的状态为状态K时,例如该状态K可以为未置灰。电子设备可以将区域12中的语音转文本内容以及区域2中的非录音信息整合成图片或者文档进行分享。该过程可以参考前述对图25中(3)的描述,此处不再赘述。在选中控件742a以及选中控件742b的状态均为状态K时。电子设备可以不将区域12中的语音转文本内容进行分析,将区域2中的非录音信息整合成图片或者文档进行分享。该过程可以参考前述对图25中(4)的描述,此处不再赘述。
在一些可能的情况下,该待分享内容可以包括选中的语音,还可以包括该选中的语音对应的文字信息。
示例性的,图27A中描述了待分享内容可以为选中的语音以及该选中的语音对应的文字信息时涉及的一组示例性用户界面。
如图27A中(1)所示,用户界面57为对语音进行编辑时涉及的一个示例性界面。对该用户界面57的相关描述可以参考前述对图22中(3)的相关描述,此处不再赘述。此时,电子设备选中的语音包括录音002(第二次录音得到的语音对应的名称)。响应于针对分享控件531的操作,电子设备可以显示一个分享选中语音时涉及的提示框C,该提示框C可以用于提示用户待分享的内容为选中语音,还可以用于接收用户输入的确认分享该选择语音的操作或者取消分享该选中语音的操作。
如图27A中(2)所示,用户界面75中显示的提示框751可以为一种示例性提示框C。该提示框751中还可以包括提示信息751a,该提示信息751a可以用于提示用户是否分享选中录音(语音)。
该提示框751中还可以包括选中控件751b,该选中控件751b可以用于接收用户是否分享语音转文本内容(选中语音对应的文字信息)的操作。在一些可能的情况下,默认选中分享语音转文本内容,此时该选中控件751b处于状态E1,例如状态E1为置灰。选中控件751b的状态为状态E1时表示在分享选中语音时,也会分享该选中语音对应的文字信息。响应于针对选中控件751b的操作,电子设备可以更改该选中控件751b的状态为状态F1,例如该状态F1为不进行置灰。选中控件751b的状态为状态F1时,表示分享选中语音时,可以不将该选中语音对应的文字信息也进行分享。选中控件751b的状态可以在状态E1以及状态F1之间进行切换,以确定在分享选中语音时是否分享该选中语音对应的文字信息。
该提示框751中还可以包括确认分享控件751c。在选中控件751的状态为E1(例如置灰)时,该确认分享控件751c可以用于触发分享选中语音以及该选择语音对应的文字信息。此时,响应于针对确定分享控件751c的操作,电子设备可以显示选择分享方式时涉及的用户界面,在选择分享方式之后,电子设备可以将选中语音以及选择语音对应的文字信息传输至其他终端,该过程可以参考前述对图26中(1)以及图26中(2)中涉及的分享方式的描述,此处不再赘述。
在选中控件751b的状态为F1(例如未置灰)时,该确认分享控件751c可以用于触发分享选中语音,但是不分享该选中语音对应的文字信息。此时,响应于针对确定分享控件751c的操作,电子设备可以将选中语音以音频的形式通过具有分享功能的应用(例如图中示出的应用1、应用2以及其他应用等)传输至其他设备。
应该理解的是,前述选中控件751b是可选的,在一些可能的情况下,在分享选中语音时可以不涉及是否分享该选中语音对应的文字信息,默认不对该选中语音对应的文字信息进行分享,本申请实施例对此不作限定。
在另一些可能的情况下该待分享内容可以包括选中的语音,还可以包括该选中的语音对应的文字信息,还可以包括该语音所属的笔记。
如图27A中(1)所示,响应于针对分享控件531的操作,电子设备可以不显示前述涉及的提示框A,可以显示一个分享选中语音时涉及的提示框D,该提示框D可以用于提示用户待分享的内容为选中语音,还可以用于接收用户输入的确认分享该选择语音的操作,或者,取消分享该选中语音的操作。还可以用于接收用户输入的确认分享该选择语音所述笔记的操作,或者,取消分享该笔记的操作。
如图27B所示,用户界面76中显示的提示框761可以为一种示例性提示框D。该提示框761中还可以包括提示信息751a,该提示信息751a可以用于提示用户是否分享选中录音(语音)。该提示框761中还可以包括选中控件751b,该选中控件751b可以用于接收用户是否分享语音转文本内容(选中语音对应的文字信息)的操作。关于该提示信息751a以及控件751b的操作,可以参考前述图27A中(2)的描述,此处暂不赘述。
该提示框761中还可以包括选中控件7611,该选中控件7611可以用于接收是否分享选择语音所属笔记(笔记1)的操作。在一些可能的情况下,默认选择分享笔记1,此时该选中控件处于状态E2,例如状态E2为置灰。选中控件7611的状态为状态E2时表示在分享选中语音时,也会分享该选中语音所属的笔记。响应于针对选中控件7611的操作,电子 设备可以更改该选中控件7611的状态为状态F2,例如该状态F2为不进行置灰。选中控件7611的状态为状态F2时,表示分享选中语音时,可以不将该选中语音所属的笔记进行分享。选中控件7611的状态可以在状态E2以及状态F2之间进行切换,以确定在分享选中语音时是否分享该选中语音所属的笔记。
在另一些可能的情况下,该待分享内容可以包括选中的语音转文本内容。
示例性的,图28中描述了待分享内容可以为选中的语音转文本内容时涉及的一组示例性用户界面。
如图28所示,用户界面17a为对选中的语音转文本内容进行处理时涉及的一个示例性用户界面,关于该用户界面17a涉及的相关内容可以参考前述对图8中(1)的描述,此处不再赘述。此时,选中标记162指示选中的语音转文本内容为:“第二部分主要讲述了XXXXXX。”。响应于针对分享控件161e的操作,电子设备可以将该选中的语音转文本内容传输至其他设备。
应该理解的是,本申请实施例中涉及的用户界面是一种示例性展示,任一用户界面中都可以包括更多或者更少的控件。各控件的形式可以变化,包括但不限于图标、文字等。各控件的位置也可以变化,不应该构成对本申请实施例的限定。例如,区域1中显示的转换控件137a除了可以显示在区域12中以外,还可以显示在区域11中,或者其他地方。例如,参考图29中(1)所示,在开始录音之后该转换控件137a可以置于区域11中。在例如,参考图29中(2)所示,在暂停录音之后,该转换控件1可以置于区域11中。
图30中描述了本申请中信息处理方法的一个示例性流程图。
在一些可能的情况下,电子设备记录信息时涉及的示例性过程可以参考下述对步骤S101-步骤S106的描述。
S101.在建立笔记1之后,电子设备显示界面1,该界面1中包括区域1、区域2以及状态为状态C的转换控件。
在一些可能的情况下,前述内容中涉及的笔记1也可以被称为第一文件。界面1也可以被称为第一界面。区域1可以被称为第一区域,区域2可以被称为第二区域。状态C可以被称为第一状态。转换控件可以被称为第一控件。
步骤S101中,建立笔记1涉及的相关过程可以参考前述对图1A以及图1B的描述,此处不再赘述。
该转换控件可以为前述涉及的转换控件137a。区域1为前述涉及的用于呈现与录音相关内容的区域。区域2为前述涉及的编辑区域。该界面1可以为前述涉及的包括区域1、区域2以及状态为状态C的转换控件的示例性用户界面。例如,可以为前述图2中(3)示出用户界面13b、图2中(4)示出用户界面13c等。
状态C的转换控件用于触发将录制的语音转换为文字。
S102.第一时间,显示控件1指示电子设备在录制语音1的情况下,将该语音1转换得 到的文字显示在区域1中,且,接收输入的信息1,将该信息1显示在区域2中。
在一些可能的情况下,该语音1也可以被称为第一语音,该控件1也可以被称为第二控件。该信息1也可以被称为第一信息。
语音1可以为前述涉及的第一次录制的语音。第一时间可以为电子设备录制语音1时的某一时间。信息1可以为电子设备通过输入工具输入的信息,例如文字或者图片等。其中,输入工具包括键盘以及手写笔工具栏。控件1可以为前述涉及的状态1下的录音条(录音条1)或者暂停录音控件136a。
电子设备建立笔记1之后,通过界面1中包括的录制控件可以触发录制语音1。在转换控件处于状态C的情况下,电子设备可以开启语音转文本功能。此时,电子设备可以在录制语音1的同时,将语音1转换得到的文本显示在区域1中,且,可以输入工具输入的信息(例信息1),并且将该信息1显示在区域2。该过程涉及的示例性用户界面可以参考前述图2中示出的用户界面13c等相关界面。
S103.响应于针对状态为状态C的转换控件的操作,电子设备停止将语音转换为文字,且,将转换控件的状态更新为状态D。
响应于针对状态为状态C的转换控件的操作,电子设备可以将转换控件的状态更新为状态D,该状态D的转换控件可以用于指示停止将语音转换为文字。
此时,电子设备仍然可以继续录制语音,在录制语音的过程中,可以通过输入工具输入的信息。
该过程涉及的示例性界面可以参考前述对图7中(1)以及对图7中(2)的描述,此处不再赘述。
S104.区域1划分为区域11以及区域12,区域11中包括控件1,区域12中包括语音转换得到的文字,响应于针对控件1的操作,电子设备将控件1替换成控件2,该控件2用于指示电子设备结束录制语音,还用于触发电子设备录制其他语音。
在一些可能的情况下,该区域11可以被称为第一子区域,该区域12可以被称为第二子区域。控件2也可以被称为第三控件。
区域11可以为前述涉及的语音录播区域、区域12可以为前述涉及的语音转文本区域。
在控件1为状态1下的录音条(录音条1)时,控件2可以为状态2下的录音条(录音条2)。在控件1为暂停录音控件136a时,控件2可以为开始录音控件141a。
步骤S104涉及的相关内容可以参考前述对图7中(2)以及对图7中(3)中相关内容的描述,此处不再赘述。
S105.在结束录制语音之后,未获取到语音转换得到的文字时,区域1中不包括区域12。
该过程涉及的相关内容可以参考前述对图3中(2)以及图3中(3)中相关内容的描述,此处不再赘述。
S106.电子设备将语音1、语音1转换得到的文字以及信息1记录到笔记1中。
下面首先介绍本申请实施例提供的示例性电子设备。
图31是本申请实施例提供的电子设备的结构示意图。
下面以电子设备为例对实施例进行具体说明。应该理解的是,电子设备可以具有比图中所示的更多的或者更少的部件,可以组合两个或多个的部件,或者可以具有不同的部件配置。图中所示出的各种部件可以在包括一个或多个信号处理和/或专用集成电路在内的硬件、软件、或硬件和软件的组合中实现。
电子设备可以包括:处理器110,外部存储器接口120,内部存储器121,通用串行总线(universal serial bus,USB)接口130,充电管理模块140,电源管理模块141,电池142,天线1,天线2,移动通信模块150,无线通信模块160,音频模块170,扬声器170A,受话器170B,麦克风170C,耳机接口170D,传感器模块180,按键190,马达191,指示器192,摄像头193,显示屏194以及用户标识模块(subscriber identificationmodule,SIM)卡接口195等。其中传感器模块180可以包括压力传感器180A,陀螺仪传感器180B,气压传感器180C,磁传感器180D,加速度传感器180E,距离传感器180F,接近光传感器180G,指纹传感器180H,温度传感器180J,触摸传感器180K,环境光传感器180L,骨传导传感器180M等。
可以理解的是,本申请实施例示意的结构并不构成对电子设备的具体限定。在本申请另一些实施例中,电子设备可以包括比图示更多或更少的部件,或者组合某些部件,或者拆分某些部件,或者不同的部件布置。图示的部件可以以硬件,软件或软件和硬件的组合实现。
处理器110可以包括一个或多个处理单元,例如:处理器110可以包括应用处理器(applicationprocessor,AP),调制解调处理器,图形处理器(graphics processing unit,GPU),图像信号处理器(image signalprocessor,ISP),控制器,存储器,视频编解码器,数字信号处理器(digital signal processor,DSP),基带处理器,和/或神经网络处理器(neural-network processingunit,NPU)等。其中,不同的处理单元可以是独立的器件,也可以集成在一个或多个处理器中。
其中,控制器可以是电子设备的神经中枢和指挥中心。控制器可以根据指令操作码和时序信号,产生操作控制信号,完成取指令和执行指令的控制。
处理器110中还可以设置存储器,用于存储指令和数据。在一些实施例中,处理器110中的存储器为高速缓冲存储器。该存储器可以保存处理器110刚用过或循环使用的指令或数据。如果处理器110需要再次使用该指令或数据,可从所述存储器中直接调用。避免了重复存取,减少了处理器110的等待时间,因而提高了***的效率。
在一些实施例中,处理器110可以包括一个或多个接口。接口可以包括集成电路(inter-integrated circuit,I2C)接口,集成电路内置音频(inter-integrated circuit sound,I2S)接口,脉冲编码调制(pulse code modulation,PCM)接口等。
可以理解的是,本申请实施例示意的各模块间的接口连接关系,只是示意性说明,并不构成对电子设备的结构限定。在本申请另一些实施例中,电子设备也可以采用上述实施例中不同的接口连接方式,或多种接口连接方式的组合。
充电管理模块140用于从充电器接收充电输入。
电源管理模块141用于连接电池142,充电管理模块140与处理器110。
电子设备的无线通信功能可以通过天线1,天线2,移动通信模块150,无线通信模块160,调制解调处理器以及基带处理器等实现。
在一些实施例中,电子设备的天线1和移动通信模块150耦合,天线2和无线通信模块160耦合,使得电子设备可以通过无线通信技术与网络以及其他设备通信。所述无线通信技术可以包括全球移动通讯***(global system for mobile communications,GSM),通用分组无线服务(general packet radio service,GPRS)等。
电子设备通过GPU,显示屏194,以及应用处理器等实现显示功能。
显示屏194用于显示图像,视频等。显示屏194包括显示面板。在一些实施例中,电子设备可以包括1个或N个显示屏194,N为大于1的正整数。
电子设备可以通过ISP,摄像头193,视频编解码器,GPU,显示屏194以及应用处理器等实现拍摄功能。
ISP用于处理摄像头193反馈的数据。
摄像头193用于捕获静态图像或视频。
数字信号处理器用于处理数字信号,除了可以处理数字图像信号,还可以处理其他数字信号。例如,当电子设备在频点选择时,数字信号处理器用于对频点能量进行傅里叶变换等。
视频编解码器用于对数字视频压缩或解压缩。
NPU为神经网络(neural-network,NN)计算处理器,通过借鉴生物神经网络结构,例如借鉴人脑神经元之间传递模式,对输入信息快速处理,还可以不断的自学习。
内部存储器121可以包括一个或多个随机存取存储器(random access memory,RAM)和一个或多个非易失性存储器(non-volatile memory,NVM)。
外部存储器接口120可以用于连接外部的非易失性存储器,实现扩展电子设备的存储能力。外部的非易失性存储器通过外部存储器接口120与处理器110通信,实现数据存储功能。例如将音乐,视频等文件保存在外部的非易失性存储器中。
电子设备可以通过音频模块170,扬声器170A,受话器170B,麦克风170C,耳机接口170D,以及应用处理器等实现音频功能。例如音乐播放,录音等。
音频模块170用于将数字音频信息转换成模拟音频信号输出,也用于将模拟音频输入转换为数字音频信号。
扬声器170A,也称“喇叭”,用于将音频电信号转换为声音信号。电子设备可以通过扬声器170A收听音乐、录音,或收听免提通话。
受话器170B,也称“听筒”,用于将音频电信号转换成声音信号。
麦克风170C,也称“话筒”,“传声器”,用于将声音信号转换为电信号。
耳机接口170D用于连接有线耳机。
按键190包括开机键,音量键等。
马达191可以产生振动提示。
指示器192可以是指示灯,可以用于指示充电状态,电量变化,也可以用于指示消息, 未接来电,通知等。
SIM卡接口195用于连接SIM卡。
本申请实施例中,该处理器110可以调用内部存储器121中存储的计算机指令,以使得电子设备执行本申请实施例中的信息处理方法。
以上所述,以上实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的范围。
上述实施例中所用,根据上下文,术语“当…时”可以被解释为意思是“如果…”或“在…后”或“响应于确定…”或“响应于检测到…”。类似地,根据上下文,短语“在确定…时”或“如果检测到(所陈述的条件或事件)”可以被解释为意思是“如果确定…”或“响应于确定…”或“在检测到(所陈述的条件或事件)时”或“响应于检测到(所陈述的条件或事件)”。
在上述实施例中,可以全部或部分地通过软件、硬件、固件或者其任意组合来实现。当使用软件实现时,可以全部或部分地以计算机程序产品的形式实现。所述计算机程序产品包括一个或多个计算机指令。在计算机上加载和执行所述计算机程序指令时,全部或部分地产生按照本申请实施例所述的流程或功能。所述计算机可以是通用计算机、专用计算机、计算机网络、或者其他可编程装置。所述计算机指令可以存储在计算机可读存储介质中,或者从一个计算机可读存储介质向另一个计算机可读存储介质传输,例如,所述计算机指令可以从一个网站站点、计算机、服务器或数据中心通过有线(例如同轴电缆、光纤、数字用户线)或无线(例如红外、无线、微波等)方式向另一个网站站点、计算机、服务器或数据中心进行传输。所述计算机可读存储介质可以是计算机能够存取的任何可用介质或者是包含一个或多个可用介质集成的服务器、数据中心等数据存储设备。所述可用介质可以是磁性介质,(例如,软盘、硬盘、磁带)、光介质(例如DVD)、或者半导体介质(例如固态硬盘)等。
本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,该流程可以由计算机程序来指令相关的硬件完成,该程序可存储于计算机可读取存储介质中,该程序在执行时,可包括如上述各方法实施例的流程。而前述的存储介质包括:ROM或随机存储记忆体RAM、磁碟或者光盘等各种可存储程序代码的介质。

Claims (13)

  1. 一种信息处理方法,其特征在于,所述方法包括:
    电子设备显示第一界面,所述第一界面包括第一区域、第二区域以及第一状态的第一控件;
    在第一时刻,所述电子设备显示第二控件和第一状态的第一控件时,所述电子设备录制第一语音,将所述第一语音转换得到的文字显示在所述第一区域;第二控件用于指示正在录音,第一状态的第一控件用于指示在录音时将录制的语音转化为文本;
    所述电子设备显示第二控件和第一状态的第一控件时,接收用户输入的第一信息,并将所述第一信息显示在所述第二区域中;
    响应于针对所述第一状态的第一控件的操作,在显示第二控件的情况下,所述电子设备将所述第一控件的状态切换为第二状态,停止将录制的语音转换为文字,所述第二状态的第一控件用于指示暂停将录制的语音转化为文本;
    所述电子设备将所述第一语音、所述第一语音转换得到的文字以及所述第一信息记录到第一文件中。
  2. 根据权利要求1所述的方法,其特征在于,所述第一区域中包括第一子区域以及第二子区域,第一子区域用于显示所述第二控件;所述第二子区域用于显示语音转换得到的文本;在所述电子设备显示第二控件时,所述方法还包括:
    响应于针对所述第二控件的操作,所述电子设备将所述第二控件替换成第三控件;所述第三控件用于指示所述电子设备结束录制语音;
    结束录制语音后,在所述电子设备未获取到所述语音转换得到的文字时,所述第一区域不包括所述第二子区域。
  3. 根据权利要求1或2所述的方法,其特征在于,所述方法还包括:
    在播放所述第一语音的情况下,所述电子设备显示进度条,所述进度条用于指示播放语音的进度;
    在所述进度条的第一位置显示第一标识的情况下,所述电子设备用第三状态显示所述第一信息,且,以第四状态显示所述第二区域中的第二信息;其中,所述第一标识在所述第一位置用于指示所述电子设备在播放所述第一时刻录制的语音;所述第二信息的输入时间为第二时刻,所述第二时刻晚于所述第一时刻。
  4. 根据权利要求3所述的方法,其特征在于,在所述进度条的第一位置显示第一标识的情况下,所述方法还包括:
    所述电子设备在所述第一区域中显示所述第一时刻录制的语音转换得到的文字。
  5. 根据权利要求3或4所述的方法,其特征在于,以第四状态显示第二区域中的第二信息之后,所述方法还包括:
    响应于针对第四控件的操作,所述电子设备以所述第三状态显示所述第二信息。
  6. 根据权利要求3-5中任一项所述的方法,其特征在于,在所述进度条的第一位置显示第一标识的情况下,所述方法还包括:
    响应于针对所述进度条中第二位置的第一操作,所述电子设备以第五状态显示第二区域中的第三信息;其中,所述第三信息的输入时间为第三时刻,且,在所述第二位置指示所述电子设备在所述第三时刻录制的语音。
  7. 根据权利要求2-6中任一项所述的方法,其特征在于,所述方法还包括:
    在建立所述第一文件后录制了Q段语音的情况下,所述电子设备在所述进度条中显示Q-1个节点;所述Q-1个节点将所述进度条划分为Q段子进度条;所述Q段子进度条中的第i段子进度条用于指示所述Q段语音中的第i段语音;所述Q为大于等于2的整数;
    在检测到针对第五控件的操作之后,在所述第一标识处于所述第i段子进度条时,所述电子设备播放所述第i段语音。
  8. 根据权利要求7所述的方法,其特征在于,所述方法还包括:
    所述电子设备选中Q段语音中L段语音;所述L为小于等于Q的整数;
    在第六控件处于第六状态的情况下,响应于针对第七控件的操作,所述电子设备删除所述L段语音,且,删除所述L段语音转换得到的文字;
    在所述第六控件不处于第六状态的情况下,响应于针对所述第七控件的操作,所述电子设备删除所述L段语音,且,保留所述L段语音转换得到的文字。
  9. 根据权利要求8所述的方法,其特征在于,所述方法还包括:
    在所述第八控件处于第八状态的情况下,响应于针对第九控件的操作,所述电子设备将所述第二区域中内容以及所述Q段语音转换得到的文字传输至其他设备;
    在所述第八控件不处于第八状态的情况下,响应于针对所述第九控件的操作,所述电子设备将所述第二区域中内容传输至其他设备。
  10. 根据权利要求1-9中任一项所述的方法,其特征在于,接收输入的第一信息,具体包括:
    所述电子设备在所述第一界面中显示第一输入工具;
    所述电子设备通过所述第一输入工具接收输入的第一信息;
    所述方法还包括:
    在显示所述第一输入工具时,还显示第十控件;
    响应于针对所述第十控件的操作,所述电子设备将所述第一输入工具切换为第二输入工具;
    在显示所述第二输入工具时,还显示第十一控件;所述第十一控件用于触发从第二输入工具切换至第一输入工具。
  11. 根据权利要求1-10中任一项所述的方法,其特征在于,所述方法还包括:
    响应于针对第十二控件的操作;所述电子设备将所述第二子区域的大小从第一尺寸变更为第二尺寸。
  12. 一种计算机存储介质,其特征在于,所述存储介质中存储有计算机程序,所述计算机程序包括可执行指令,所述可执行指令当被处理器执行时使所述处理器执行如权利要求1至11中任一项所述的方法。
  13. 一种电子设备,其特征在于,所述电子设备包括:一个或多个处理器和存储器;所述存储器与所述一个或多个处理器耦合,所述存储器用于存储计算机程序代码,所述计算机程序代码包括计算机指令,所述一个或多个处理器调用所述计算机指令以使得所述电子设备执行如权利要求1至11中任一项所述的方法。
PCT/CN2023/117685 2022-11-04 2023-09-08 信息处理方法和电子设备 WO2024093525A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202211379551.4A CN117998005A (zh) 2022-11-04 2022-11-04 信息处理方法和电子设备
CN202211379551.4 2022-11-04

Publications (1)

Publication Number Publication Date
WO2024093525A1 true WO2024093525A1 (zh) 2024-05-10

Family

ID=90901719

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/117685 WO2024093525A1 (zh) 2022-11-04 2023-09-08 信息处理方法和电子设备

Country Status (2)

Country Link
CN (1) CN117998005A (zh)
WO (1) WO2024093525A1 (zh)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103065659A (zh) * 2012-12-06 2013-04-24 广东欧珀移动通信有限公司 一种多媒体记录方法
US20130297308A1 (en) * 2012-05-07 2013-11-07 Lg Electronics Inc. Method for displaying text associated with audio file and electronic device
KR20170005590A (ko) * 2015-07-06 2017-01-16 주식회사 엘지유플러스 음성 통화 녹음 방법 및 이를 수행하는 단말
CN111819830A (zh) * 2018-09-13 2020-10-23 华为技术有限公司 一种通信过程中信息的记录及显示方法及终端
CN112866469A (zh) * 2021-01-19 2021-05-28 维沃移动通信(杭州)有限公司 通话内容的记录方法及装置
CN114512132A (zh) * 2022-01-27 2022-05-17 Oppo广东移动通信有限公司 内容显示方法、装置以及电子设备

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130297308A1 (en) * 2012-05-07 2013-11-07 Lg Electronics Inc. Method for displaying text associated with audio file and electronic device
CN103065659A (zh) * 2012-12-06 2013-04-24 广东欧珀移动通信有限公司 一种多媒体记录方法
KR20170005590A (ko) * 2015-07-06 2017-01-16 주식회사 엘지유플러스 음성 통화 녹음 방법 및 이를 수행하는 단말
CN111819830A (zh) * 2018-09-13 2020-10-23 华为技术有限公司 一种通信过程中信息的记录及显示方法及终端
CN112866469A (zh) * 2021-01-19 2021-05-28 维沃移动通信(杭州)有限公司 通话内容的记录方法及装置
CN114512132A (zh) * 2022-01-27 2022-05-17 Oppo广东移动通信有限公司 内容显示方法、装置以及电子设备

Also Published As

Publication number Publication date
CN117998005A (zh) 2024-05-07

Similar Documents

Publication Publication Date Title
JP7181320B2 (ja) 背景音楽を選択して動画を撮影する方法、装置、端末機及び媒体
US9900414B2 (en) Mobile terminal and menu control method thereof
JP7414842B2 (ja) コメント追加方法及び電子デバイス
US8032571B2 (en) Method and apparatus for automatically classifying files on mobile communication terminal
US10782856B2 (en) Method and device for displaying application function information, and terminal device
KR101569427B1 (ko) 휴대 단말기의 터치 입력 장치 및 그의 운영 방법
JP5563650B2 (ja) 音声ファイルに関連するテキストの表示方法及びこれを実現した電子機器
KR102020335B1 (ko) 메시지 운용 방법 및 이를 지원하는 단말기
KR20230042523A (ko) 멀티미디어 데이터의 처리 방법, 생성 방법 및 관련 기기
AU2017221872A1 (en) Method for providing contents, device and system for the same
US20120210201A1 (en) Operation method for memo function and portable terminal supporting the same
US20240061560A1 (en) Audio sharing method and apparatus, device and medium
KR100685975B1 (ko) 이동통신단말기에서의 통화 중 메모 작성 방법 및 이를구현할 수 있는 이동통신단말기
JP2009533780A (ja) マルチメディアモバイル装置によるノートテイキングのユーザ体験
US20230317117A1 (en) Video generation method and apparatus, device, and storage medium
WO2022121877A1 (zh) 消息处理方法、装置和电子设备
WO2023061414A1 (zh) 一种文件生成方法、装置及电子设备
JP2016524767A (ja) インスタントメッセージの管理方法、装置、プログラム及び記録媒体
WO2024078514A1 (zh) 投屏方法、装置、电子设备和存储介质
WO2024093525A1 (zh) 信息处理方法和电子设备
WO2024109699A1 (zh) 显示方法、装置、电子设备及介质
CN114298000A (zh) 基于应用的笔记生成方法、设备、存储介质及程序产品
WO2024041514A1 (zh) 视频播放方法、装置和电子设备
CN112866469A (zh) 通话内容的记录方法及装置
CN111913641A (zh) 一种实现图片语音化的方法和***

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23884458

Country of ref document: EP

Kind code of ref document: A1