WO2018120820A1 - 一种演示文稿的制作方法和装置 - Google Patents

一种演示文稿的制作方法和装置 Download PDF

Info

Publication number
WO2018120820A1
WO2018120820A1 PCT/CN2017/094599 CN2017094599W WO2018120820A1 WO 2018120820 A1 WO2018120820 A1 WO 2018120820A1 CN 2017094599 W CN2017094599 W CN 2017094599W WO 2018120820 A1 WO2018120820 A1 WO 2018120820A1
Authority
WO
WIPO (PCT)
Prior art keywords
presentation
switching
speech
time
action
Prior art date
Application number
PCT/CN2017/094599
Other languages
English (en)
French (fr)
Inventor
吴亮
黄薇
高峰
钟恒
Original Assignee
北京奇虎科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京奇虎科技有限公司 filed Critical 北京奇虎科技有限公司
Publication of WO2018120820A1 publication Critical patent/WO2018120820A1/zh

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/12Use of codes for handling textual entities
    • G06F40/14Tree-structured documents
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/12Use of codes for handling textual entities
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/166Editing, e.g. inserting or deleting

Definitions

  • the present application relates to the field of web technologies, and in particular, to a method for fabricating a presentation and a device for making a presentation.
  • the user In order to realize distance learning, the user usually records the operation of the presentation while the user is speaking, keeping the user's speech synchronized with the presentation.
  • the video data obtained by recording the operation of the presentation is bulky and takes up a lot of storage space.
  • the video data is often compressed to reduce the resolution of the video data, resulting in blurry content of the presentation.
  • the present application has been made in order to provide a method for fabricating a presentation and a corresponding apparatus for producing a presentation that overcomes the above problems or at least partially solves the above problems.
  • a method of making a presentation including:
  • a presentation switching action is configured on the presentation element to play the presentation document element in accordance with the presentation switching action.
  • a production apparatus for a presentation including:
  • a web page loading module adapted to load a web page generated for the presentation
  • a presentation element configuration module adapted to configure a presentation element in the web page
  • An audio data adding module adapted to add audio data to the presentation element on a time axis to synchronously play the audio data when the presentation element is played according to the time axis;
  • the presentation switching action configuration module is adapted to configure a presentation switching action on the presentation element to play the presentation document element according to the speech switching action.
  • a computer program comprising computer readable code causing the terminal device to perform the production of any of the aforementioned presentations when the computer readable code is run on a terminal device method.
  • a computer readable medium storing a computer program of a method of fabricating a presentation as described above.
  • the embodiment of the present application loads a web page generated for a presentation in a client, and configures a presentation element in the web page, and further adds audio data to the presentation element on the timeline, so that the presentation can be played according to the timeline
  • the elements are synchronized to play audio data
  • the web page is used as a carrier to create a presentation
  • the audio data is used to synchronize the presentation of the presentation elements and audio data, allowing the user to view the contents of the presentation and listen to the presentation of the presentation.
  • Using the web element as a presentation element compared to the video data, can greatly reduce the mention, reduce the occupation of the storage space, and, because the web element is directly drawn and loaded on the web page, without compression processing, the web element can be guaranteed.
  • Sharpness on the other hand, the presentation switching action is configured on the presentation element, so that the presentation document element can be played according to the speech switching action during the presentation, which increases the synchronization precision of the presentation and the audio data.
  • FIG. 1 is a flow chart showing the steps of an embodiment of a method for creating a presentation according to an embodiment of the present application
  • FIG. 2A-2C illustrate a configuration presentation element in accordance with one embodiment of the present application.
  • 3A-3D illustrate example diagrams of editing a presentation element and audio data playback order, in accordance with one embodiment of the present application
  • FIGS. 4A-4D illustrate example diagrams of playing presentation elements and audio data in accordance with one embodiment of the present application
  • FIGS. 5A-5B illustrate example diagrams of recording audio data in accordance with one embodiment of the present application
  • FIGS. 6A-6D illustrate example diagrams of an additional speech switching action in accordance with an embodiment of the present application
  • FIGS. 7A-7B are diagrams showing an example of deleting a speech switching action according to an embodiment of the present application.
  • FIGS. 8A-8B illustrate example diagrams of a mobile speech switching action in accordance with one embodiment of the present application
  • FIG. 9 is a structural block diagram of a device for fabricating a presentation according to an embodiment of the present application.
  • Figure 10 schematically shows a block diagram of a terminal device for performing the method according to the present application
  • Fig. 11 schematically shows a storage unit for holding or carrying program code implementing the method according to the present application.
  • FIG. 1 is a flow chart showing the steps of an embodiment of a method for creating a presentation according to an embodiment of the present application. Specifically, the method may include the following steps:
  • Step 101 Load a web page generated for the presentation.
  • the user can log in to the server by using a user account on a client such as a browser, and send a request for generating a presentation to the server.
  • a client such as a browser
  • the server can configure a new presentation and configure the presentation with a unique presentation identifier, such as slide_id (slide ID), which is used to generate a unique one for the presentation.
  • Slide_id segment ID
  • Edited URL Uniform Resource Locator, Uniform Resource Locator
  • the client accesses the URL for editing to load a web page, which is the carrier of the presentation, ie the presentation can edit the content in the web page.
  • the information of the presentation can be displayed in the area such as the user center.
  • the client can directly load the web page by using the URL for editing, which is not used in the embodiment of the present application. limit.
  • the presentation ID is used to generate a unique URL for the presentation, and the URL for the presentation is returned to the client.
  • the client can access the URL for the presentation to load the web page, which is the carrier of the presentation, ie the presentation can be played in the web page.
  • Step 102 configuring a presentation element in the web page.
  • the presentation elements can include one or more of the following:
  • Text images, images of specified shapes, lines, tables, frames, and code.
  • the user can trigger the presentation element to edit state by clicking or the like.
  • the editing operation bar of the presentation element is popped up in the web page, and the user can display the element of the presentation element in the editing operation column. Parameters for the user to adjust.
  • the edit operation column of the table may be popped up in the web page, and the user may set the number of rows, the number of columns, and the cell.
  • Element parameters such as margins, border width, and border color.
  • the user can save it manually, or the script of the client executing the web page can be automatically saved.
  • the parameters configured in the presentation element of the web page can be synchronized with the server during saving, and the server takes the parameter. Store under the presentation (represented by the presentation ID) for subsequent loading.
  • the client loads the web page with the URL for editing, and according to the previously set element parameters.
  • the corresponding presentation element is loaded for the user to continue editing. This embodiment of the present application does not limit this.
  • Step 103 adding audio data to the presentation element on a time axis to synchronously play the audio data when the presentation element is played according to the time axis.
  • the client in order to control the playing of the presentation, can configure a timeline and set the playing time of the presentation element on the timeline.
  • the user can record audio data
  • the client adds audio data to the presentation element, such as a user's speech, so that the presentation elements can be played while the audio data is being played on the time axis, so that the two can be synchronized.
  • the user can set the playing time of the presentation element. With the passage of time, when the audio data is set to be played, the speech can be set to be switched in order.
  • the manuscript elements that is, the text "Quiet Night Thinking", “Li Bai”, “Before the Moon”.
  • the timing control is displayed in the lower left corner, and as time passes, the audio presentation data is played, and the presentation document elements are switched in order, that is, the text is displayed. "Quiet night thinking”, “Li Bai”, “before the bed bright moonlight”.
  • step 103 may include the following sub-steps:
  • Sub-step S11 the recorder is called to record audio data to the presentation element.
  • the microphone can be called to collect the original audio data, and the recorder is called to record the audio data.
  • a recording control can be loaded, after clicking the recording control, recording is started, and a visual element of the audio data is displayed on the axis element of the visual axis of the time axis.
  • the sub-step S11 may include the following sub-steps:
  • Sub-step S111 acquiring original audio stream data collected by the microphone
  • Sub-step S112 the original audio stream data is transmitted to the recorder
  • Sub-step S113 the original audio stream data is visualized in the recorder according to the recording parameters, and the original audio stream data is converted into audio data of a specified format.
  • the client can obtain the original audio stream data collected by the microphone through the getUserMedia interface provided by WebRTC (Web Real-Time Communication).
  • WebRTC Web Real-Time Communication
  • a script processing node is created by the createScriptProcess method of the Web Audio API, which is used to process raw audio stream data using Javascript.
  • the audio source node is connected to the processing node, and the processing node is connected to the audio output node to form a complete processing flow.
  • the processing node can listen to the AudioProcessingEvent event through the onaudioprocess method, and the event acquires a certain length of data from the original audio stream data for processing at regular intervals.
  • the original audio stream data is visualized by the drawAudioWave method (the visualized elements are generated based on the frequency, waveform and other attributes of the original audio stream data), and the audio data is transmitted to the Web Worker for audio.
  • the drawAudioWave method the visualized elements are generated based on the frequency, waveform and other attributes of the original audio stream data
  • the audio data is transmitted to the Web Worker for audio.
  • the audio processing is paused, and a format file such as WAV is requested from the Web Worker, and the Web Worker converts the existing original audio stream data into audio data of a format such as WAV and returns it.
  • a format file such as WAV
  • the Web Worker also opens a thread to temporarily store and process the original audio stream data, so that other processing of the client (such as the browser) can be performed normally.
  • step 103 may include the following sub-steps:
  • Sub-step S21 inputting text information to the presentation element
  • Sub-step S22 converting the text information into audio data.
  • the terminal where the client is located is not configured with a microphone
  • the user can input text information to the presentation element, and the text information can be converted into audio data through the voice synthesis (The Emperor Waltz, TEW).
  • Speech synthesis also known as Text to Speech (TTS) technology
  • TTS Text to Speech
  • the characteristics of the segment such as pitch, length and intensity, are made, so that the synthesized speech can correctly express the semantics and sound more natural.
  • the phonetic primitives of the single words or phrases corresponding to the processed text are extracted from the speech synthesis library, and the prosody characteristics of the speech primitives are adjusted and modified by using a specific speech synthesis technique, and finally synthesized. Meet the required voice data.
  • the manner of adding audio data is only an example.
  • other manners of adding audio data may be set according to actual conditions, for example, directly importing existing audio data, and the like. This is not limited.
  • those skilled in the art may also adopt other manners of adding audio data according to actual needs, and the embodiment of the present application does not limit this.
  • the audio data on the time axis can be uploaded to the server.
  • the audio data can be retrieved from the Web Worker, and the audio file is compressed by the amrnb.js library, and compressed to specify the amr.
  • the format is then uploaded to the server, which is stored under the presentation (represented by the presentation ID) for subsequent loading.
  • Step 104 Configure a presentation switching action on the presentation element to play the presentation document element according to the speech switching action.
  • the user may define a series of speech switching actions in the recording editor, and play the presentation document elements according to the speech switching actions.
  • the speech switching action includes a switching time and a switching operation mode, that is, the speech switching action has a time point corresponding to the audio data, and the corresponding time point is triggered according to the time point of the audio data playing during the playing of the presentation.
  • the switching mode of operation switches the presentation document elements.
  • step 104 may include the following sub-steps:
  • Sub-step S31 receiving an increase instruction of the speech switching action
  • Sub-step S32 setting a switching time of the speech switching action and a switching operation mode of the speech document element according to the adding instruction, to perform the switching operation mode when the speech document is played to the switching time Switch the presentation document elements.
  • a speech switching action may be added, and the presentation document element is switched according to the switching operation mode when the speech presentation is played to the switching time.
  • the newly added speech switching action may include the following two categories:
  • the switching operation mode of the presentation document element is recorded to switch the presentation document element according to the switching operation mode when the speech presentation is played to the time point.
  • the recording editor will record the position of the action switching (corresponding time point) and the corresponding switching operation mode. (one action, one action as above) and displayed in real time in the visualization area.
  • a speech switching action indicator (a symbol with a triangle in a circle) indicates a speech switching action, and the switching operation mode of the speech switching action in the vicinity of 2.3 seconds is the next action, that is, the display.
  • the user can click the speech to switch the blank of the action identifier to add the speech switching action.
  • the recording editor responds to the final state of the action switching in real time and displays it in real time in the visualization area.
  • a speech switching action identifier (a symbol with a triangle in a circle) represents a speech switching action, and a speech switching action is added outside the time axis, and the switching operation mode is the next action. That is, the text "Moonlight in front of the window" is displayed, and the next action can be updated in real time, that is, the text "Moonlight in front of the window” is displayed.
  • step 104 may include the following sub-steps:
  • Sub-step S41 receiving a delete instruction of the speech action
  • Sub-step S42 deleting the switching time of the speech switching operation according to the deletion instruction, to switch the speech document element according to the switching operation mode when the speech document is played to the switching time of the previous speech switching operation.
  • the user can click the speech to switch the action identifier, pop up the delete box, and click the delete button to delete the existing talk switch action.
  • the presentation document element can also be updated in real time, and the presentation document element is updated in real time to the previous speech switching action.
  • a speech switching action identifier (a symbol with a triangle in a circle) indicates a speech switching action, and the speech switching action is deleted near 16.2 seconds, and the switching operation mode is the next action, that is, The text "Bearing Moonlight” is displayed, and the speech switching action near 16.1 seconds can be updated in real time to the next action, that is, the text "Bearing Moonlight” is displayed.
  • step 104 may include the following sub-steps:
  • Sub-step S51 receiving a movement instruction of the speech action
  • Sub-step S52 changing the switching time of the speech switching operation according to the movement instruction, and switching the presentation document element according to the switching operation mode when playing to the switching time after the change.
  • the user can change the position associated with the recorded audio (ie, the switching time) by clicking the existing speech switching action identifier and dragging the movement of the speech switching action identifier.
  • an effective time interval can be set, and the change of the switching time is at the effective time.
  • the area is valid.
  • the effective time interval may be calculated, and the effective time interval is between the switching time of the previous speech switching action and the switching time of the next speech switching action, and is not related to the switching time of the previous speech switching action, and the next The switching time of the presentation switching action overlaps.
  • a time point is determined as the switching time of the speech switching action in the effective time interval to switch the presentation document element according to the switching operation mode when playing to the time point after the change.
  • a speech switching action indicator (a symbol with a triangle in a circle) indicates a speech switching action
  • the switching operation mode of the speech switching action in the vicinity of 6.5 seconds is the next action, that is, Display the text "suspected ground frost", click on the speech to switch the action ID, calculate the effective time interval, between 5.5 seconds - 10.3 seconds, that is, the area covered by the rectangular figure, if the user moves the speech switching action flag to 8.5 seconds Nearby, when the playback is near 7.3 seconds, the text "Suspected Ground Cream" is canceled.
  • the configuration of the above-mentioned speech switching action is only an example.
  • the configuration mode of the other speech switching action may be set according to the actual situation, which is not limited by the embodiment of the present application.
  • those skilled in the art may also adopt other configuration modes of the speech switching action according to actual needs, which is not limited in the embodiment of the present application.
  • the embodiment of the present application loads a web page generated for a presentation in a client, and configures a presentation element in the web page, and further adds audio data to the presentation element on the timeline, so that the presentation can be played according to the timeline
  • the elements are synchronized to play audio data
  • the web page is used as a carrier to create a presentation
  • the audio data is used to synchronize the presentation of the presentation elements and audio data, allowing the user to view the contents of the presentation and listen to the presentation of the presentation.
  • Using the web element as a presentation element compared to the video data, can greatly reduce the mention, reduce the occupation of the storage space, and, because the web element is directly drawn and loaded on the web page, without compression processing, the web element can be guaranteed.
  • Sharpness on the other hand, the presentation switching action is configured on the presentation element, so that the presentation document element can be played according to the speech switching action during the presentation, which increases the synchronization precision of the presentation and the audio data.
  • FIG. 9 a structural block diagram of a device for creating a presentation according to an embodiment of the present application is shown, which may specifically include the following modules:
  • a web page loading module 901 configured to load a web page generated for the presentation
  • a presentation element configuration module 902 adapted to configure a presentation element in the web page
  • An audio data adding module 903, configured to add audio data to the presentation element on a time axis to synchronously play the audio data when the presentation element is played according to the time axis;
  • the presentation switching action configuration module 904 is adapted to configure a presentation switching action on the presentation element to play the presentation document element in accordance with the presentation switching action.
  • the audio data adding module 903 includes:
  • a recording sub-module adapted to call the recorder to record audio data to the presentation element.
  • the recording submodule includes:
  • the original audio stream data acquiring unit is adapted to acquire original audio stream data collected in the microphone
  • a recorder incoming unit adapted to transmit the raw audio stream data to the recorder
  • a recorder processing unit adapted to visualize the original audio stream data in the recorder according to recording parameters, and convert the original audio stream data into audio data of a specified format.
  • the audio data adding module 903 includes:
  • a text information input submodule adapted to input text information to the presentation element
  • a text information conversion sub-module adapted to convert the text information into audio data.
  • the speech switching action includes a switching time and a switching operation mode
  • the speech switching action configuration module 904 includes:
  • a speech switching action adding submodule configured to set a switching time of the speech switching action and a switching operation mode of the speech document element according to the adding instruction, to perform, when the speech presentation is played to the switching time, The switching mode of operation switches the presentation document element.
  • the speech switching action adding submodule includes:
  • a first switching time recording unit configured to record a time point of the speech switching action on the time axis as a switching time when a speech switching action is added in the time axis;
  • a first switching operation mode recording unit adapted to record a switching operation on the presentation document element a manner of switching the presentation document element according to the switching operation mode when the speech presentation is played to the time point;
  • a second switching time recording unit adapted to record an end time of the time axis as a switching time when the speech switching action is added outside the time axis;
  • the second switching operation mode recording unit is adapted to record a switching operation mode of the presentation document element to switch the presentation document element according to the switching operation mode when the speech presentation is played to the end time.
  • the speech switching action configuration module 904 further includes:
  • a speech switching action deletion sub-module configured to delete a switching time of the speech switching action according to the deletion instruction, to switch the switching presentation mode to a switching time of the last speech switching action according to the switching operation mode Presentation document elements.
  • the speech switching action configuration module 904 further includes:
  • a mobile instruction receiving submodule adapted to receive a move instruction of a speech action
  • the speech switching action moving submodule is adapted to change the switching time of the speech switching operation according to the movement instruction to switch the presentation document element according to the switching operation mode when playing to the changed switching time.
  • the speech switching action moving submodule includes:
  • the effective time interval calculation unit is adapted to calculate an effective time interval between the switching time of the last speech switching action and the switching time of the next speech switching action;
  • the switching time determining unit is adapted to determine a time point as the switching time of the speech switching action in the valid time interval to switch the presentation document element according to the switching operation mode when playing to the changed time point.
  • the method further includes:
  • An audio uploading module adapted to upload audio data on the timeline to a server.
  • the description is relatively simple, and the relevant parts can be referred to the description of the method embodiment.
  • the various component embodiments of the present application can be implemented in hardware, or in a software module running on one or more processors, or in a combination thereof. It will be understood by those skilled in the art that a microprocessor or digital signal processor (DSP) can be used in practice to implement some of the production equipment of the presentation according to embodiments of the present application or Some or all of the features of all components.
  • DSP digital signal processor
  • the application can also be implemented as a device or device program (e.g., a computer program and a computer program product) for performing some or all of the methods described herein.
  • Such a program implementing the present application may be stored on a computer readable medium or may be in the form of one or more signals. Such signals may be downloaded from an Internet website, provided on a carrier signal, or provided in any other form.
  • FIG. 10 illustrates a terminal device that can implement the production of a presentation according to the present application.
  • the terminal device conventionally includes a processor 1010 and a computer program product or computer readable medium in the form of a memory 1020.
  • the memory 1020 may be an electronic memory such as a flash memory, an EEPROM (Electrically Erasable Programmable Read Only Memory), an EPROM, a hard disk, or a ROM.
  • the memory 1020 has a memory space 1030 for executing program code 1031 of any of the above method steps.
  • storage space 1030 for program code may include various program code 1031 for implementing various steps in the above methods, respectively.
  • the program code can be read from or written to one or more computer program products.
  • Such computer program products include program code carriers such as hard disks, compact disks (CDs), memory cards or floppy disks.
  • Such a computer program product is typically a portable or fixed storage unit as described with reference to FIG.
  • the storage unit may have a storage section, a storage space, and the like arranged similarly to the storage 1020 in the terminal device of FIG.
  • the program code can be compressed, for example, in an appropriate form.
  • the storage unit comprises computer readable code 1031', ie code that can be read by, for example, a processor such as 1010, which when executed by the terminal device causes the terminal device to perform each of the methods described above step.
  • "an embodiment," or "an embodiment," or "one or more embodiments" as used herein means that the particular features, structures, or characteristics described in connection with the embodiments are included in at least one embodiment of the present application.
  • phrase "in one embodiment" is not necessarily referring to the same embodiment.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Telephonic Communication Services (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

本申请实施例提供了一种演示文稿的制作方法和装置,该方法包括:加载对演示文稿生成的web页面;在所述web页面中配置演示文稿元素;在时间轴上对所述演示文稿元素添加音频数据,以按照所述时间轴播放所述演示文稿元素时同步播放所述音频数据;对所述演示文稿元素配置演讲切换动作,以按照所述演讲切换动作播放所述演讲文稿元素。本申请实施例将web元素作为演示文稿元素,相比视频数据,可以大大减少提及,减少存储空间的占用,并且,由于web元素是直接在web页面绘制、加载,无需进行压缩处理,可以保证web元素的清晰度。

Description

一种演示文稿的制作方法和装置 技术领域
本申请涉及web技术领域,特别是涉及一种演示文稿的制作方法和一种演示文稿的制作装置。
背景技术
在人们的日常工作、学习和技术交流中,为了方便而有条理的讲解,经常采用演示文稿辅助演讲。
为了实现进行远程教学,用户在进行演讲时,通常对演示文稿的操作进行录屏,保持用户的演讲与演示文稿同步。
此后,其他用户可以播放录屏获得的视频数据,收听该用户的演讲以及观看演示文稿的内容,而无需该用户实时面对其他用户进行演讲。
但是,对演示文稿的操作进行录屏获得的视频数据,体积较大,占用大量的存储空间。
为了减少存储空间,经常会对视频数据进行压缩,降低视频数据的分辨率,导致演示文稿的内容模糊。
发明内容
鉴于上述问题,提出了本申请以便提供一种克服上述问题或者至少部分地解决上述问题的一种演示文稿的制作方法和相应的一种演示文稿的制作装置。
依据本申请的一个方面,提供了一种演示文稿的制作方法,包括:
加载对演示文稿生成的web页面;
在所述web页面中配置演示文稿元素;
在时间轴上对所述演示文稿元素添加音频数据,以按照所述时间轴播放所述演示文稿元素时同步播放所述音频数据;
对所述演示文稿元素配置演讲切换动作,以按照所述演讲切换动作播放所述演讲文稿元素。
根据本申请的另一方面,提供了一种演示文稿的制作装置,包括:
web页面加载模块,适于加载对演示文稿生成的web页面;
演示文稿元素配置模块,适于在所述web页面中配置演示文稿元素;
音频数据添加模块,适于在时间轴上对所述演示文稿元素添加音频数据,以按照所述时间轴播放所述演示文稿元素时同步播放所述音频数据;
演讲切换动作配置模块,适于对所述演示文稿元素配置演讲切换动作,以按照所述演讲切换动作播放所述演讲文稿元素。
根据本申请的另一方面,提供了一种计算机程序,包括计算机可读代码,当所述计算机可读代码在终端设备上运行时,导致所述终端设备执行前述任一个所述演示文稿的制作方法。
根据本申请的另一方面,提供了一种计算机可读介质,其中存储了如前述演示文稿的制作方法的计算机程序。
本申请实施例在客户端中加载对演示文稿生成的web页面,并在该web页面中配置演示文稿元素,此外,在时间轴上对演示文稿元素添加音频数据,使得可以按照时间轴播放演示文稿元素时同步播放音频数据,将web页面作为载体制作演示文稿,并配以音频数据,实现在同步播放演示文稿元素及音频数据,可供用户观看演示文稿的内容、收听演示文稿的演讲,一方面,将web元素作为演示文稿元素,相比视频数据,可以大大减少提及,减少存储空间的占用,并且,由于web元素是直接在web页面绘制、加载,无需进行压缩处理,可以保证web元素的清晰度;另一方面,对演示文稿元素配置演讲切换动作,使得在演播时,可以按照演讲切换动作播放演讲文稿元素,增加了演示文稿与音频数据的同步精确度。
上述说明仅是本申请技术方案的概述,为了能够更清楚了解本申请的技术手段,而可依照说明书的内容予以实施,并且为了让本申请的上述和其它目的、特征和优点能够更明显易懂,以下特举本申请的具体实施方式。
附图说明
通过阅读下文优选实施方式的详细描述,各种其他的优点和益处对于本领域普通技术人员将变得清楚明了。附图仅用于示出优选实施方式的目的,而并不认为是对本申请的限制。而且在整个附图中,用相同的参考符号表示相同的部件。在附图中:
图1示出了根据本申请一个实施例的一种演示文稿的制作方法实施例的步骤流程图;
图2A-图2C示出了根据本申请一个实施例的一种配置演示文稿元素的 示例图;
图3A-图3D示出了根据本申请一个实施例的一种编辑演示文稿元素与音频数据播放顺序的示例图;
图4A-图4D示出了根据本申请一个实施例的一种播放演示文稿元素与音频数据的示例图;
图5A-图5B示出了根据本申请一个实施例的一种录制音频数据的示例图;
图6A-图6D示出了根据本申请一个实施例的一种增加演讲切换动作的示例图;
图7A-图7B示出了根据本申请一个实施例的一种删除演讲切换动作的示例图;
图8A-图8B示出了根据本申请一个实施例的一种移动演讲切换动作的示例图;以及
图9示出了根据本申请一个实施例的一种演示文稿的制作装置的结构框图;
图10示意性地示出了用于执行根据本申请的方法的终端设备的框图;
图11示意性地示出了用于保持或者携带实现根据本申请的方法的程序代码的存储单元。
具体实施方式
下面将参照附图更详细地描述本公开的示例性实施例。虽然附图中显示了本公开的示例性实施例,然而应当理解,可以以各种形式实现本公开而不应被这里阐述的实施例所限制。相反,提供这些实施例是为了能够更透彻地理解本公开,并且能够将本公开的范围完整的传达给本领域的技术人员。
参照图1,示出了根据本申请一个实施例的一种演示文稿的制作方法实施例的步骤流程图,具体可以包括如下步骤:
步骤101,加载对演示文稿生成的web页面。
在本申请实施例中,用户可以在浏览器等客户端采用用户账号登录服务器,并向服务器发送演示文稿的生成请求。
服务器若接收到客户端发送的生成请求,则可以配置一个新的演示文稿,并对该演示文稿配置唯一的演示文稿标识,如slide_id(幻灯片ID),采用该演示文稿标识生成唯一的用于编辑的URL(Uniform Resource Locator, 统一资源定位符),并将该用于编辑的URL返回客户端。
例如,假设某个演示文稿标识的slide_id为abcd1234,则可以生成用于编辑的URL为https://ppt.***.com/editor?slide_id=abcd1234,其中,“***”为该服务器的域名。
客户端访问该用于编辑的URL加载web(网页)页面,该web页面为演示文稿的载体,即演示文稿可以在该web页面中编辑内容。
当然,若在先已生成演示文稿,则在用户中心等区域可以显示该演示文稿的信息,用户点击之后,客户端可以直接采用用于编辑的URL加载web页面,本申请实施例对此不加以限制。
此外,当演示文稿编辑完成之后,采用演示文稿标识生成唯一的用于演讲的URL,并将该用于演讲的URL返回客户端。
例如,假设某个演示文稿标识的slide_id为abcd1234,则可以生成用于编辑的URL为http://ppt.***.com/present?slide_id=abcd1234#/,其中,“***”为该服务器的域名。
客户端可以访问该用于演讲的URL加载web页面,该web页面为演示文稿的载体,即可以在该web页面中播放演示文稿。
步骤102,在所述web页面中配置演示文稿元素。
在web页面中具有一个或多个web操作控件,用户可以通过鼠标点击等方式触发该web操作控件,从而添加演示文稿元素。
在一个示例中,如图2A所示,演示文稿元素可以包括如下的一种或多种:
文字、图片、指定形状的图像、直线、表格、框架(iframe)、代码。
在web页面中,可以直接对这些演示文稿元素进行操作,编辑部分元素参数,例如,调整字体的文本框、图片、图像、直线、表格框架、代码等演示文稿元素在web页面的大小,旋转字体的文本框、图片、图像、直线、表格框架、代码等演示文稿元素,拖动字体的文本框、图片、图像、直线、表格框架、代码等演示文稿元素在web页面的位置,在文字的文本框、表格等演示文稿元素中输入文本,等等。
此外,用户可以通过点击等方式触发该演示文稿元素进行编辑状态,此时,在web页面中弹出该演示文稿元素的编辑操作栏,用户可以在该编辑操作栏中,显示该演示文稿元素的元素参数,以供用户进行调整。
例如,如图2B所示,如果在web页面中***了文字的文本框,若用户 点击该文本框,则可以在web页面中弹出该文本框的编辑操作栏,用户可以设置字体对齐方式、字体播放倍数、字体颜色、行间距、字体间距等元素参数。
又例如,如图2C所示,如果在web页面中***了表格,若用户点击该表格,则可以在web页面中弹出该表格的编辑操作栏,用户可以设置行数、列数、单元格内边距、边框宽度、边框颜色等元素参数。
需要说明的是,这些元素参数与可直接对演示文稿元素进行的编辑的种类可以相同,也可以不相同,本申请实施例对此不加以限制。
对于演示文稿元素配置的编辑,用户可以手动保存,也可以客户端执行web页面的脚本自动保存,在web页面中对演示文稿元素配置的参数,可以在保存时与服务器进行同步,服务器将该参数存储至该演示文稿(以演示文稿标识表示)之下,以便后续进行加载。
当然,若在先已生成演示文稿,则在用户中心等区域可以显示该演示文稿的信息,用户点击之后,客户端在采用用于编辑的URL加载web页面的同时,按照在先设置的元素参数加载相应的演示文稿元素,以供用户继续进行编辑,本申请实施例对此不加以限制。
步骤103,在时间轴上对所述演示文稿元素添加音频数据,以按照所述时间轴播放所述演示文稿元素时同步播放所述音频数据。
在本申请实施例中,为了控制演示文稿的播放,客户端可以配置一时间轴,并设置演示文稿元素在该时间轴的播放时间。
此外,用户可以录制音频数据,客户端对演示文稿元素添加音频数据,例如用户的演讲讲话,以便按照时间轴在播放音频数据进行演讲等操作的同时,播放演示文稿元素,使得两者可以同步。
在编辑状态下,如图3A、图3B、图3C、图3D所示,用户可以设置演示文稿元素的播放时间,随着时间的推移,在设置播放音频数据的同时,可以设置按照顺序切换演讲文稿元素,即显示文本“静夜思”、“李白”、“床前明月光”。
在演讲状态下,如图4A、图4B、图4C、图4D所示,左下角显示计时的控件,随着时间的推移,在播放音频数据的同时,按照顺序切换演讲文稿元素,即显示文本“静夜思”、“李白”、“床前明月光”。
在本申请的一个实施例中,步骤103可以包括如下子步骤:
子步骤S11,对所述演示文稿元素调用录音器录制音频数据。
在本申请实施例中,若客户端所处的终端配置有麦克风,则可以调用该麦克风采集原始的音频数据,并调用录音器录制成音频数据。
例如,如图5A和图5B所示,在web页面中,可以加载一录音控件,点击该录音控件之后,开始录音,并生成时间轴的可视化的轴元素上,显示音频数据的可视化的音频元素。
在本申请实施例的一个示例中,子步骤S11可以包括如下子步骤:
子步骤S111,获取麦克风采集的原始音频流数据;
子步骤S112,将所述原始音频流数据传入录音器;
子步骤S113,在所述录音器中按照录音参数对所述原始音频流数据进行可视化处理,并将所述原始音频流数据转换为指定格式的音频数据。
在本示例中,客户端可以通过WebRTC(Web Real-Time Communication,源自网页实时通信)提供的getUserMedia接口,获取麦克风采集到的原始音频流数据。
通过Web Audio API(Application Programming Interface,应用程序编程接口)的createMediaStreamSource方法创建一个音频流源节点,并将该音频源传入录音器Recorder中,进行录音器的初始化。
在录音器中,通过Web Audio API的createScriptProcess方法创建一个脚本处理节点,该处理节点用于使用Javascript处理原始音频流数据。
此外,将音频源节点与处理节点相连,处理节点与音频输出节点相连,形成一个完整的处理流程。
处理节点可以通过onaudioprocess方法监听AudioProcessingEvent事件,该事件每隔一定时间从原始音频流数据中获取一定长度的数据进行处理。
开始录音时,向Web Worker传递录音相关的参数(如开始的位置、选区的大小、选区右侧剩余的宽度),并通知onaudioprocess进行音频处理。
参数设置完毕后,在onaudioprocess中,通过drawAudioWave方法对原始音频流数据进行可视化处理(基于原始音频流数据的频率、波形等属性生成可视化的元素),并将音频数据传递到Web Worker,进行音频的暂存与初步处理。
暂停录音时,暂停音频处理,并向Web Worker请求WAV等格式文件,Web Worker将已有的原始音频流数据为WAV等格式的音频数据并传回。
在本示例中,由于客户端(如浏览器)的计算能力大多比较有限,而对原始音频流数据的暂存以及处理一般需要很大的计算能力,因此,通过引入 Web Worker另开启一个线程,进行原始音频流数据的暂存以及处理,从而保证客户端(如浏览器)的其他处理能够正常进行。
在本申请的另一个实施例中,步骤103可以包括如下子步骤:
子步骤S21,对所述演示文稿元素输入文本信息;
子步骤S22,将所述文本信息转换为音频数据。
在本申请实施例中,若客户端所处的终端未配置麦克风,则可以让用户对演示文稿元素输入文本信息,可以通过语音合成(The Emperor Waltz,TEW)将文本信息转换为音频数据。
语音合成,又称文语转换(Text to Speech,TTS)技术,通常可以包括如下处理:
(1)文本分析
对输入的文本信息进行语言学分析(模拟人对自然语言的理解过程),逐句进行词汇的、语法的和语义的分析,以确定句子的低层结构和每个字的音素的组成,包括文本的断句、字词切分、多音字的处理、数字的处理、缩略语的处理等。使计算机对输入的文本能完全理解,并给出后两部分所需要的各种发音提示。
(2)韵律建模
为合成语音规划出音段特征,如音高、音长和音强等,使合成语音能正确表达语意,听起来更加自然。
(3)语音合成(核心模块)
根据韵律建模的结果,把处理好的文本所对应的单字或短语的语音基元从语音合成库中提取,利用特定的语音合成技术对语音基元进行韵律特性的调整和修改,最终合成出符合要求的语音数据。
当然,上述添加音频数据的方式只是作为示例,在实施本申请实施例时,可以根据实际情况设置其他添加音频数据的方式,例如,直接导入已有的音频数据,等等,本申请实施例对此不加以限制。另外,除了上述判断处理方法外,本领域技术人员还可以根据实际需要采用其它添加音频数据的方式,本申请实施例对此也不加以限制。
在本申请实施例中,当添加音频数据完成时,可以将时间轴上的音频数据上传至服务器。
在具体实现中,对于录制音频数据的情况,可以从Web Worker取回音频数据,并通过amrnb.js库等方式进行音频文件的压缩,压缩为amr等指定 的格式,然后上传至服务器,服务器存储至该演示文稿(以演示文稿标识表示)之下,以便后续进行加载。
步骤104,对所述演示文稿元素配置演讲切换动作,以按照所述演讲切换动作播放所述演讲文稿元素。
在本申请实施例中,用户在制作演示文稿的过程中,可以在录音编辑器中定义一系列的演讲切换动作,按照这些演讲切换动作播放演讲文稿元素。
在具体实现中,演讲切换动作包括切换时间和切换操作方式,即演讲切换动作具有一个和音频数据对应的时间点,在演示文稿的播放过程中,根据音频数据播放的时间点来触发对应时间点的切换操作方式切换演讲文稿元素。
在本申请的一个实施例中,步骤104可以包括如下子步骤:
子步骤S31,接收演讲切换动作的增加指令;
子步骤S32,按照所述增加指令设置所述演讲切换动作的切换时间和对所述演讲文稿元素的切换操作方式,以在所述演讲文稿播放至所述切换时间时、按照所述切换操作方式切换所述演讲文稿元素。
在本申请实施例中,可以新增演讲切换动作,在演讲文稿播放至切换时间时、按照切换操作方式切换演讲文稿元素。
在具体实现中,新增收演讲切换动作可以包括如下两类:
一、在录音过程中添加动作
在此类方式中,在时间轴内增加演讲切换动作时,记录演讲切换动作在时间轴的时间点,作为切换时间。
记录对演讲文稿元素的切换操作方式,以在演讲文稿播放至所述时间点时、按照切换操作方式切换演讲文稿元素。
在实际应用中,用户在录音的过程中,可以通过左右键、空格键等进行演讲切换动作的切换,此时录音编辑器会记录动作切换的位置(对应的时间点)及对应的切换操作方式(如上一个动作、下一个动作),并在可视化区域实时显示出来。
例如,如图6A和图6B所示,一个演讲切换动作标识(圆形中带三角形的符号)表示一个演讲切换动作,在2.3秒附近的演讲切换动作的切换操作方式为下一个动作、即显示文本“静夜思”,在3.6秒秒附近的演讲切换动作的切换操作方式为下一个动作、即显示文本“李白”。
二、在非录音过程中添加动作
在此类方式中,在时间轴外增加演讲切换动作时,记录时间轴的结束时间,作为切换时间。
记录对演讲文稿元素的操作方式,以在演讲文稿播放至结束时间时、按照切换操作方式切换演讲文稿元素。
在实际应用中,用户可以通过点击演讲切换动作标识的空白处,进行演讲切换动作的添加,演讲切换动作添加后,录音编辑器实时响应动作切换的最终状态,并在可视化区域实时显示出来。
例如,如图6C和图6D所示,一个演讲切换动作标识(圆形中带三角形的符号)表示一个演讲切换动作,在时间轴之外增加一个演讲切换动作,其切换操作方式为下一个动作、即显示文本“窗前明月光”,即可实时更新到下一个动作、即显示文本“窗前明月光”。
在本申请的另一个实施例中,步骤104可以包括如下子步骤:
子步骤S41,接收演讲动作的删除指令;
子步骤S42,按照所述删除指令删除所述演讲切换动作的切换时间,以在所述演讲文稿播放至上一个演讲切换动作的切换时间时、按照所述切换操作方式切换所述演讲文稿元素。
在本申请实施例中,用户可以点击演讲切换动作标识,弹出删除框,点击删除按钮,则可以删除已有演讲切换动作。
删除演讲切换动作之后,也可以实时进行演讲文稿元素的更新,将演讲文稿元素实时更新到上一个演讲切换动作。
例如,如图7A和图7B所示,一个演讲切换动作标识(圆形中带三角形的符号)表示一个演讲切换动作,在16.2秒附近删除演讲切换动作,其切换操作方式为下一个动作、即显示文本“窗前明月光”,即可在16.1秒附近的演讲切换动作实时更新到下一个动作、即显示文本“窗前明月光”。
在本申请的另一个实施例中,步骤104可以包括如下子步骤:
子步骤S51,接收演讲动作的移动指令;
子步骤S52,按照所述移动指令变更所述演讲切换动作的切换时间,以在播放至变更后的切换时间时、按照所述切换操作方式切换所述演讲文稿元素。
在本申请实施例中,用户可以通过点击已有演讲切换动作标识并拖动演讲切换动作标识的移动,改变与录制音频关联的位置(即切换时间)。
在一种情况中,可以设置有效时间区间,切换时间的更改在该有效时间 区内有效。
在具体实现中,可以计算有效时间区间,有效时间区间位于上一个演讲切换动作的切换时间与下一个演讲切换动作的切换时间之间,并不与于上一个演讲切换动作的切换时间、下一个演讲切换动作的切换时间重叠。
在有效时间区间中确定一时间点作为演讲切换动作的切换时间,以在播放至变更后的时间点时、按照切换操作方式切换演讲文稿元素。
例如,如图8A和图8B所示,一个演讲切换动作标识(圆形中带三角形的符号)表示一个演讲切换动作,时间在6.5秒附近的演讲切换动作的切换操作方式为下一个动作、即显示文本“疑似地上霜”,点击该演讲切换动作标识时,计算有效时间区间,在5.5秒-10.3秒之间,即被矩形图形覆盖的区域,若用户将该演讲切换动作标识移动至8.5秒附近,则在播放至7.3秒附近时,取消显示文本“疑似地上霜”。
当然,上述演讲切换动作的配置方式只是作为示例,在实施本申请实施例时,可以根据实际情况设置其他演讲切换动作的配置方式,本申请实施例对此不加以限制。另外,除了上述演讲切换动作的配置方式外,本领域技术人员还可以根据实际需要采用其它演讲切换动作的配置方式,本申请实施例对此也不加以限制。
本申请实施例在客户端中加载对演示文稿生成的web页面,并在该web页面中配置演示文稿元素,此外,在时间轴上对演示文稿元素添加音频数据,使得可以按照时间轴播放演示文稿元素时同步播放音频数据,将web页面作为载体制作演示文稿,并配以音频数据,实现在同步播放演示文稿元素及音频数据,可供用户观看演示文稿的内容、收听演示文稿的演讲,一方面,将web元素作为演示文稿元素,相比视频数据,可以大大减少提及,减少存储空间的占用,并且,由于web元素是直接在web页面绘制、加载,无需进行压缩处理,可以保证web元素的清晰度;另一方面,对演示文稿元素配置演讲切换动作,使得在演播时,可以按照演讲切换动作播放演讲文稿元素,增加了演示文稿与音频数据的同步精确度。
对于方法实施例,为了简单描述,故将其都表述为一系列的动作组合,但是本领域技术人员应该知悉,本申请实施例并不受所描述的动作顺序的限制,因为依据本申请实施例,某些步骤可以采用其他顺序或者同时进行。其次,本领域技术人员也应该知悉,说明书中所描述的实施例均属于优选实施 例,所涉及的动作并不一定是本申请实施例所必须的。
参照图9,示出了根据本申请一个实施例的一种演示文稿的制作装置的结构框图,具体可以包括如下模块:
web页面加载模块901,适于加载对演示文稿生成的web页面;
演示文稿元素配置模块902,适于在所述web页面中配置演示文稿元素;
音频数据添加模块903,适于在时间轴上对所述演示文稿元素添加音频数据,以按照所述时间轴播放所述演示文稿元素时同步播放所述音频数据;
演讲切换动作配置模块904,适于对所述演示文稿元素配置演讲切换动作,以按照所述演讲切换动作播放所述演讲文稿元素。
在本申请的一个实施例中,所述音频数据添加模块903包括:
录音子模块,适于对所述演示文稿元素调用录音器录制音频数据。
在本申请的一个实施例中,所述录音子模块包括:
原始音频流数据获取单元,适于获取麦克风中采集的原始音频流数据;
录音器传入单元,适于将所述原始音频流数据传入录音器;
录音器处理单元,适于在所述录音器中按照录音参数对所述原始音频流数据进行可视化处理,并将所述原始音频流数据转换为指定格式的音频数据。
在本申请的一个实施例中,所述音频数据添加模块903包括:
文本信息输入子模块,适于对所述演示文稿元素输入文本信息;
文本信息转换子模块,适于将所述文本信息转换为音频数据。
在本申请的一个实施例中,所述演讲切换动作包括切换时间和切换操作方式;
所述演讲切换动作配置模块904包括:
增加指令接收子模块,适于接收演讲切换动作的增加指令;
演讲切换动作增加子模块,适于按照所述增加指令设置所述演讲切换动作的切换时间和对所述演讲文稿元素的切换操作方式,以在所述演讲文稿播放至所述切换时间时、按照所述切换操作方式切换所述演讲文稿元素。
在本申请的一个实施例中,所述演讲切换动作增加子模块包括:
第一切换时间记录单元,适于在所述时间轴内增加演讲切换动作时,记录所述演讲切换动作在所述时间轴的时间点,作为切换时间;
第一切换操作方式记录单元,适于记录对所述演讲文稿元素的切换操作 方式,以在所述演讲文稿播放至所述时间点时、按照所述切换操作方式切换所述演讲文稿元素;
或者,
第二切换时间记录单元,适于在所述时间轴外增加演讲切换动作时,记录所述时间轴的结束时间,作为切换时间;
第二切换操作方式记录单元,适于记录对所述演讲文稿元素的切换操作方式,以在所述演讲文稿播放至所述结束时间时、按照所述切换操作方式切换所述演讲文稿元素。
在本申请的一个实施例中,所述演讲切换动作配置模块904还包括:
删除指令接收子模块,适于接收演讲动作的删除指令;
演讲切换动作删除子模块,适于按照所述删除指令删除所述演讲切换动作的切换时间,以在所述演讲文稿播放至上一个演讲切换动作的切换时间时、按照所述切换操作方式切换所述演讲文稿元素。
在本申请的一个实施例中,所述演讲切换动作配置模块904还包括:
移动指令接收子模块,适于接收演讲动作的移动指令;
演讲切换动作移动子模块,适于按照所述移动指令变更所述演讲切换动作的切换时间,以在播放至变更后的切换时间时、按照所述切换操作方式切换所述演讲文稿元素。
在本申请的一个实施例中,所述演讲切换动作移动子模块包括:
有效时间区间计算单元,适于计算有效时间区间,所述有效时间区间位于上一个演讲切换动作的切换时间与下一个演讲切换动作的切换时间之间;
切换时间确定单元,适于在所述有效时间区间中确定一时间点作为演讲切换动作的切换时间,以在播放至变更后的时间点时、按照所述切换操作方式切换所述演讲文稿元素。
在本申请的一个实施例中,还包括:
音频上传模块,适于将所述时间轴上的音频数据上传至服务器。
对于装置实施例而言,由于其与方法实施例基本相似,所以描述的比较简单,相关之处参见方法实施例的部分说明即可。
本申请的各个部件实施例可以以硬件实现,或者以在一个或者多个处理器上运行的软件模块实现,或者以它们的组合实现。本领域的技术人员应当理解,可以在实践中使用微处理器或者数字信号处理器(DSP)来实现根据本申请实施例的演示文稿的制作设备中的一些或者 全部部件的一些或者全部功能。本申请还可以实现为用于执行这里所描述的方法的一部分或者全部的设备或者装置程序(例如,计算机程序和计算机程序产品)。这样的实现本申请的程序可以存储在计算机可读介质上,或者可以具有一个或者多个信号的形式。这样的信号可以从因特网网站上下载得到,或者在载体信号上提供,或者以任何其他形式提供。
例如,图10示出了可以实现根据本申请的演示文稿的制作的终端设备。该终端设备传统上包括处理器1010和以存储器1020形式的计算机程序产品或者计算机可读介质。存储器1020可以是诸如闪存、EEPROM(电可擦除可编程只读存储器)、EPROM、硬盘或者ROM之类的电子存储器。存储器1020具有用于执行上述方法中的任何方法步骤的程序代码1031的存储空间1030。例如,用于程序代码的存储空间1030可以包括分别用于实现上面的方法中的各种步骤的各个程序代码1031。这些程序代码可以从一个或者多个计算机程序产品中读出或者写入到这一个或者多个计算机程序产品中。这些计算机程序产品包括诸如硬盘,紧致盘(CD)、存储卡或者软盘之类的程序代码载体。这样的计算机程序产品通常为如参考图11所述的便携式或者固定存储单元。该存储单元可以具有与图10的终端设备中的存储器1020类似布置的存储段、存储空间等。程序代码可以例如以适当形式进行压缩。通常,存储单元包括计算机可读代码1031’,即可以由例如诸如1010之类的处理器读取的代码,这些代码当由终端设备运行时,导致该终端设备执行上面所描述的方法中的各个步骤。
本文中所称的“一个实施例”、“实施例”或者“一个或者多个实施例”意味着,结合实施例描述的特定特征、结构或者特性包括在本申请的至少一个实施例中。此外,请注意,这里“在一个实施例中”的词语例子不一定全指同一个实施例。
在此处所提供的说明书中,说明了大量具体细节。然而,能够理解,本申请的实施例可以在没有这些具体细节的情况下被实践。在一些实例中,并未详细示出公知的方法、结构和技术,以便不模糊对本说明书的理解。
应该注意的是上述实施例对本申请进行说明而不是对本申请进行限制,并且本领域技术人员在不脱离所附权利要求的范围的情况下可设计出替换实施例。在权利要求中,不应将位于括号之间的任何参考符号构造成对权利要求的限制。单词“包含”不排除存在未列在权利要求中的元件或步骤。位于元件之前的单词“一”或“一个”不排除存在多个这样的元件。本申请可以借助于包括有若干不同元件的硬件以及借助于适当编程的计算机来实现。在列举了若干装置的单元权利要求中,这些装置中的若干个可以是通过同一个硬件项来具体体现。单词第一、第二、以及第三等的使用不表示任何顺序。可将这些单词解释为名称。
此外,还应当注意,本说明书中使用的语言主要是为了可读性和教导的目的而选择的,而不是为了解释或者限定本申请的主题而选择的。因此,在不偏离所附权利要求书的范围和精神的情况下,对于本技术领域的普通技术人员来说许多修改和变更都是显而易见的。对于本申请的范围,对本申请所做的公开是说明性的,而非限制性的,本申请的范围由所附权利要求书限定。

Claims (22)

  1. 一种演示文稿的制作方法,包括:
    加载对演示文稿生成的web页面;
    在所述web页面中配置演示文稿元素;
    在时间轴上对所述演示文稿元素添加音频数据,以按照所述时间轴播放所述演示文稿元素时同步播放所述音频数据;
    对所述演示文稿元素配置演讲切换动作,以按照所述演讲切换动作播放所述演讲文稿元素。
  2. 如权利要求1所述的方法,其中,所述在时间轴上对所述演示文稿元素添加音频数据的步骤包括:
    对所述演示文稿元素调用录音器录制音频数据。
  3. 如权利要求2所述的方法,其中,所述对所述演示文稿元素调用录音器录制音频数据的步骤包括:
    获取麦克风中采集的原始音频流数据;
    将所述原始音频流数据传入录音器;
    在所述录音器中按照录音参数对所述原始音频流数据进行可视化处理,并将所述原始音频流数据转换为指定格式的音频数据。
  4. 如权利要求1所述的方法,其中,所述在时间轴上对所述演示文稿元素添加音频数据的步骤包括:
    对所述演示文稿元素输入文本信息;
    将所述文本信息转换为音频数据。
  5. 如权利要求1或2或3或4所述的方法,其中,所述演讲切换动作包括切换时间和切换操作方式;
    所述对所述演示文稿元素配置演讲切换动作,以按照所述演讲切换动作播放所述演讲文稿元素的步骤包括:
    接收演讲切换动作的增加指令;
    按照所述增加指令设置所述演讲切换动作的切换时间和对所述演讲文稿元素的切换操作方式,以在所述演讲文稿播放至所述切换时间时、按照所述切换操作方式切换所述演讲文稿元素。
  6. 如权利要求5所述的方法,其中,所述按照所述增加指令设置所述演讲切换动作的切换时间和对所述演讲文稿元素的切换操作方式,以在所述演讲文稿播放至所述切换时间时、按照所述切换操作方式切换所述演讲文稿元素的步骤包括:
    在所述时间轴内增加演讲切换动作时,记录所述演讲切换动作在所述时间轴的时间点,作为切换时间;
    记录对所述演讲文稿元素的切换操作方式,以在所述演讲文稿播放至所述时间点时、按照所述切换操作方式切换所述演讲文稿元素;
    或者,
    在所述时间轴外增加演讲切换动作时,记录所述时间轴的结束时间,作为切换时间;
    记录对所述演讲文稿元素的切换操作方式,以在所述演讲文稿播放至所述结束时间时、按照所述切换操作方式切换所述演讲文稿元素。
  7. 如权利要求5所述的方法,其中,所述对所述演示文稿元素配置演讲切换动作,以按照所述演讲切换动作播放所述演讲文稿元素的步骤还包括:
    接收演讲动作的删除指令;
    按照所述删除指令删除所述演讲切换动作的切换时间,以在所述演讲文稿播放至上一个演讲切换动作的切换时间时、按照所述切换操作方式切换所述演讲文稿元素。
  8. 如权利要求5所述的方法,其中,所述对所述演示文稿元素配置演讲切换动作,以按照所述演讲切换动作播放所述演讲文稿元素的步骤还包括:
    接收演讲动作的移动指令;
    按照所述移动指令变更所述演讲切换动作的切换时间,以在播放至变更后的切换时间时、按照所述切换操作方式切换所述演讲文稿元素。
  9. 如权利要求8所述的方法,其中,所述按照所述移动指令变更所述演讲切换动作的切换时间,以在所述演讲文稿播放至变更后的切换时间时、 按照所述操作方式切换所述演讲文稿元素的步骤包括:
    计算有效时间区间,所述有效时间区间位于上一个演讲切换动作的切换时间与下一个演讲切换动作的切换时间之间;
    在所述有效时间区间中确定一时间点作为演讲切换动作的切换时间,以在播放至变更后的时间点时、按照所述切换操作方式切换所述演讲文稿元素。
  10. 如权利要求1或2或3或4或6或7或8或9所述的方法,其中,还包括:
    将所述时间轴上的音频数据上传至服务器。
  11. 一种演示文稿的制作装置,包括:
    web页面加载模块,适于加载对演示文稿生成的web页面;
    演示文稿元素配置模块,适于在所述web页面中配置演示文稿元素;
    音频数据添加模块,适于在时间轴上对所述演示文稿元素添加音频数据,以按照所述时间轴播放所述演示文稿元素时同步播放所述音频数据;
    演讲切换动作配置模块,适于对所述演示文稿元素配置演讲切换动作,以按照所述演讲切换动作播放所述演讲文稿元素。
  12. 如权利要求11所述的装置,其中,所述音频数据添加模块包括:
    录音子模块,适于对所述演示文稿元素调用录音器录制音频数据。
  13. 如权利要求12所述的装置,其中,所述录音子模块包括:
    原始音频流数据获取单元,适于获取麦克风中采集的原始音频流数据;
    录音器传入单元,适于将所述原始音频流数据传入录音器;
    录音器处理单元,适于在所述录音器中按照录音参数对所述原始音频流数据进行可视化处理,并将所述原始音频流数据转换为指定格式的音频数据。
  14. 如权利要求11所述的装置,其中,所述音频数据添加模块包括:
    文本信息输入子模块,适于对所述演示文稿元素输入文本信息;
    文本信息转换子模块,适于将所述文本信息转换为音频数据。
  15. 如权利要求11或12或13或14所述的装置,其中,所述演讲切换动作包括切换时间和切换操作方式;
    所述演讲切换动作配置模块包括:
    增加指令接收子模块,适于接收演讲切换动作的增加指令;
    演讲切换动作增加子模块,适于按照所述增加指令设置所述演讲切换动作的切换时间和对所述演讲文稿元素的切换操作方式,以在所述演讲文稿播放至所述切换时间时、按照所述切换操作方式切换所述演讲文稿元素。
  16. 如权利要求15所述的装置,其中,所述演讲切换动作增加子模块包括:
    第一切换时间记录单元,适于在所述时间轴内增加演讲切换动作时,记录所述演讲切换动作在所述时间轴的时间点,作为切换时间;
    第一切换操作方式记录单元,适于记录对所述演讲文稿元素的切换操作方式,以在所述演讲文稿播放至所述时间点时、按照所述切换操作方式切换所述演讲文稿元素;
    或者,
    第二切换时间记录单元,适于在所述时间轴外增加演讲切换动作时,记录所述时间轴的结束时间,作为切换时间;
    第二切换操作方式记录单元,适于记录对所述演讲文稿元素的切换操作方式,以在所述演讲文稿播放至所述结束时间时、按照所述切换操作方式切换所述演讲文稿元素。
  17. 如权利要求15所述的装置,其中,所述演讲切换动作配置模块还包括:
    删除指令接收子模块,适于接收演讲动作的删除指令;
    演讲切换动作删除子模块,适于按照所述删除指令删除所述演讲切换动作的切换时间,以在所述演讲文稿播放至上一个演讲切换动作的切换时间时、按照所述切换操作方式切换所述演讲文稿元素。
  18. 如权利要求15所述的装置,其中,所述演讲切换动作配置模块还包括:
    移动指令接收子模块,适于接收演讲动作的移动指令;
    演讲切换动作移动子模块,适于按照所述移动指令变更所述演讲切换动作的切换时间,以在播放至变更后的切换时间时、按照所述切换操作方式切换所述演讲文稿元素。
  19. 如权利要求18所述的装置,其中,所述演讲切换动作移动子模块包括:
    有效时间区间计算单元,适于计算有效时间区间,所述有效时间区间位于上一个演讲切换动作的切换时间与下一个演讲切换动作的切换时间之间;
    切换时间确定单元,适于在所述有效时间区间中确定一时间点作为演讲切换动作的切换时间,以在播放至变更后的时间点时、按照所述切换操作方式切换所述演讲文稿元素。
  20. 如权利要求11或12或13或14或16或17或18或19所述的装置,其中,还包括:
    音频上传模块,适于将所述时间轴上的音频数据上传至服务器。
  21. 一种计算机程序,包括计算机可读代码,当所述计算机可读代码在终端设备上运行时,导致所述终端设备执行根据权利要求1-10中的任一个所述的演示文稿的制作方法。
  22. 一种计算机可读介质,其中存储了如权利要求21所述的计算机程序。
PCT/CN2017/094599 2016-12-26 2017-07-27 一种演示文稿的制作方法和装置 WO2018120820A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201611219547.6 2016-12-26
CN201611219547.6A CN108241597A (zh) 2016-12-26 2016-12-26 一种演示文稿的制作方法和装置

Publications (1)

Publication Number Publication Date
WO2018120820A1 true WO2018120820A1 (zh) 2018-07-05

Family

ID=62701920

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/094599 WO2018120820A1 (zh) 2016-12-26 2017-07-27 一种演示文稿的制作方法和装置

Country Status (2)

Country Link
CN (1) CN108241597A (zh)
WO (1) WO2018120820A1 (zh)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112533054A (zh) * 2019-09-19 2021-03-19 腾讯科技(深圳)有限公司 在线视频的播放方法、装置及存储介质

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110347848A (zh) * 2019-07-11 2019-10-18 深圳云智教育科技有限公司 一种演示文稿管理方法及装置
CN111221452B (zh) * 2020-02-14 2022-02-25 青岛希望鸟科技有限公司 方案讲解控制方法
CN113177126A (zh) * 2021-03-24 2021-07-27 珠海金山办公软件有限公司 一种处理演示文稿的方法、装置、计算机存储介质及终端

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101299250A (zh) * 2007-04-30 2008-11-05 深圳华飚科技有限公司 在线协同幻灯片制作服务***
CN101802816A (zh) * 2007-09-18 2010-08-11 微软公司 同步幻灯片显示事件与音频
CN105450944A (zh) * 2015-11-13 2016-03-30 北京自由坊科技有限责任公司 一种幻灯片和现场讲演语音同步录制与重现的方法及装置

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7312803B2 (en) * 2004-06-01 2007-12-25 X20 Media Inc. Method for producing graphics for overlay on a video source
CN101344883A (zh) * 2007-07-09 2009-01-14 宇瞻科技股份有限公司 记录演示文稿的方法
US20120317486A1 (en) * 2011-06-07 2012-12-13 Microsoft Corporation Embedded web viewer for presentation applications

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101299250A (zh) * 2007-04-30 2008-11-05 深圳华飚科技有限公司 在线协同幻灯片制作服务***
CN101802816A (zh) * 2007-09-18 2010-08-11 微软公司 同步幻灯片显示事件与音频
CN105450944A (zh) * 2015-11-13 2016-03-30 北京自由坊科技有限责任公司 一种幻灯片和现场讲演语音同步录制与重现的方法及装置

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112533054A (zh) * 2019-09-19 2021-03-19 腾讯科技(深圳)有限公司 在线视频的播放方法、装置及存储介质

Also Published As

Publication number Publication date
CN108241597A (zh) 2018-07-03

Similar Documents

Publication Publication Date Title
WO2018120819A1 (zh) 一种演示文稿的制作方法和装置
WO2018120821A1 (zh) 一种演示文稿的制作方法和装置
US10210769B2 (en) Method and system for reading fluency training
US9552807B2 (en) Method, apparatus and system for regenerating voice intonation in automatically dubbed videos
US8548618B1 (en) Systems and methods for creating narration audio
JP5030617B2 (ja) デジタル・オーディオ・プレーヤ上でrssコンテンツをレンダリングするためのrssコンテンツ管理のための方法、システム、およびプログラム(デジタル・オーディオ・プレーヤ上でrssコンテンツをレンダリングするためのrssコンテンツ管理)
CN108831437B (zh) 一种歌声生成方法、装置、终端和存储介质
WO2020098115A1 (zh) 字幕添加方法、装置、电子设备及计算机可读存储介质
US20200058288A1 (en) Timbre-selectable human voice playback system, playback method thereof and computer-readable recording medium
WO2016037440A1 (zh) 视频语音转换方法、装置和服务器
WO2018120820A1 (zh) 一种演示文稿的制作方法和装置
US20080027726A1 (en) Text to audio mapping, and animation of the text
US20130246063A1 (en) System and Methods for Providing Animated Video Content with a Spoken Language Segment
US20090006965A1 (en) Assisting A User In Editing A Motion Picture With Audio Recast Of A Legacy Web Page
US20110112835A1 (en) Comment recording apparatus, method, program, and storage medium
US20180226101A1 (en) Methods and systems for interactive multimedia creation
WO2012086356A1 (ja) ファイルフォーマット、サーバ、電子コミックのビューワ装置および電子コミック生成装置
JPH0778074A (ja) マルチメディアのスクリプト作成方法とその装置
KR101164379B1 (ko) 사용자 맞춤형 컨텐츠 제작이 가능한 학습 장치 및 이를 이용한 학습 방법
JP6746923B2 (ja) 情報処理システム、情報処理装置、情報処理方法および情報処理プログラム
Pauletto et al. Exploring expressivity and emotion with artificial voice and speech technologies
US9087512B2 (en) Speech synthesis method and apparatus for electronic system
US20080243510A1 (en) Overlapping screen reading of non-sequential text
KR102353797B1 (ko) 영상 컨텐츠에 대한 합성음 실시간 생성에 기반한 컨텐츠 편집 지원 방법 및 시스템
JP2008217447A (ja) コンテンツ生成装置及びコンテンツ生成プログラム

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17886990

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17886990

Country of ref document: EP

Kind code of ref document: A1