CN115514987A - System and method for automated narrative video production by using script annotations - Google Patents

System and method for automated narrative video production by using script annotations Download PDF

Info

Publication number
CN115514987A
CN115514987A CN202110697345.7A CN202110697345A CN115514987A CN 115514987 A CN115514987 A CN 115514987A CN 202110697345 A CN202110697345 A CN 202110697345A CN 115514987 A CN115514987 A CN 115514987A
Authority
CN
China
Prior art keywords
video
audio
script
stream
text
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110697345.7A
Other languages
Chinese (zh)
Inventor
周昌印
余飞
金伟成
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
See Technology Hangzhou Co ltd
Original Assignee
See Technology Hangzhou Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by See Technology Hangzhou Co ltd filed Critical See Technology Hangzhou Co ltd
Priority to CN202110697345.7A priority Critical patent/CN115514987A/en
Publication of CN115514987A publication Critical patent/CN115514987A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/23412Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs for generating or manipulating the scene composition of objects, e.g. MPEG-4 objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44012Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving rendering scenes according to scene graphs, e.g. MPEG-4 scene graphs

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

The present disclosure describes systems and methods for audio/visual authoring. An example system includes a text script parser configured to receive a text script and provide a first output and a second output, wherein the first output includes an expected spoken script, and wherein the second output includes a set of instructions corresponding to a predefined set of actions. The example system also includes an audio script alignment module configured to receive the audio stream and provide an alignment anchor. The example system yet further includes a video/audio rendering module configured to: receiving the instruction set, receiving the alignment anchor and providing rendered video/audio. The example system additionally includes a real-time video processor configured to: receiving a video stream, receiving the audio stream, receiving the rendered video/audio, and processing the video stream, audio stream, and the rendered video/audio in real-time to provide a video output.

Description

System and method for automated narrative video production by using script annotations
Background
Video media has become a popular format for expressing opinions and providing information, compared to conventional text media and still images. Major media companies, as well as individuals, create and distribute video media on a common video platform. However, conventional video production processes may involve multiple individuals from different specialties and may take a long time to complete. In general, conventional video creation may require the following steps: writing a text script of content contained in a video; (2) Collecting materials including shot fragments or prerecorded media according to the text script; (3) editing the video based on the script; and (4) rendering and auditing the video according to the script. Even a simple video-style oral presentation of opinions, the editing step can be time consuming and tedious. Such editing may include video overlay of video and still image material, media material time alignment, coarse/fine cropping, subtitle alignment, and the like. Another problem with conventional workflows is that the user is unaware of the final video results when recording a video clip. Thus, it may be difficult to consider the material added during the editing step. It usually takes a long time to process video editing. Accordingly, improved systems and methods for editing video media are desired.
Disclosure of Invention
The present disclosure describes systems and methods for audio/visual authoring.
In a first aspect, a system is described. The system includes a text script parser configured to receive a text script. The text script parser is configured to parse a text script to provide an expected spoken script and a set of instructions corresponding to a predefined set of actions. The system additionally includes an audio script alignment module configured to receive the audio stream and provide an alignment anchor. The alignment anchor indicates a dynamic progress of the audio stream relative to the expected spoken script. The system also includes a video/audio rendering module configured to: receiving an instruction set, receiving an alignment anchor, and providing rendered video/audio based on the instruction set and the alignment anchor. The system additionally includes a real-time video processor configured to: the method includes receiving a video stream, receiving an audio stream, receiving rendered video/audio, and processing the video stream, the audio stream, and the rendered video/audio in real-time to provide a video output.
In a second aspect, an automated real-time camera system is described. The automated real-time camera system includes a Graphical User Interface (GUI) configured to accept user input from a user and generate a text script based on the user input. The automated real-time camera system also includes a text script parser configured to receive a text script. The text script parser is configured to parse a text script to provide an expected spoken script and a set of instructions corresponding to a predefined set of actions. An automated real-time camera system comprising: a camera configured to capture a video stream; a microphone configured to capture an audio stream; and an audio script alignment module configured to receive the audio stream and provide an alignment anchor. The alignment anchor indicates a dynamic progress of the audio stream relative to the intended spoken script. The automated real-time camera system further includes a video/audio rendering module configured to: receiving an instruction set, receiving an alignment anchor, and providing rendered video/audio based on the instruction set and the alignment anchor. The automated real-time camera system includes a real-time video processor configured to: the method includes receiving a video stream, receiving an audio stream, receiving rendered video/audio, and processing the video stream, the audio stream, and the rendered video/audio in real-time to provide a video output.
In a third aspect, a method is described. The method includes parsing a text script to provide an expected spoken script and a set of instructions associated with a corresponding set of predefined actions; the method also includes receiving a video stream, receiving an audio stream, and determining an alignment anchor. The alignment anchor indicates a dynamic progress of the audio stream relative to the intended spoken script. The method also includes providing rendered video/audio based on the instruction set and the alignment anchor point and processing the video stream, the audio stream, and the rendered video/audio in real-time to provide a video output.
These and other embodiments, aspects, advantages, and alternatives will become apparent to one of ordinary skill in the art by reading the following detailed description, where appropriate, with reference to the accompanying drawings. Further, it is to be understood that this summary as well as the other descriptions and drawings provided herein are intended to illustrate embodiments by way of example only and that, as such, many variations are possible. For example, structural elements and process steps may be rearranged, combined, distributed, eliminated, or otherwise altered while remaining within the scope of the embodiments as claimed.
Drawings
FIG. 1 illustrates a system according to an example embodiment.
FIG. 2 illustrates an operational scenario in accordance with an example embodiment.
FIG. 3 illustrates a graphical user interface according to an example embodiment.
FIG. 4 illustrates an automated real-time camera system according to an example embodiment.
Fig. 5 illustrates a method according to an example embodiment.
Detailed Description
Example methods, devices, and systems are described herein. It should be understood that the words "example" and "exemplary" are used herein to mean "serving as an example, instance, or illustration. Any embodiment or feature described herein as "exemplary" or "exemplary" is not necessarily to be construed as preferred or advantageous over other embodiments or features. Other embodiments may be utilized, and other changes may be made, without departing from the scope of the subject matter presented herein.
Accordingly, the example embodiments described herein are not limiting. The aspects of the present disclosure may be arranged, substituted, combined, separated, and designed in a wide variety of different configurations, all of which are contemplated herein, as generally described herein and illustrated in the figures.
Further, features illustrated in each figure may be used in combination with each other, unless the context suggests otherwise. Thus, the figures should generally be considered as forming aspects of one or more unitary embodiments, with the understanding that not all illustrated features may be required of each embodiment.
I. Overview
In the present disclosure, systems and methods for narrative video production are disclosed. Based on the annotation script, the disclosed systems and methods can accumulate all necessary auto/video/text material and automatically render the output video. Audio input (e.g., reading a pre-written text script) whether captured with video or not may be used as a timeline. If there is no recorded audio input, automatically generated audio input may be generated via a text-to-speech process. Audio/text alignment is performed continuously and text annotations may trigger corresponding video effects or material overlays, etc. at specified times determined by the annotation location.
Example System
Fig. 1 illustrates a system 100 according to an example embodiment. The system 100 includes a text script parser 110 configured to receive a text script 10. The text script parser 110 is configured to parse the text script 10 to provide an intended spoken script 20 and a set of instructions 30 corresponding to a predefined set of actions 122.
The system 100 also includes an audio script alignment module 120 configured to receive the audio stream 40 and provide the alignment anchor 50. The alignment anchor 50 indicates the dynamic progress of the audio stream 40 relative to the intended spoken script 20.
The system 100 additionally includes a video/audio rendering module 130. In an example embodiment, video/audio rendering module 130 is configured to receive instruction set 30 and receive alignment anchor 50. Video/audio rendering module 130 is also configured to provide rendered video/audio 60 based on instruction set 30 and alignment anchor 50.
The system 100 additionally includes a real-time video processor 140. The real-time video processor 140 is configured to receive the video stream 70 and an audio stream corresponding thereto, such as video captured by a video camera having a microphone. In various embodiments, in addition to generating video/audio content, video/audio rendering module 130 may be configured to generate messages and/or signals. Such messages or signals may be utilized by the real-time video processor 140 to change its behavior and/or mode of operation. For example, messages and/or signals provided by the video/audio rendering module 130 may be utilized by the real-time video processor 140 to reduce the final output video volume, trigger special effects, trigger transitions between video or audio elements, or otherwise adjust the operating mode of the real-time video processor 140.
The real-time video processor 140 is further configured to receive the rendered video/audio 60 and process the video stream 70, the audio stream 40, and the rendered video/audio 60 in real-time to provide the video output 80, wherein the processing may include a mixing operation. In some example embodiments, the real-time video processor 140 may perform various audio/video/effect processing tasks. In such a scenario, the real-time video processor 140 may perform functions such as adjusting audio volume, applying special filters/effects to the video stream, and several possibilities.
In various embodiments, the video/audio rendering module 130 may include a computer vision module 138. In one embodiment, the computer vision module 138 may be used to extract certain information (e.g., metadata) from the video stream 70 to aid or guide the dynamic rendering performed by the video/audio rendering module 130 and/or the video processing performed by the real-time video processor 140.
In some embodiments, the video/audio rendering module 130 may include specific events 132 and trigger markers 134. The specific events 132 and trigger tags 134 may be obtained from the text script 10 by the text script parser 110 and sent to the video/audio rendering module 130.
In some embodiments, video/audio rendering module 130 may further include media assets 136, wherein providing rendered video/audio is further based on the set of media assets 136. In such a scenario, the media assets 136 may include at least one of: video content, audio content, still images, text content, three-dimensional still object content, or three-dimensional animated object content.
In some examples, video/audio rendering module 130 may be configured to overlay some or all of media assets 136 onto the original video (e.g., video stream 70) at particular temporal and spatial locations based at least in part on real-time object detection results. For example, if the computer vision module 138 determines that the video stream 70 includes a human hand, the video/audio rendering module 130 may be operable to virtually attach an apple to the detected human hand. Additionally or alternatively, the video/audio rendering module 130 may be operable to replace the background of the video stream 70 with a particular image defined in the instruction set 30. Still further, the rendered video/audio 60 provided by the video/audio rendering module 130 may be further based on contextual information (e.g., objects in the video, characters in the video, etc.) extracted from the video stream 70.
In some examples, the text script 10 includes information indicating the intended spoken script 20, the particular event 132 to be triggered, and the trigger flag 134 for the particular event 132, all of which may be parsed by the text script resolver. In such a scenario, the trigger markers 134 for a particular event 132 may include spoken text prompts that correspond to desired event trigger points within the intended spoken script 20.
In various examples, the audio stream 40 includes a spoken version of the text script 10. Alternatively or additionally, the audio stream 40 may include recorded audio.
A further example of the system 100 includes a text-to-speech module 112 configured to automatically generate the audio stream 40 based on the text script 10.
The system 100 may additionally include a playback monitor 170 configured to display the video output 80 in real-time. In some embodiments, the playback monitor 170 is further configured to display at least a portion of the intended spoken script 20.
Additionally or alternatively, the system 100 may include a controller 150 having at least one processor 152 and a memory 154. In this scenario, at least one processor 152 executes instructions stored in memory 154 in order to implement the instructions. The instructions may include some or all of the method 600 as described and illustrated with respect to fig. 5. In some embodiments, controller 150 may be configured to implement some or all of the blocks of method 600 as described and illustrated with respect to fig. 5.
In some embodiments, the controller 150 may further include a trained artificial intelligence model 156.
An advantageous feature of the described system and method is that the video capture/recording and video editing processes are merged together based on video content scripts.
FIG. 2 illustrates an operational scenario 200 according to an example embodiment. As illustrated in operational scenario 200, inputs to the system may include at least one or more of:
1) A text script for audio input. The script will be read later using audio input acquired by the system. The text script will also include the necessary annotations.
2) An audio stream, such as a spoken version of a text script.
3) Optionally a video stream input, time synchronized with the audio stream.
4) All necessary assets to be used for outputting video, including: video content, audio content, still images, text content.
5) 3D static objects, with or without animation.
6) Other essential components.
In some embodiments, four main components may be included in the system 100:
1) A text script parser 110 that parses a text script input from a specific format into two outputs. The first output is the intended spoken script 20 corresponding to the audio stream 40. The second output is an instruction set 30 of predefined actions that can be triggered during real-time video production.
2) An audio script alignment module 120 that will accept the audio stream 40 and dynamically generate alignment anchors 50. The alignment anchor 50 indicates the audio progress or placeholder of the intended spoken script.
3) A video/audio rendering module 130 that hosts a structured script with predefined actions and alignment anchors. When the audio progress reaches a certain point, the video/audio rendering module 130 renders the video or audio as output, as specified in a structured script with predefined actions.
4) A real-time video processor 140 that accepts the audio stream 40, the video stream 70, and the rendered video/audio 60 through the video/audio rendering module 130 and generates the final video output 80. The video output 80 of the real-time video processor may be the final production video.
A. Script format and parsing
The input script conveys the following information:
1) An expected spoken script;
2) A specific event to be triggered; and
3) Number of triggers for a particular event.
The first two pieces of information are relatively straight white. However, establishing the number of triggers for a particular event requires careful design and planning. In the present disclosure, a timeline is associated with an audio stream. Thus, from scratch, the time stamp cannot be directly applied to the particular event to be triggered because the system does not have this information due to the variability of the audio stream. However, based on information from the spoken text/audio stream alignment module, the present system and method can accurately identify timestamps for aligned text. Thus, the event trigger may be annotated into the intended spoken text. In such a scenario, the present systems and methods are configured to trigger those events at the expected times.
An example of a script format is as follows:
this is an example script [ FIG. 1 is shown within 2 seconds ]. An event will be triggered when a different location of the script is read [ zoom in to the upper corner of FIG. 1]. You can also [ show the avatar ] control the virtual avatar and let it speak some sentences. [ head portrait: "yes, i can speak a certain sentence at a certain point. "]. Thank you for precious time!
Other script formats and syntax are possible as long as the above requirements are met. The analysis of the text outputs two main components (1) an expected spoken script; and (2) actions/events to trigger.
Some possible annotations may include:
1) Trigger notes: we show fast animation \ here [: action \ j ].
2) And displaying the annotation: this is \ beautiful flower. \8230, action \ except flowers, has a big tree.
3) Duration annotation: we show here a 5 second fast animation \ action 5.0\ for.
4) Start and end notes: this is \ in start: id =1: action \ beautiful flower. \8230 [ < end > < 1\ except for flowers, there is a big tree.
5) Head portrait action: [ head portrait: "yes, i can speak a certain sentence at a certain point. "]
Various different actions/events may be triggered in association with the present disclosure. For example, various elements may be played or displayed, including textual content, still images, video content, 3D models, transitions, and/or other special effects. Additionally or alternatively, other actions or events may include: changing the background of the video when transitioning to another scene; adjusting various audio/video parameters, such as audio volume, frame layout; turning on or off certain gestural triggers, such as gestural control of virtual objects; start a real-time video call and/or display video from another camera. Other actions or events are possible within the scope of the present disclosure.
B. Timeline and Material alignment
Text-to-speech alignment is an important element of the video production process described herein. Various materials (including video/audio/images) may be dynamically rendered onto the shared timeline based on the audio script. The shared timeline is based on the audio input and its alignment with the text script. The audio input may be provided in a variety of ways, such as:
1) Provided as a streaming input from a recording camera and corresponding audio receiver;
2) Is provided from a recorded piece of audio; or
3) Audio generated from text using text-to-speech (TTS) technology is provided.
To further control the time at which material appears in the rendered video, further conditions may be added to trigger certain important events. For example, during a live broadcast, triggering a video or audio element within a rendered video may depend on 1) aligning the script and audio input up to a certain point; and 2) capturing body gestures in the video, such as a hand sliding from left to right, hand signals, or head tilt, etc. For example, the trigger states for such a scenario may include: when the audio stream matches the script text at some point, the script will not only directly trigger an event, but may also trigger a new state, where some body gesture detection will be the actual event trigger. After the audio and text align passes a certain text script point or action timeout duration, the system may exit this state so that body gesture detection will no longer trigger an action.
C. Dynamic rendering
Within the scope of the present disclosure, the final video may be dynamically rendered by templates and/or artificial intelligence based models to determine various rendering parameters. In such a scenario, the following information from different types of sources may be required when outputting the final video.
For a video source, the video/audio rendering module may determine and render with information about: 1) Playback speed, duration, start timestamp; 2) A location; 3) Scale; and/or 4) other required attributes. In some examples, a video/audio rendering module (e.g., video/audio rendering module 130) may accept three inputs: 1) A set of instructions 30 from a text script parser 110; 2) An alignment anchor 50; and 3) a video stream 70. Further, the video/audio rendering module may include a computer vision module 138 configured to process the video stream 70 and provide contextual instructions/metadata to perform dynamic rendering. Such dynamic rendering may include: 1) Replacing the background; 2) Cartoonizing or other graphic styles; 3) Special filtration, etc. Such functionality may require information from the computer vision module 138 to properly locate and/or trigger such effects/modes.
For audio sources, the video/audio rendering module may determine and render with information such as: 1) Audio volume; 2) Audio playback speed, duration, start timestamp; and/or 3) other required attributes.
In the case of 3D avatar and/or 3D animation, there may be several parameters to utilize. For example, the video/audio rendering module needs information about whether the 3D avatar should smile when "speaking" a sentence and/or whether the head and body should move when the avatar speaks. Such information may not be provided in its entirety in the text script. That is, the provided text script may provide incomplete information. In such a scenario, default values, templates, and/or AI-determined values may be utilized. For example, some examples may utilize Natural Language Processing (NLP) techniques to determine the mood of the script, which may be used to adjust or otherwise affect the movement and/or facial expression of the avatar.
Messages and signals in dynamic rendering
In some example embodiments, a triggering event may be included in the signal generated by the video/audio rendering module 130 that may include rendering instructions that may assist the functionality of the real-time video processor. In such a scenario, there may be several video streams input into the real-time video processor, and the triggering event may be a signal comprising:
1) Switching the video source to a different camera;
2) Speeding up/slowing down video playback speed;
3) Reducing the volume of the audio;
4) Applying a filter to the video;
5) Applying a scene transition between two video scenes; and
6) Video zoom in/out, video cropping.
D. Automatic real-time video production system
Real-time production
Using the presently described systems and methods, an automated real-time camera system may be provided. In such a scenario, an example system may include three main components: (1) a Graphical User Interface (GUI) for editing scripts; (2) the video production system introduced above; and (3) a camera, a microphone, and a playback monitor configured to display the output video.
Fig. 3 illustrates a graphical user interface 300 according to an example embodiment. The GUI 300 may be configured to accept user input from a user and generate and edit text scripts (e.g., text script 10) based on the user input, such as adding actions to the text scripts and specifically configuring the actions.
Fig. 4 illustrates an automated real-time camera system 400 according to an example embodiment.
The example automated real-time camera system 400 may include a Graphical User Interface (GUI), such as the GUI 300 shown in fig. 3.
The automated real-time camera system 400 also includes a text script parser (e.g., text script parser 110) configured to receive a text script. The text script parser is also configured to parse the text script to provide a desired spoken script (e.g., desired spoken script 20) and a set of instructions (e.g., set of instructions 30) corresponding to a predefined set of actions (e.g., predefined set of actions 122).
The automated real-time camera system 400 includes a camera configured to capture a video stream (e.g., video stream 70) and a microphone configured to capture an audio stream (e.g., audio stream 40).
The automated real-time camera system 400 additionally includes an audio script alignment module (e.g., audio script alignment module 120) configured to receive the audio stream and provide an alignment anchor (e.g., alignment anchor 50). The alignment anchor indicates a dynamic progress of the audio stream relative to the intended spoken script.
The automated real-time camera system 400 includes a video/audio rendering module (e.g., video/audio rendering module 130). The video/audio rendering module is configured to receive an instruction set and receive an alignment anchor. The video/audio rendering module is further configured to provide rendered video/audio (e.g., rendered video/audio 60) based on the instruction set and the alignment anchor.
The automated real-time camera system 400 includes a real-time video processor (e.g., the real-time video processor 140) configured to: the video stream, audio stream, rendered video/audio are received, and the video stream, audio stream, and rendered video/audio are processed in real-time to provide a video output (e.g., video output 80).
The automated real-time camera system 400 may also include a playback monitor (e.g., playback monitor 170) configured to display video output in real-time. In some examples, the playback monitor is further configured to display a portion of the expected spoken script.
In order to operate such a system, a user needs to create a text script with the necessary annotations. In some example embodiments, a Graphical User Interface (GUI) may provide a way for a user to enter desired spoken scripts and insert necessary annotations with corresponding metadata (e.g., action type, 3D avatar, timing, trigger conditions, etc.). After the text script is generated, the user may enter a recording phase. During the recording phase, audio and/or video content may be captured with a camera and/or microphone. In such a scenario, the monitor may show the final video production results in real-time. Spoken scripts may be displayed on a monitor to prompt/remind the user what to expect next and/or say. After the recording phase is complete, the user can review the final produced video (e.g., via a monitor) and the process is complete without further editing steps.
Edit extension/post-processing
Although the present system and method may provide real-time video production capabilities, some editing functions may be performed in post-processing. For example, the systems and methods may include adding/removing some audio/video assets without re-recording the entire video again. In such a scenario, post-production following initial content capture may be performed after video production is ready. Our system also easily meets this requirement. After the first recording is complete, we have collected accurate timestamps for all assets that were dynamically rendered. For the second rendering after reconfiguring the script annotations, no new recording from the user is needed as long as the overall structure remains similar. The user only needs to run the rendering process again using the pre-recorded audio and video. Furthermore, since the system has recorded all timestamps, the system can export even those configurations into items compatible with popular video editing tools (such as Adobe PR, etc.). Where the user can make all possible modifications.
Example method
Fig. 5 illustrates a method 500 according to an example embodiment. It will be appreciated that the method 500 may include fewer or more steps or blocks than are explicitly illustrated or otherwise disclosed herein. Further, the respective steps or blocks of the method 500 may be performed in any order and each step or block may be performed one or more times. In some embodiments, some or all of the blocks or steps of the method 500 may be implemented by the controller 150 and/or other elements of the system 100, as illustrated and described with respect to fig. 1, 2, 3, and 4.
Block 502 includes parsing a text script (e.g., text script 10) to provide a desired spoken script (e.g., desired spoken script 20) and a set of instructions (e.g., set of instructions 30) associated with a corresponding set of predefined actions (e.g., set of predefined actions 122).
Block 504 includes receiving an audio stream (e.g., audio stream 40).
Block 506 includes determining an alignment anchor (e.g., alignment anchor 50). In such a scenario, the alignment anchor indicates a dynamic progress of the audio stream with respect to the intended spoken script.
Block 508 includes rendering video/audio (e.g., rendered video/audio 60) based on the instruction set and the alignment anchor.
Block 510 includes processing the video stream (e.g., video stream 70), the audio stream, and the rendered video/audio in real-time to provide a video output (e.g., video output 80).
In some example embodiments, the text script may include information indicating an expected spoken script, a particular event to trigger (e.g., particular event 132), and a trigger for the particular event (e.g., trigger 134).
In some examples, processing the video stream, the audio stream, and the rendered video/audio is based on spoken text prompts within an intended spoken script.
In some embodiments, method 500 may include displaying the video output in real-time by a playback monitor (e.g., playback monitor 170). In some examples, method 500 may include displaying, by a playback monitor, at least a portion of an expected spoken script.
In various examples, at least one of the determining (block 506), rendering (block 508), or processing (block 510) steps is performed based at least in part on the trained artificial intelligence model 156. In some examples, the trained artificial intelligence model 156 may include a Natural Language Processing (NLP) model.
The particular arrangements shown in the drawings should not be considered limiting. It should be understood that other embodiments may include more or less of each element shown in a given figure. Further, some of the illustrated elements may be combined or omitted. Still further, illustrative embodiments may include elements not illustrated in the figures.
The steps or blocks representing the processing of information may correspond to circuitry that may be configured to perform particular logical functions of the methods or techniques described herein. Alternatively or in addition, the steps or blocks representing the processing of information may correspond to modules, segments, or portions of program code (including related data). The program code may include one or more instructions executable by a processor for implementing specific logical functions or actions in the methods or techniques. The program code and/or related data may be stored on any type of computer-readable medium, such as a storage device, including a diskette, hard drive, or other storage medium.
The computer readable medium may also include non-transitory computer readable media such as computer readable media like register memory, processor cache memory, and Random Access Memory (RAM) that store data for short periods of time. The computer-readable medium may also include a non-transitory computer-readable medium that stores program code and/or data for longer periods of time. Thus, for example, a computer-readable medium may include secondary or permanent long term storage devices such as Read Only Memory (ROM), optical or magnetic disks, compact disk read only memory (CD-ROM). The computer readable medium can also be any other volatile or non-volatile storage system. For example, the computer readable medium may be considered a computer readable storage medium or a tangible storage device.
While various examples and embodiments have been disclosed, other examples and embodiments will be apparent to those skilled in the art. The various disclosed examples and embodiments are for purposes of illustration and are not intended to be limiting, with the true scope being indicated by the following claims.

Claims (20)

1. A system for providing video output, comprising:
a text script parser configured to receive a text script, wherein the text script parser is configured to parse the text script to provide an expected spoken script and a set of instructions corresponding to a predefined set of actions;
an audio script alignment module configured to receive an audio stream and provide an alignment anchor, wherein the alignment anchor indicates a dynamic progress of the audio stream relative to the expected spoken script;
a video/audio rendering module configured to:
receiving the set of instructions;
receiving the alignment anchor point; and
providing rendered video/audio based on the instruction set and the alignment anchor; and a real-time video processor configured to:
receiving a video stream;
receiving the audio stream;
receiving the rendered video/audio; and
processing the video stream, audio stream, and the rendered video/audio in real-time to provide a video output.
2. The system of claim 1, wherein the text script comprises information indicating the expected spoken script, a particular event to trigger, and a trigger flag for the particular event.
3. The system of claim 2, wherein the trigger marker for the particular event comprises a spoken text prompt corresponding to a desired event trigger point within the expected spoken script.
4. The system of claim 1, wherein the audio stream comprises a spoken version of the text script.
5. The system of claim 1, wherein the audio stream comprises recorded audio.
6. The system of claim 1, further comprising a text-to-speech module configured to automatically generate the audio stream based on the text script.
7. The system of claim 1, wherein the video/audio rendering module further comprises a media asset, wherein providing the rendered video/audio is further based on a set of media assets, wherein the media assets comprise at least one of: video content, audio content, still images, text content, three-dimensional still object content, or three-dimensional animated object content.
8. The system of claim 1, wherein the video/audio rendering module is further configured to receive the video stream, wherein the rendered video/audio provided by the video/audio rendering module is further based on contextual information extracted from the video stream.
9. The system of claim 1, further comprising a playback monitor configured to display the video output in real-time.
10. The system of claim 9, wherein the playback monitor is further configured to display at least a portion of the intended spoken script.
11. An automated real-time camera system, comprising:
a Graphical User Interface (GUI) configured to accept user input from a user and generate a text script based on the user input;
a text script parser configured to receive a text script, wherein the text script parser is configured to parse the text script to provide an expected spoken script and a set of instructions corresponding to a predefined set of actions;
a camera configured to capture a video stream;
a microphone configured to capture an audio stream;
an audio script alignment module configured to receive the audio stream and provide an alignment anchor, wherein the alignment anchor indicates a dynamic progress of the audio stream relative to the expected spoken script;
a video/audio rendering module configured to:
receiving the set of instructions;
receiving the alignment anchor point; and
providing rendered video/audio based on the instruction set and the alignment anchor; and a real-time video processor configured to:
receiving the video stream;
receiving the audio stream;
receiving the rendered video/audio; and
processing the video stream, audio stream, and the rendered video/audio in real-time to provide a video output.
12. The automated real-time camera system of claim 11, further comprising a playback monitor configured to display the video output in real-time.
13. The automated real-time camera system of claim 12, wherein the playback monitor is further configured to display a portion of the intended spoken script.
14. A method for providing video output, comprising:
parsing the text script to provide an expected spoken script and a set of instructions related to a corresponding set of predefined actions;
receiving a video stream;
receiving an audio stream;
determining an alignment anchor, wherein the alignment anchor indicates a dynamic progress of the audio stream relative to the expected spoken script;
rendering video/audio based on the instruction set and the alignment anchor; and
processing the video stream, the audio stream and the rendered video/audio in real-time to provide a video output.
15. The method of claim 14, wherein the text script comprises information indicating the expected spoken script, a particular event to trigger, and a trigger flag for the particular event.
16. The method of claim 14, wherein processing the video stream, the audio stream, and the rendered video/audio is based on spoken text prompts within the intended spoken script.
17. The method of claim 14, further comprising:
displaying the video output in real-time through a playback monitor.
18. The method of claim 14, further comprising:
displaying, by a playback monitor, at least a portion of the intended spoken script.
19. The method of claim 14, wherein at least one of the determining, rendering, or processing steps is performed based at least in part on a trained artificial intelligence model.
20. The method of claim 19, wherein the trained artificial intelligence model comprises a Natural Language Processing (NLP) model.
CN202110697345.7A 2021-06-23 2021-06-23 System and method for automated narrative video production by using script annotations Pending CN115514987A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110697345.7A CN115514987A (en) 2021-06-23 2021-06-23 System and method for automated narrative video production by using script annotations

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110697345.7A CN115514987A (en) 2021-06-23 2021-06-23 System and method for automated narrative video production by using script annotations

Publications (1)

Publication Number Publication Date
CN115514987A true CN115514987A (en) 2022-12-23

Family

ID=84499102

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110697345.7A Pending CN115514987A (en) 2021-06-23 2021-06-23 System and method for automated narrative video production by using script annotations

Country Status (1)

Country Link
CN (1) CN115514987A (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130124203A1 (en) * 2010-04-12 2013-05-16 II Jerry R. Scoggins Aligning Scripts To Dialogues For Unmatched Portions Based On Matched Portions
CN107172476A (en) * 2017-06-09 2017-09-15 创视未来科技(深圳)有限公司 A kind of system and implementation method of interactive script recorded video resume
US20180063602A1 (en) * 2014-05-22 2018-03-01 Idomoo Ltd. System and method to generate a video on-the-fly
US20180213299A1 (en) * 2017-01-26 2018-07-26 Electronics And Telecommunications Research Institute Apparatus and method for tracking temporal variation of video content context using dynamically generated metadata
US20190087870A1 (en) * 2017-09-15 2019-03-21 Oneva, Inc. Personal video commercial studio system
CN110166816A (en) * 2019-05-29 2019-08-23 上海乂学教育科技有限公司 The video editing method and system based on speech recognition for artificial intelligence education
CN110622109A (en) * 2017-04-11 2019-12-27 朗德费尔股份有限公司 Computer animation based on natural language
CN112004163A (en) * 2020-08-31 2020-11-27 北京市商汤科技开发有限公司 Video generation method and device, electronic equipment and storage medium

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130124203A1 (en) * 2010-04-12 2013-05-16 II Jerry R. Scoggins Aligning Scripts To Dialogues For Unmatched Portions Based On Matched Portions
US20180063602A1 (en) * 2014-05-22 2018-03-01 Idomoo Ltd. System and method to generate a video on-the-fly
US20180213299A1 (en) * 2017-01-26 2018-07-26 Electronics And Telecommunications Research Institute Apparatus and method for tracking temporal variation of video content context using dynamically generated metadata
CN110622109A (en) * 2017-04-11 2019-12-27 朗德费尔股份有限公司 Computer animation based on natural language
CN107172476A (en) * 2017-06-09 2017-09-15 创视未来科技(深圳)有限公司 A kind of system and implementation method of interactive script recorded video resume
US20190087870A1 (en) * 2017-09-15 2019-03-21 Oneva, Inc. Personal video commercial studio system
CN110166816A (en) * 2019-05-29 2019-08-23 上海乂学教育科技有限公司 The video editing method and system based on speech recognition for artificial intelligence education
CN112004163A (en) * 2020-08-31 2020-11-27 北京市商汤科技开发有限公司 Video generation method and device, electronic equipment and storage medium

Similar Documents

Publication Publication Date Title
US8966360B2 (en) Transcript editor
CN110858408B (en) Animation production system
US8223152B2 (en) Apparatus and method of authoring animation through storyboard
JP4250301B2 (en) Method and system for editing video sequences
JP4599244B2 (en) Apparatus and method for creating subtitles from moving image data, program, and storage medium
CA2956566C (en) Custom video content
US20180226101A1 (en) Methods and systems for interactive multimedia creation
US20100183278A1 (en) Capturing and inserting closed captioning data in digital video
KR20070020252A (en) Method of and system for modifying messages
US20120026174A1 (en) Method and Apparatus for Character Animation
JP2001333379A (en) Device and method for generating audio-video signal
JP2010509859A (en) System and method for high speed subtitle creation
US9666211B2 (en) Information processing apparatus, information processing method, display control apparatus, and display control method
KR20140141408A (en) Method of creating story book using video and subtitle information
US8223153B2 (en) Apparatus and method of authoring animation through storyboard
JP6641045B1 (en) Content generation system and content generation method
CN114911448A (en) Data processing method, device, equipment and medium
CN115514987A (en) System and method for automated narrative video production by using script annotations
US20240244299A1 (en) Content providing method and apparatus, and content playback method
Türk The technical processing in smartkom data collection: a case study
JP7133367B2 (en) MOVIE EDITING DEVICE, MOVIE EDITING METHOD, AND MOVIE EDITING PROGRAM
JP4235635B2 (en) Data retrieval apparatus and control method thereof
CN113033357A (en) Subtitle adjusting method and device based on mouth shape features
Park et al. Automatic subtitles localization through speaker identification in multimedia system
US20230420001A1 (en) Enhanced trim assistance for video editing environments

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination