CN113709575B - Video editing processing method and device, electronic equipment and storage medium - Google Patents

Video editing processing method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN113709575B
CN113709575B CN202110371612.1A CN202110371612A CN113709575B CN 113709575 B CN113709575 B CN 113709575B CN 202110371612 A CN202110371612 A CN 202110371612A CN 113709575 B CN113709575 B CN 113709575B
Authority
CN
China
Prior art keywords
shot
video
editing
picture
lens
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110371612.1A
Other languages
Chinese (zh)
Other versions
CN113709575A (en
Inventor
韩瑞
王丽云
沈艳慧
张仁寿
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Technology Shenzhen Co Ltd
Original Assignee
Tencent Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Technology Shenzhen Co Ltd filed Critical Tencent Technology Shenzhen Co Ltd
Priority to CN202110371612.1A priority Critical patent/CN113709575B/en
Publication of CN113709575A publication Critical patent/CN113709575A/en
Application granted granted Critical
Publication of CN113709575B publication Critical patent/CN113709575B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47205End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for manipulating displayed content, e.g. interacting with MPEG-4 objects, editing locally
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/166Editing, e.g. inserting or deleting
    • G06F40/186Templates
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/451Execution arrangements for user interfaces
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/435Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
    • H04N21/4355Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream involving reformatting operations of additional data, e.g. HTML pages on a television screen
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • H04N21/4398Processing of audio elementary streams involving reformatting operations of audio signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44008Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • H04N21/440245Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display the reformatting operation being performed only on part of the stream, e.g. a region of the image or a time segment
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • Databases & Information Systems (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

The application provides a video editing processing method, a video editing processing device, electronic equipment and a computer readable storage medium; to computer vision techniques in the field of artificial intelligence, the method comprising: displaying a video editing script template in a document editing interface; displaying the set plurality of shot pictures and parameters corresponding to each of the shot pictures in the video editing script template in response to a video editing operation; in response to a video preview operation received during video editing, generating a preview video based on at least one shot and parameters corresponding to the at least one shot, and displaying the preview video. According to the method and the device, the corresponding preview video can be checked in time in the script editing process, and therefore video editing efficiency is improved.

Description

Video editing processing method and device, electronic equipment and storage medium
Technical Field
The present disclosure relates to artificial intelligence technology and internet technology, and in particular, to a video editing processing method, apparatus, electronic device, and computer readable storage medium.
Background
Artificial intelligence (AI, artificial Intelligence) is the theory, method and technique and application system that uses a digital computer or a machine controlled by a digital computer to simulate, extend and extend human intelligence, sense the environment, acquire knowledge and use the knowledge to obtain optimal results. As artificial intelligence technology research and advances, artificial intelligence technology has been developed and applied in a variety of fields.
Taking video editing as an example, a shot script (Storyboard) is a series of sketches prepared by a video planner for formally producing video, and is a visual expectation before video formation. In the related art, a video planner can formulate a static shot script according to a video theme, and the static shot script cannot acquire the rhythm of a video and cannot accurately estimate the transition and stay time, so that parameters in the shot script need to be continuously optimized in the video editing process, and the video editing efficiency is low. For this, there is no effective solution in the related art.
Disclosure of Invention
The embodiment of the application provides a video editing processing method, a video editing processing device, electronic equipment and a computer readable storage medium, which can support timely viewing of corresponding preview videos in a script editing process, so that video editing efficiency is improved.
The editing efficiency of the video is improved by generating the preview video corresponding to the shot picture and the parameters.
The technical scheme of the embodiment of the application is realized as follows:
the embodiment of the application provides a video editing processing method, which comprises the following steps:
displaying a video editing script template in a document editing interface;
displaying the set plurality of shot pictures and parameters corresponding to each of the shot pictures in the video editing script template in response to a video editing operation;
Generating a preview video based on at least one shot picture and parameters corresponding to the at least one shot picture in response to a video preview operation received during video editing, and
and displaying the preview video.
In the above scheme, the performing parameter identification processing on each lens frame to obtain parameters adapted to the plurality of lens frames includes:
performing one of the following processes for each of the lens pictures:
performing object identification processing on the lens picture, and determining a time length matched with the lens picture according to the number of identified objects, wherein the number of identified objects is positively correlated with the time length;
and identifying the similarity between the shot picture and the adjacent shot picture, and determining the time length matched with the shot picture according to the similarity, wherein the similarity is in negative correlation with the time length.
In the above scheme, the performing parameter identification processing on each lens frame to obtain parameters adapted to the plurality of lens frames includes:
the following processing is performed for each of the lens pictures:
performing object identification processing on the lens picture, and inquiring sound effects with mapping relations between the identified objects in a mapping table, wherein the mapping table comprises a plurality of objects and a plurality of sound effects which are in one-to-one correspondence with the plurality of objects;
And determining the sound effect with the mapping relation with the identified object as the sound effect matched with the lens picture.
In the above scheme, the performing parameter identification processing on each lens frame to obtain parameters adapted to the plurality of lens frames includes:
the following processing is performed for each of the lens pictures:
identifying the similarity between the lens picture and the adjacent lens picture, and inquiring transition matched with the similarity;
and determining the transition matched with the similarity as the transition matched with the lens picture.
In the above scheme, the performing parameter identification processing on each lens frame to obtain parameters adapted to the plurality of lens frames includes:
the following processing is performed for each of the lens pictures:
identifying a history parameter adapted to the lens picture as a parameter adapted to the lens picture;
wherein the type of the history parameter includes one of:
history parameters corresponding to the history lens picture with highest similarity between the lens pictures;
setting a history parameter with highest frequency in the video editing process;
historical parameters set in the video editing process which are most closely spaced from the current time point.
In the above scheme, the method further comprises:
responding to a collaborative editing triggering operation received in the document editing interface, and displaying a collaborative account setting page of a first account;
the first account is an account logged in the document editing interface, and the collaborative account setting page comprises at least one candidate account;
and in response to an account selection operation received on the collaborative account setting page, determining at least one selected candidate account as a second account for collaborating with the first account to edit the video editing script template, and sending the video editing script template to the second account.
In the above solution, before the sending the video editing script template to the second account, the method further includes:
displaying an editing authority setting inlet on the collaborative account setting page;
obtaining the set rights in response to a rights setting operation for the editing rights setting portal, wherein the types of rights include: viewing rights, editing rights;
and determining that the process of sending a video editing script template applying the rights to the second account is to be executed.
In the above aspect, when the preview video is displayed, the method further includes:
displaying a video modification page in response to a modification operation for the preview video, wherein the video modification page comprises at least one shot picture and parameters corresponding to each of the shot pictures;
and responding to the parameter modification operation received at the video modification page, and updating the displayed preview video according to the modified parameter.
In the above scheme, the displaying the video editing script template in the document editing interface includes:
responding to video editing triggering operation, and displaying a video type selection page, wherein the video type selection page comprises a plurality of candidate video types;
and responding to the video type selection operation received in the video type selection page, and displaying a video editing script template corresponding to the selected video type.
The embodiment of the application provides a video editing processing device, which comprises:
the display module is used for displaying the video editing script template in the document editing interface;
an editing module for displaying a plurality of set shot pictures and parameters corresponding to each of the shot pictures in the video editing script template in response to a video editing operation;
A generation module for generating a preview video based on at least one shot picture and parameters corresponding to the at least one shot picture in response to a video preview operation received in a video editing process;
and the display module is also used for displaying the preview video.
In the above scheme, the parameters include a clipping parameter and a linking parameter; the generating module is further configured to perform the following processing for each of the lens frames: according to the clipping parameters of the shot pictures, clipping the shot pictures to obtain shot fragments corresponding to the shot pictures; when the number of the at least one shot picture is one, determining a shot segment corresponding to the shot picture as the preview video; and when the number of the at least one lens picture is a plurality of, combining the lens segments corresponding to each lens picture according to the connection parameter of each lens picture to obtain the preview video.
In the above aspect, the clipping parameters include at least one of: a document, a duration and an audio effect; the generation module is further configured to clip the shot frame into a preprocessing segment, where a playing duration of the preprocessing segment is a duration in a clipping parameter of the shot frame; and adding the text in the clipping parameters of the shot picture into the preprocessing segment, and adding the sound effect in the clipping parameters into the preprocessing segment to obtain the shot segment corresponding to the shot picture.
In the above scheme, the engagement parameters include at least one of the following: lens sequence number, transition; the generating module is further configured to sort the plurality of lens segments according to the sequence of the lens sequence numbers in the linking parameters, so as to obtain a lens segment sequence; the following connection processing is sequentially performed for each of the shot segments in the shot segment sequence: connecting the lens segment with the adjacent lens segment according to transition in the connection parameters corresponding to the lens segment; and taking the shot fragment sequence after the connection processing as the preview video.
In the above scheme, the editing module is further configured to respond to a video editing operation submitted in the video editing script template by at least one of the first account number and the second account number, and display, in the video editing script template, a plurality of shot frames set by the video editing operation, and parameters corresponding to each of the shot frames; the first account is an account logged in the document editing interface, and the second account is an account cooperated with the first account to edit the video editing script template.
In the above scheme, the editing module is further configured to query a status of a first shot frame, where the first shot frame is a shot frame that the video editing operation requests to edit in the video editing script template, or a shot frame that the video editing operation requests to edit in the video editing script template corresponds to a parameter; determining that a process responsive to the video editing operation is to be performed when the first shot and a second shot related to the first shot are both in an unedited state; and when the first lens picture and the second lens picture related to the first lens picture are in an editing state, displaying first prompt information, wherein the first prompt information is used for prompting that the video editing operation cannot be responded immediately due to editing conflict.
In the above scheme, the first lens picture and the second lens picture meet at least one of the following association conditions: the first lens picture and the second lens picture are edited into the video editing script template by the same account number; the first shot picture and the second shot picture belong to the same scene.
In the above scheme, the editing module is further configured to query a state of a first parameter, where the first parameter is a parameter corresponding to a third shot picture edited in the video editing script template by the video editing operation request; determining that processing responsive to the video editing operation is to be performed when the first parameter and a second parameter related to the first parameter are both in an unedited state; and when the first parameter and the second parameter related to the first parameter are both in an editing state, displaying second prompt information, wherein the second prompt information is used for prompting that the video editing operation cannot be responded immediately due to editing conflict.
In the above scheme, the first parameter and the second parameter meet at least one of the following association conditions: the first parameter and the second parameter are edited into the video editing script template by the same account; the lens picture corresponding to the first parameter is the same as the lens picture corresponding to the second parameter; the first parameter is of the same type as the second parameter.
In the above scheme, the video editing operation includes a shot screen editing operation and a parameter editing operation; the editing module is further used for responding to the shot picture editing operation and displaying a plurality of shot pictures set by the shot picture editing operation in the video editing script template; and responding to the parameter editing operation, and displaying parameters set by the parameter editing operation and corresponding to the plurality of shot pictures in the video editing script template.
In the above scheme, the editing module is further configured to perform parameter identification processing on each of the lens frames to obtain parameters adapted to the plurality of lens frames, and display the parameters adapted to the plurality of lens frames as default parameters in the video editing script template; in response to the parameter editing operation for the plurality of shot pictures, replacing the default parameters of the plurality of shot pictures displayed on the video editing script template with parameters set by the parameter editing operation.
In the above aspect, the editing module is further configured to perform, for each of the lens frames, the following processing: performing object identification processing on the lens picture, and determining the identified object as a document matched with the lens picture; wherein the types of the objects include: scene, person, event.
In the above aspect, the editing module is further configured to perform, for each of the lens frames, one of the following processes: performing object identification processing on the lens picture, and determining a time length matched with the lens picture according to the number of identified objects, wherein the number of identified objects is positively correlated with the time length; and identifying the similarity between the shot picture and the adjacent shot picture, and determining the time length matched with the shot picture according to the similarity, wherein the similarity is in negative correlation with the time length.
In the above aspect, the editing module is further configured to perform, for each of the lens frames, the following processing: performing object identification processing on the lens picture, and inquiring sound effects with mapping relations between the identified objects in a mapping table, wherein the mapping table comprises a plurality of objects and a plurality of sound effects which are in one-to-one correspondence with the plurality of objects; and determining the sound effect with the mapping relation with the identified object as the sound effect matched with the lens picture.
In the above aspect, the editing module is further configured to perform, for each of the lens frames, the following processing: identifying the similarity between the lens picture and the adjacent lens picture, and inquiring transition matched with the similarity; and determining the transition matched with the similarity as the transition matched with the lens picture.
In the above aspect, the editing module is further configured to perform, for each of the lens frames, the following processing: identifying a history parameter adapted to the lens picture as a parameter adapted to the lens picture; wherein the type of the history parameter includes one of: history parameters corresponding to the history lens picture with highest similarity between the lens pictures; setting a history parameter with highest frequency in the video editing process; historical parameters set in the video editing process which are most closely spaced from the current time point.
In the above scheme, the editing module is further configured to display a collaborative account setting page of the first account in response to a collaborative editing trigger operation received in the document editing interface; the first account is an account logged in the document editing interface, and the collaborative account setting page comprises at least one candidate account; and in response to an account selection operation received on the collaborative account setting page, determining at least one selected candidate account as a second account for collaborating with the first account to edit the video editing script template, and sending the video editing script template to the second account.
In the above scheme, the editing module is further configured to display an editing permission setting entry on the collaborative account setting page; obtaining the set rights in response to a rights setting operation for the editing rights setting portal, wherein the types of rights include: viewing rights, editing rights; and determining that the process of sending a video editing script template applying the rights to the second account is to be executed.
In the above aspect, the video editing processing apparatus further includes: a modification module, configured to display a video modification page in response to a modification operation for the preview video, where the video modification page includes at least one shot frame and a parameter corresponding to each of the shot frames; and responding to the parameter modification operation received at the video modification page, and updating the displayed preview video according to the modified parameter.
In the above scheme, the display module is further configured to display a video type selection page in response to a video editing trigger operation, where the video type selection page includes a plurality of candidate video types; and responding to the video type selection operation received in the video type selection page, and displaying a video editing script template corresponding to the selected video type.
An embodiment of the present application provides an electronic device, including:
a memory for storing computer executable instructions;
and the processor is used for realizing the video editing processing method provided by the embodiment of the application when executing the computer executable instructions stored in the memory.
The embodiment of the application provides a computer readable storage medium, which stores computer executable instructions for implementing the video editing processing method provided by the embodiment of the application when being executed by a processor.
Embodiments of the present application provide a computer program product, where the computer program product includes computer executable instructions for implementing the video editing processing method provided in the embodiments of the present application when the computer executable instructions are executed by a processor.
The embodiment of the application has the following beneficial effects:
the shot picture and the corresponding parameters in the video editing script template can be accurately and efficiently identified on the document editing interface to generate the preview video, so that a user can visually and intuitively know the visual effect of the video editing script, the optimization times of the shot picture and the corresponding parameters can be reduced, editing resources can be saved, and the editing efficiency of the video is improved.
Drawings
FIG. 1 is a schematic diagram of a related art provided shot script;
FIG. 2 is a schematic architecture diagram of a video editing processing system 100 according to an embodiment of the present application;
fig. 3 is a schematic structural diagram of a terminal 400 according to an embodiment of the present application;
fig. 4 is a flow chart of a video editing processing method provided in an embodiment of the present application;
fig. 5 is a flowchart of a video editing processing method according to an embodiment of the present application;
fig. 6 is a flowchart of a video editing processing method according to an embodiment of the present application;
fig. 7A and fig. 7B are schematic application scenarios of the video editing processing method provided in the embodiments of the present application;
fig. 8 is a flowchart of a video editing processing method according to an embodiment of the present application;
fig. 9A and 9B are schematic application scenarios of the video editing processing method provided in the embodiment of the present application;
fig. 10 is an application scenario schematic diagram of a video editing processing method provided in an embodiment of the present application;
fig. 11A and 11B are application scene diagrams of a video editing processing method provided in an embodiment of the present application;
fig. 12 is an application scenario schematic diagram of a video editing processing method provided in an embodiment of the present application;
fig. 13 is an application scenario schematic diagram of a video editing processing method provided in an embodiment of the present application.
Detailed Description
For the purpose of making the objects, technical solutions and advantages of the present application more apparent, the present application will be described in further detail with reference to the accompanying drawings, and the described embodiments should not be construed as limiting the present application, and all other embodiments obtained by those skilled in the art without making any inventive effort are within the scope of the present application.
In the following description, reference is made to "some embodiments" which describe a subset of all possible embodiments, but it is to be understood that "some embodiments" can be the same subset or different subsets of all possible embodiments and can be combined with one another without conflict.
In the following description, the terms "first" and "second" are used merely to distinguish between similar objects and do not represent a particular ordering of the objects, it being understood that the "first" and "second" may be interchanged with a particular order or precedence where allowed, to enable embodiments of the present application described herein to be implemented in other than those illustrated or described herein.
Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this application belongs. The terminology used herein is for the purpose of describing embodiments of the present application only and is not intended to be limiting of the present application.
Before further describing embodiments of the present application in detail, the terms and expressions that are referred to in the embodiments of the present application are described, and are suitable for the following explanation.
1) In response to a condition or state that is used to represent the condition or state upon which the performed operation depends, the performed operation or operations may be in real-time or with a set delay when the condition or state upon which it depends is satisfied; without being specifically described, there is no limitation in the execution sequence of the plurality of operations performed.
2) Computer Vision (CV) is a science of researching how to make a machine "look at", and more specifically, to replace human eyes with a camera and a Computer to perform machine Vision such as recognition, detection and measurement on a target, and further perform graphic processing, so that the Computer processes the target into an image more suitable for human eyes to observe or transmit to an instrument to detect. As a scientific discipline, computer vision research-related theory and technology has attempted to build artificial intelligence systems that can acquire information from images or multidimensional data. Computer vision techniques typically include image processing, image recognition, image semantic understanding, image retrieval, OCR, video processing, video semantic understanding, video content/behavior recognition, three-dimensional object reconstruction, 3D techniques, virtual reality, augmented reality, synchronous positioning, and map construction, among others, as well as common biometric recognition techniques such as face recognition, fingerprint recognition, and others.
3) The shot script, or video editing script, is a series of sketches prepared by a video planner for formal production (shooting), and is a visual prediction before video formation. It is not the final form of video, but the basis of early communication, and also requires collaboration by multiple persons, thus requiring repeated modifications. Generally comprises the following elements: shot number, picture content, text (caption/subtitle), jing Bie (full/medium/near/close), transition (cut/push-pull pan/stack), time length (seconds), music, etc.
Referring to fig. 1, fig. 1 is a schematic diagram of a shot script provided by the related art, and fig. 1 is a handwriting version of the shot script currently used in the film industry, wherein a draft drawn by a lens splitter is included in a shot picture in fig. 1. The process of editing video in the related art generally includes: after establishing the video theme, the planner imagines by editing the shot script as shown in fig. 1 and communicates with multiple persons internally; then, the shot script is sent to a clipping personnel to generate a video through professional clipping software, and the learning cost of the clipping software is high; meanwhile, the static shot script cannot confirm the rhythm of the video and can not accurately estimate the transition and stay time, so that the output video needs to be regenerated after the parameters are modified each time, and the editing cost is high. Still, the static shot script is a traditional story board script, and a user needs to imagine generating videos by himself and cannot implement multi-person collaboration aiming at the shot script.
Aiming at the technical problems, the embodiment of the application provides a video editing processing method which can support timely viewing of corresponding preview videos in the script editing process, so that video editing efficiency is improved. The following describes an exemplary application of the video editing processing method provided by the embodiment of the present application, where the video editing processing method provided by the embodiment of the present application may be implemented by various electronic devices, for example, may be applied to various types of user terminals (hereinafter also simply referred to as terminals) such as smart phones, tablet computers, vehicle-mounted terminals, smart wearable devices, and the like.
Next, taking an electronic device as an example of a terminal, an exemplary application system architecture of the terminal implementing the video editing processing method provided in the embodiment of the present application is described, referring to fig. 2, and fig. 2 is a schematic architecture diagram of the video editing processing system 100 provided in the embodiment of the present application. The video editing processing system 100 includes: the server 200, the network 300, and the terminal 400 will be described separately.
The server 200 is a background server of the client 410, and is configured to receive a plurality of lens frames sent by the client 410 and parameters corresponding to each lens frame; and is further configured to generate a preview video based on the at least one shot and the parameters corresponding to the at least one shot, and send the preview video to the client 410.
The network 300 may be a wide area network or a local area network, or a combination of both, for mediating communication between the server 200 and the terminal 400.
A terminal 400 for running a client 410. A client 410 for displaying a video editing script template in a document editing interface; and is further configured to display the set plurality of shot pictures and parameters corresponding to each shot picture in the video editing script template in response to the video editing operation, and transmit the plurality of shot pictures and parameters corresponding to each shot picture to the server 200; and also receives the preview video transmitted from the server 200 and displays the preview video.
The embodiment of the application can be realized by means of Cloud Technology (Cloud Technology), wherein the Cloud Technology refers to a hosting Technology for integrating serial resources such as hardware, software, network and the like in a wide area network or a local area network to realize calculation, storage, processing and sharing of data.
The cloud technology is a generic term of network technology, information technology, integration technology, management platform technology, application technology and the like based on cloud computing business model application, can form a resource pool, and is flexible and convenient as required. Cloud computing technology will become an important support. Background services of technical network systems require a large amount of computing and storage resources.
As an example, the server 200 may be a stand-alone physical server, a server cluster or a distributed system formed by a plurality of physical servers, or may be a cloud server that provides cloud services, cloud databases, cloud computing, cloud functions, cloud storage, network services, cloud communication, middleware services, domain name services, security services, CDNs, and basic cloud computing services such as big data and artificial intelligence platforms. The terminal 400 and the server 200 may be directly or indirectly connected through wired or wireless communication, which is not limited in the embodiment of the present application.
Next, the structure of the terminal 400 in fig. 2 will be described. Referring to fig. 3, fig. 3 is a schematic structural diagram of a terminal 400 provided in an embodiment of the present application, and the terminal 400 shown in fig. 3 includes: at least one processor 410, a memory 450, at least one network interface 420, and a user interface 430. The various components in terminal 400 are coupled together by a bus system 440. It is understood that the bus system 440 is used to enable connected communication between these components. The bus system 440 includes a power bus, a control bus, and a status signal bus in addition to the data bus. But for clarity of illustration the various buses are labeled in fig. 3 as bus system 440.
The processor 410 may be an integrated circuit chip having signal processing capabilities such as a general purpose processor, such as a microprocessor or any conventional processor, or the like, a digital signal processor (DSP, digital Signal Processor), or other programmable logic device, discrete gate or transistor logic, discrete hardware components, or the like.
The user interface 430 includes one or more output devices 431, including one or more speakers and/or one or more visual displays, that enable presentation of the media content. The user interface 430 also includes one or more input devices 432, including user interface components that facilitate user input, such as a keyboard, mouse, microphone, touch screen display, camera, other input buttons and controls.
Memory 450 may be removable, non-removable, or a combination thereof. Exemplary hardware devices include solid state memory, hard drives, optical drives, and the like. Memory 450 optionally includes one or more storage devices physically remote from processor 410.
Memory 450 includes volatile memory or nonvolatile memory, and may also include both volatile and nonvolatile memory. The non-volatile memory may be read only memory (ROM, read Only Me mory) and the volatile memory may be random access memory (RAM, random Access Memor y). The memory 450 described in the embodiments herein is intended to comprise any suitable type of memory.
In some embodiments, memory 450 is capable of storing data to support various operations, examples of which include programs, modules and data structures, or subsets or supersets thereof, as exemplified below.
The operating system 451 includes system programs, such as a framework layer, a core library layer, a driver layer, etc., for handling various basic system services and performing hardware-related tasks, for implementing various basic services and handling hardware-based tasks.
Network communication module 452 for reaching other computing devices via one or more (wired or wireless) network interfaces 420, exemplary network interfaces 420 include: bluetooth, wireless compatibility authentication (WiFi), and universal serial bus (USB, universal Serial Bus), etc.
A presentation module 453 for enabling presentation of information (e.g., a user interface for operating peripheral devices and displaying content and information) via one or more output devices 431 (e.g., a display screen, speakers, etc.) associated with the user interface 430.
An input processing module 454 for detecting one or more user inputs or interactions from one of the one or more input devices 432 and translating the detected inputs or interactions.
In some embodiments, the video editing processing apparatus provided in the embodiments of the present application may be implemented in a software manner, and fig. 3 shows a video editing processing apparatus 455 stored in a memory 450, which may be software in the form of a program and a plug-in, and includes the following software modules: the display module 4551, the editing module 4552 and the generation module 4553 are logical, and thus may be arbitrarily combined or further split according to the functions implemented. The functions of the respective modules will be described hereinafter.
Next, a video editing processing method provided in the embodiment of the present application is described by way of example in which the terminal 400 in fig. 2 alone is executed. Referring to fig. 4, fig. 4 is a flowchart of a video editing processing method according to an embodiment of the present application, and will be described with reference to the steps shown in fig. 4.
It should be noted that the method shown in fig. 4 may be executed by various computer programs executed by the terminal 400, and is not limited to the above client 410, such as the operating system 451, the software modules, and the scripts described above, and thus the examples of the client should not be considered as limiting the embodiments of the present application.
In step S101, a video editing script template is displayed in a document editing interface.
In some embodiments, the document editing interface may be an editing interface of a local document, where the local document refers to a document opened in a document editing program installed and running locally at the terminal, for example, an editing interface of a local text (Word) document, an editing interface of a local form (Excel) document, an editing interface of a local slide (PPT, power Point) document, and the like; the online document may also be an editing interface of an online document, where the online document refers to a document that runs a text editing program in a server (e.g., cloud end) and displays a document opened in the text editing program in a terminal, for example, an editing interface of an online Word document, an editing interface of an online excel document, an editing interface of an online PPT document, and the like.
In some embodiments, the video editing script template includes a shot screen editing region and a parameter editing region.
As an example, referring to fig. 9A, fig. 9A is an application scene diagram of a video editing processing method provided in the embodiment of the present application, in fig. 9A, a video editing script template 901 includes a shot screen editing area 902 and a parameter editing area 903, and a user may edit a shot screen in the shot screen editing area 902, for example, the user may directly upload a picture or a video in the shot screen editing area 902 as a shot screen; the user can edit parameters such as a shot number, a document, a time length, a transition, a scene, sound effects, music, and the like in the parameter editing area 903.
In some embodiments, in response to a video editing trigger operation, displaying a video type selection page, wherein the video type selection page includes a plurality of candidate video types therein; and responding to the video type selection operation received in the video type selection page, and displaying a video editing script template corresponding to the selected video type.
As an example, the candidate video types include a plurality of video types of different styles, such as a lovely video type, a cool video type, a delicately video type, a fashionable video type, and the like, wherein video editing script templates corresponding to the different video types are also different, and thus, the corresponding video editing script templates can be generated according to the video types selected by the user so as to meet the personalized video editing requirements of the user.
For example, referring to fig. 9B, fig. 9B is a schematic application scenario of the video editing processing method provided in the embodiment of the present application, in fig. 9B, when a user triggers the video editing script portal 904, a video type selection page 905 is presented, where the video type selection page 905 includes a plurality of candidate video types, and when the user triggers the portal of "lovely video type", a video editing script template corresponding to "lovely video type" is presented.
In step S102, in response to the video editing operation, a plurality of lens frames set, and parameters corresponding to each lens frame are displayed in the video editing script template.
In some embodiments, in response to a video editing operation submitted in a video editing script template by at least one of a first account number and a second account number, displaying a plurality of shot pictures set by the video editing operation and parameters corresponding to each shot picture in the video editing script template in real time; the first account is an account for logging in a document editing interface, and the second account is an account for editing a video editing script template in cooperation with the first account.
As an example, when the document editing interface is an editing interface of an online document, the first account may be an account that logs in to the online document.
As an example, the video editing script template supports the first account and the second account to edit, and the display effects of the lens frames set in the video editing script template and the corresponding parameters of different accounts are different, for example, the frame color of the lens frames set in the video editing script template by the first account is different from the frame color of the lens frames set in the video editing script template by the second account, and the fonts (or font size and font color) of the parameters set in the video editing script template by the first account and the fonts (or font size and font color) of the parameters set in the video editing script template by the second account are different, so that a user can intuitively know the account for editing each lens frame and parameter in the video editing script template without communication, thereby saving communication resources.
For example, referring to fig. 10, fig. 10 is a schematic view of an application scenario of the video editing processing method provided in the embodiment of the present application, in fig. 10, parameter 101 is a parameter set in a video editing script template by a first account, parameter 102 is a parameter set in the video editing script template by a second account, and font sizes of parameter 101 and parameter 102 are different.
As an example, prior to the video editing operations submitted in the video editing script template in response to at least one of the first account number and the second account number, may further include: inquiring the state of a first shot picture, wherein the first shot picture is a shot picture which is requested by video editing operation to be edited in a video editing script template or a shot picture which is requested by video editing operation to be edited in the video editing script template and corresponds to parameters; determining that a process responsive to the video editing operation is to be performed when both the first shot and the second shot related to the first shot are in an unedited state; when the first shot picture and the second shot picture related to the first shot picture are in an editing state, first prompt information is displayed, wherein the first prompt information is used for prompting that video editing operation cannot be responded immediately due to editing conflict.
For example, the first shot frame may be any shot frame in the video editing script template, and may also be a shot frame specified by the first account number or the second account number.
For example, the first prompting information is further used for prompting that the first lens frame, the second lens frame, a parameter corresponding to the first lens frame, or a parameter corresponding to the second lens frame is in an editing state; the first prompt information may further include an account number that is editing the first lens frame, the second lens frame, a parameter corresponding to the first lens frame, or a parameter corresponding to the second lens frame.
For example, when the first shot screen is updated from the editing state to the unedited state, it may be that the first shot screen is edited or the first shot screen is edited and withdrawn, if the first shot screen is edited, the editing result needs to be synchronously displayed in the video editing script template, and prompt information capable of responding to the video editing operation is displayed; if the first shot picture is edited and removed, removing the editing result in the video editing script template, and displaying prompt information capable of responding to the video editing operation.
For example, referring to fig. 11A, fig. 11A is a schematic view of an application scenario of the video editing processing method according to the embodiment of the present application, and in fig. 11A, the upper right corner of each column of the shot frame displays a state 111 corresponding to the shot frame, including an edited state and an unedited state. When the user edits the shot screen in the editing state, the first prompt information 112 is presented, and the first prompt information 112 includes an account number which is editing the shot screen. When the user edits the shot screen in the unedited state, the user's editing operation can be directly responded.
For example, the first shot and the second shot satisfy at least one of the following association conditions: the first shot picture and the second shot picture are edited into the video editing script template by the same account number; the first shot and the second shot belong to the same scene, for example, the first shot and the second shot are both indoor scenes.
In the embodiment of the application, the association degree of the shot pictures belonging to the same scene is higher, so that when a user edits a certain shot picture, the possibility of editing the rest shot pictures belonging to the same scene is higher, thereby locking the shot pictures belonging to the same scene in an exclusive mode, reducing the optimization times of the shot pictures in the script, and further saving editing resources. Similarly, the association degree of the lens pictures edited by the same account number into the video editing script template is higher, so that the lens pictures edited by the same account number into the video editing script template are locked in an exclusive mode, the optimization times of the lens pictures can be reduced, and editing resources are saved.
As an example, prior to the video editing operations submitted in the video editing script template in response to at least one of the first account number and the second account number, may further include: inquiring the state of a first parameter, wherein the first parameter is a parameter corresponding to a third shot picture edited in a video editing script template by a video editing operation request; determining that processing responsive to the video editing operation is to be performed when the first parameter and the second parameter associated with the first parameter are both in an unedited state; when the first parameter and the second parameter related to the first parameter are both in the editing state, displaying second prompt information, wherein the second prompt information is used for prompting that the video editing operation cannot be responded immediately due to editing conflict.
For example, the first parameter may be any parameter in the video editing script template, and may also be a parameter specified by the first account or the second account.
For example, the second prompting information is further used for prompting that the parameter corresponding to the third lens frame is in an editing state; the second prompt information may further include an account number that is editing parameters corresponding to the third shot screen.
For example, when the first parameter is updated from the editing state to the unedited state, it may be that the first parameter editing is completed or the first parameter editing is cancelled, if the first parameter editing is completed, the editing result needs to be synchronously displayed in the video editing script template, and a prompt message that can respond to the video editing operation is displayed; if the first parameter editing is cancelled, the editing result is cancelled in the video editing script template, and prompt information which can respond to the video editing operation is displayed.
For example, referring to fig. 11B, fig. 11B is a schematic view of an application scenario of the video editing processing method according to the embodiment of the present application, and in fig. 11B, the upper right corner of each column of parameters displays a state 113 corresponding to the parameter, including an edited state and an unedited state. When the user edits the parameter in the editing state, a second prompt 114 is presented, and the second prompt 114 includes an account number that is editing the parameter. When the user edits the parameter in the unedited state, the user's editing operation may be directly responded.
For example, the first parameter and the second parameter satisfy at least one of the following associated conditions: the first parameter and the second parameter are edited by the same account number into a video editing script template; the lens picture corresponding to the first parameter is the same as the lens picture corresponding to the second parameter; the first parameter is of the same type as the second parameter.
In the embodiment of the application, the parameter edited by the same account number into the video editing script template has higher association degree, so that when a user edits a certain parameter, the possibility of editing the parameter also edited by the user into the video editing script template is higher, the parameter edited by the same account number into the video editing script template is locked in an exclusive mode, the optimization times of the parameter in the script can be reduced, and editing resources can be saved. Similarly, the parameter association degree and the parameter association degree with the same type corresponding to the same lens picture are also higher, so that the parameter with the same type corresponding to the same lens picture is locked in an exclusive mode, the optimization times of the parameter can be reduced, and editing resources are saved.
In some embodiments, the video editing operations include a shot editing operation and a parameter editing operation; referring to fig. 5, fig. 5 is a flowchart of a video editing processing method provided in an embodiment of the present application, and based on fig. 4, step S102 may include steps S1021 to S1022.
In step S1021, in response to the shot screen editing operation, a plurality of shot screens set by the shot screen editing operation are displayed in the video editing script template.
In some embodiments, in response to a shot screen editing operation for a shot screen editing region, a plurality of shot screens set by the shot screen editing operation are acquired through the shot screen editing region, and the plurality of shot screens set by the shot screen editing operation are displayed in the shot screen editing region.
As an example, the shot screen editing operation may be submitted by the first account or the second account. The display effects of the lens frames set by the different accounts in the video editing script template are different, for example, the border colors of the lens frames set by the first account in the video editing script template are different from the border colors of the lens frames set by the second account in the video editing script template.
For example, in fig. 9A, the user may edit the shot in the shot editing area 902, for example, the user may upload a picture or video directly in the shot editing area 902 as the shot.
In step S1022, in response to the parameter editing operation, parameters of the corresponding plurality of shot pictures set by the parameter editing operation are displayed in the video editing script template.
In some embodiments, in response to a parameter editing operation for a parameter editing area, parameters of a plurality of shots set by the parameter editing operation are acquired through the parameter editing area, and the parameters of the plurality of shots set by the parameter editing operation are displayed in the parameter editing area.
As an example, the parameter editing operation may be submitted by the first account or the second account. The display effects of the parameters set by the different accounts in the video editing script template are different, for example, the fonts (or font sizes and font colors) of the parameters set by the first account in the video editing script template are different from the fonts (or font sizes and font colors) of the parameters set by the second account in the video editing script template.
For example, in fig. 9A, the user can edit parameters such as a lens number, a document, a duration, a transition, a scene, an audio effect, and music in the parameter editing area 903.
In some embodiments, performing parameter identification processing on each lens picture to obtain parameters adapted to a plurality of lens pictures, using the parameters adapted to the plurality of lens pictures as default parameters, and displaying the parameters in a video editing script template; in response to a parameter editing operation for a plurality of shot pictures, replacing default parameters of the plurality of shot pictures displayed in the video editing script template with parameters set by the parameter editing operation to replace the default parameters.
The procedure of performing the parameter identification processing for each lens frame is described below in connection with a plurality of examples.
As a first example, the following processing is performed for each lens screen: performing object recognition processing on the lens picture, and determining the recognized object as a document matched with the lens picture; wherein, the types of the objects include: scene, person, event.
For example, when the shot screen is a cyclist's screen, the text "cyclist" may be automatically populated in the text parameters. Therefore, by automatically filling the text parameters, a user does not need to fill corresponding text in the video editing script template according to the lens picture, so that the editing efficiency of the video is improved.
As a second example, one of the following processes is performed for each lens screen: performing object recognition processing on the lens picture, and determining the time length matched with the lens picture according to the number of recognized objects, wherein the number of recognized objects is positively correlated with the time length; and identifying the similarity between the shot picture and the adjacent shot picture, and determining the time length matched with the shot picture according to the similarity, wherein the similarity is inversely related to the time length.
For example, the more the number of objects in a shot, the more the amount of information characterizing the shot, and the more time it takes for a user to understand the shot, so a duration parameter that is positively correlated to the number of identified objects can be set, so that the number of operations of the user can be reduced while ensuring video editing accuracy. Similarly, the higher the similarity between the shot picture and the adjacent shot picture, the higher the association degree of the two shot pictures, the less time is consumed for users to understand the association between the two shot pictures, therefore, the duration parameter negatively related to the similarity can be set, and the operation times of the users can be reduced while the video editing accuracy is ensured.
As a third example, the following processing is performed for each lens screen: identifying a history parameter matched with the lens picture as a parameter matched with the lens picture; wherein the type of history parameter includes one of: historical parameters corresponding to the historical lens picture with highest similarity with the lens picture; setting a history parameter with highest frequency in the video editing process; historical parameters set in the video editing process which are most closely spaced from the current time point. Therefore, by multiplexing the history parameters, a user does not need to upload corresponding parameters in the video editing script template according to the shot picture, so that the editing efficiency of the video is improved.
As a fourth example, the following processing is performed for each lens screen: performing object recognition processing on the lens picture, and inquiring sound effects with mapping relations between the recognized objects in a mapping table, wherein the mapping table comprises a plurality of objects and a plurality of sound effects which are in one-to-one correspondence with the plurality of objects; and determining the sound effect with the mapping relation with the identified object as the sound effect matched with the lens picture.
For example, when the shot screen is a screen of a rider, it can be identified that the object included in the shot screen is a "bicycle", so that the sound effect corresponding to the "bicycle", for example, a "bell sound", can be queried in the local mapping table as the sound effect parameter, and of course, the sound effect corresponding to the "bicycle" can also be queried directly in the network as the sound effect parameter. Here, by automatically filling the sound effect parameters, the user does not need to upload the corresponding sound effect in the video editing script template according to the lens picture, thereby improving the editing efficiency of the video.
As a fifth example, the following processing is performed for each lens screen: identifying the similarity between the lens picture and the adjacent lens picture, and inquiring transition matched with the similarity; and determining the transition matched with the similarity as the transition matched with the lens picture.
For example, the process of identifying the similarity between a shot and an adjacent shot may include: the method comprises the steps of extracting a first image feature from a lens picture, extracting a second image feature from an adjacent lens picture, determining a geometric distance (such as chebyshev distance, euclidean distance, minkowski distance, and the like) between the first image feature and the second image feature, and determining the similarity between the lens picture and the adjacent lens picture according to the geometric distance.
For example, when the similarity is smaller than a similarity threshold, setting the transition adaptive to the shot image as "hard cut", wherein the similarity threshold may be a default value or a value set by a user, a client or a server; when the similarity is not smaller than the similarity threshold, the transition matched with the lens picture is set as push-pull or overlapping, so that the transition between the lens picture with higher similarity and the adjacent lens picture is smooth and natural.
In step S103, in response to the video preview operation received in the video editing process, a preview video is generated based on at least one shot picture and parameters corresponding to the at least one shot picture.
In some embodiments, the client may invoke a corresponding service of the terminal (e.g., a preview video generation service), through which the process of generating the preview video is completed. The client may also invoke a corresponding service of the server (e.g., a preview video generation service) to complete the process of generating the preview video through the server.
As an example, when the client invokes the corresponding service of the server to complete the process of generating the preview video, the alternative step of step S103 may be: the client side responds to video preview operation received in the video editing process and sends a preview video generation request to the server; the server responds to the preview video generation request and generates a preview video based on at least one shot picture and parameters corresponding to the at least one shot picture; and sending the preview video to the client.
In the following, a process of generating a preview video by a terminal is described by taking a procedure of calling a corresponding service of the terminal by the client terminal as an example. It should be noted that, the process of the client calling the corresponding service of the server to complete the generation of the preview video is similar to the following process, and will not be described in detail.
As an example, in fig. 7B, when the user triggers the video generation portal 703, a preview video 704 may be generated from the content edited by the user in the video editing script template 702.
In some embodiments, the preview video may be generated based on all of the shots in the video editing script template and corresponding parameters; preview videos may also be generated based on the partial shot frames and corresponding parameters in the video editing script template.
As an example, in response to a shot selection operation, a preview video is generated based on the selected shot and corresponding parameters. Therefore, the user can watch the preview video comprising the selected lens picture, so that the personalized requirement of the user is met, and the resources consumed for generating the preview video are saved.
Here, the parameters include a clip parameter and a join parameter; the clip parameters include at least one of: a document, a duration and an audio effect; the engagement parameters include at least one of: lens serial number, transition.
In some embodiments, referring to fig. 6, fig. 6 is a flowchart of a video editing processing method provided in an embodiment of the present application, and based on fig. 4, step S103 may include steps S1031 to S1033.
In step S1031, in response to the video preview operation received in the video editing process, the following processing is performed for each shot: and according to the clipping parameters of the shot pictures, clipping the shot pictures to obtain shot fragments of the corresponding shot pictures.
In some embodiments, the following processing is performed for each shot: clipping the shot frames into pre-processing fragments, wherein the playing time length of the pre-processing fragments is the time length in clipping parameters of the shot frames; and adding the text in the clipping parameters of the shot picture into the preprocessing segment, and adding the sound effect in the clipping parameters into the preprocessing segment to obtain the shot segment of the corresponding shot picture.
For example, firstly, a shot frame is clipped into a pre-processing segment, then, a document in the clipping parameters is overlapped in the pre-processing segment in a floating layer mode, and finally, the pre-processing segment is filled with sound effects in the clipping parameters to obtain the shot segment of the corresponding shot frame.
In step S1032, when the number of at least one shot is one, the shot segment of the corresponding shot is determined as the preview video.
In some embodiments, when generating the preview video based on one shot frame and the corresponding parameters, since one shot frame corresponds to only one shot segment and one shot segment does not need to be spliced, the splicing parameters of the shot frame do not need to be considered, and the corresponding preview video can be generated only according to the clipping parameters corresponding to the shot frame.
In step S1033, when the number of at least one shot is plural, the shot segments corresponding to each shot are combined according to the linking parameter of each shot to obtain the preview video.
In some embodiments, when the number of at least one shot frame is a plurality of, sequencing the plurality of shot segments according to the sequence of the shot sequence numbers in the linking parameter to obtain a shot segment sequence; the following connection processing is performed in sequence for each shot segment in the shot segment sequence: connecting the lens segment with the adjacent lens segment according to transition in the connection parameters corresponding to the lens segment; and taking the shot fragment sequence after the connection processing as a preview video.
As an example, the manner of splicing between the lens segments is based on the transition in the splicing parameter, for example, when the transition in the splicing parameter is the superimposition for N seconds, the lens segment may be lengthened for N seconds, the transparency of the lens segment is gradually reduced from 100 to 0 during this period, and the transparency of the next adjacent lens segment is gradually increased from 0 to 100, so as to realize the smooth process presentation between the lens segments.
In step S104, a preview video is displayed.
In some embodiments, displaying the preview video may be by integrating a display box of the preview video into a document editing interface for display; the video editing program may call a popup window of the player program to display the preview video, or the video editing program may call the player program to display the preview video in a window of the video editing program itself.
In some embodiments, after step S104, the preview video (i.e., the export video file) may also be downloaded in response to the video editing completion operation, so that the user may save the generated video after previewing the video.
By way of example, in FIG. 7B, preview video 704 is displayed and a video download portal 705 is included in the document editing interface, where preview video 704 may be downloaded when the user triggers video download portal 705.
In the embodiment of the application, the shot picture and the corresponding parameters in the video editing script template can be accurately and efficiently identified on the document editing interface to generate the preview video, so that a user can visually and intuitively know the content logic of the video, the technical problems that in the related technology, the video rhythm cannot be confirmed due to static shot script, and the video needs to be regenerated after the parameters are modified each time, so that the video editing cost is high can be solved, the optimization times of the shot picture and the corresponding parameters can be reduced, editing resources can be saved, and the video editing efficiency is improved.
In some embodiments, after step S104, it may further include: in response to a modification operation for the preview video, displaying a video modification page, wherein the video modification page comprises at least one shot picture and parameters corresponding to each shot picture; and responding to the parameter modification operation received in the video modification page, and updating the displayed preview video according to the modified parameter.
As an example, referring to fig. 12, fig. 12 is an application scenario schematic diagram of a video editing processing method provided in the embodiment of the present application, in fig. 12, when a user clicks a certain frame of shot frame in a preview video 121, a floating window displays a video modification page 122 corresponding to the clicked shot frame, a user may modify parameters corresponding to the shot frame in the video modification page 122, and after the user modifies, the preview video 121 may be updated by triggering a determination button 123, so that man-machine interaction efficiency may be improved in the modification process of the preview video.
In some embodiments, after step S101, it may further include: responding to a collaborative editing triggering operation received in a document editing interface, and displaying a collaborative account setting page of a first account; the first account is an account logged in a document editing interface, and the collaborative account setting page comprises at least one candidate account; and in response to the account selection operation received on the collaborative account setting page, determining the selected at least one candidate account as a second account for collaborating with the first account to edit the video editing script template, and sending the video editing script template to the second account.
As an example, an editing right setting entry is displayed on a collaborative account setting page; in response to a rights setting operation for an editing rights setting portal, acquiring the set rights, wherein the types of rights include: viewing rights, editing rights; a determination is made that a process of sending the video editing script template of the application rights to the second account is to be performed.
For example, for an account with viewing rights, updates to the video editing script template may be viewed in real-time; for the account with the editing authority, not only can the update of the video editing script template be checked in real time, but also the video editing script template can be edited.
For example, the same authority may be globally set for all the shot images and parameters in the video editing script template, for example, the second account may view and edit all the shot images and parameters in the video editing script template; different permissions can also be set for each shot picture and parameter or each type of shot picture and parameter in the video editing script template, for example, the second account can view the update of parameter 1 in real time and edit parameter 1, but can only view the update of parameter 2 in real time and cannot edit parameter 2.
For example, in addition to the above manner of manually distributing the rights of the video editing script template through the first account, the rights of the second account may be automatically distributed through the client, for example, corresponding rights may be automatically distributed according to roles of the second account, where the roles of the second account may be set by the first account or the second account, or corresponding rights may be automatically distributed according to the activity level of the second account, where the activity level of the second account is positively correlated with the number of times the second account participates in editing, and the activity level of the second account is positively correlated with the number of times the second account interacts based on editing.
For example, referring to fig. 13, fig. 13 is a schematic view of an application scenario of a video editing processing method provided in an embodiment of the present application, in fig. 13, when a user triggers a collaboration portal 131, a collaboration account setting page 132 is displayed, the user may select an account to be sent in the collaboration account setting page 132 and set the authority of a video editing script template in an editing authority setting portal 133, and after the user selects the authority and the account to be sent, the video editing script template of the application authority may be sent to the selected account.
According to the embodiment of the application, a multi-person cooperation function is provided for the video editing script template, so that the technical problems that in the related art, due to the fact that multi-person cooperation cannot be implemented for static shot scripts, editing cost is high and editing efficiency is low can be solved, communication efficiency in a video editing process can be improved, optimization times of shot pictures and corresponding parameters are reduced, and editing resources are saved.
The video editing processing method provided in the embodiment of the present application is described below by taking an online document as an example.
According to the method and the system for generating the video, the video can be generated rapidly through literal editing of the shot script, for example, after a video planner fills out and selects the structured words (namely the parameters) corresponding to the shot painting, the system can automatically identify and generate the preview video, and therefore the efficiency of video production is improved. In addition, the embodiment of the application also utilizes the characteristic that the online document can be cooperated by multiple persons and modified in real time, improves the communication efficiency, carries out structural processing on the constituent information of the storyboard shot script, directly generates the preview video without imagination of a user, and further improves the video generation efficiency on the basis of not changing the industry flow and the habit of the user.
Referring to fig. 7A, fig. 7B, and fig. 8, fig. 7A and fig. 7B are schematic application scenarios of the video editing processing method provided in the embodiment of the present application, and fig. 8 is a flow diagram of the video editing processing method provided in the embodiment of the present application. A specific implementation of the embodiment of the present application will be described next with reference to fig. 7A, 7B, and 8.
In step S801, the terminal displays a video editing script template in response to a video editing trigger operation received in an online document.
In some embodiments, in FIG. 7A, when a user clicks on a video editing script entry 701 in an online document, a video editing script template 702 as shown in FIG. 7B may be created and displayed in a human-machine interaction interface.
In step S802, the terminal displays a plurality of shot pictures and parameters corresponding to each shot picture in the video editing script template in response to the video editing operation.
In some embodiments, in FIG. 7B, the user may fill in (or upload) multiple shot frames in the video editing script template 702, as well as parameters (including shot sequence number, text, duration, transition, sound effects, music, shot frame internal motion) corresponding to each shot frame. The editing of the shot picture can copy and paste or upload pictures or videos, if a plurality of pictures exist, the pictures are arranged in sequence, and the videos need to fill in the using period in the uploading stage. The shot number edit is typically a default fill. The text editing can directly edit the corresponding text, and can also select whether to use the text as a subtitle or a side note, or else, the text is not displayed by default. Duration edit, i.e., indicating the dwell time of the line (shot), where the user can fill in the "number + seconds(s)", otherwise the default presentation is 1s. Transition editing, namely indicating the connection relation between the shot picture and the next shot picture, can be selected (options comprise hard cutting, push-pull, overlapping and the like), and needs to input time length, otherwise, the shot picture is directly hard cut by default. In the sound effect editing, the user is supported to upload an audio file in an mp3 format or input a text (such as clock sound), and the server can perform network searching and downloading according to the input text to obtain a corresponding sound effect, wherein the using time of the sound effect depends on the duration of a shot picture. In the music editing, the user is supported to upload the audio file in the mp3 format, and the using time period is filled in the uploading stage to be used as a whole. The specific data editing method is not limited to this specific one, and depends on the actual scenario.
In some embodiments, the content in the video editing script template may be edited by an account logged into the online document (i.e., the first account described above) or by a collaborative account (i.e., the second account described above).
In step S803, the terminal transmits the plurality of shot pictures and the parameters corresponding to each shot picture to the server, and the server generates a preview video according to the plurality of shot pictures and the parameters corresponding to each shot picture and transmits the preview video to the terminal.
In some embodiments, the process of generating the preview video may include: firstly, respectively generating lens fragments of each lens picture according to lens sequence numbers, and then combining all the lens fragments to generate a complete preview video, wherein the duration of the preview video is X seconds, a music parameter in a document can be called, and full-film playing is started according to the lens sequence number of a row where the music parameter is located, and the playing duration of music is X seconds.
As an example, the process of generating a shot segment for each shot may include: for the video part, a "shot picture" in the document can be called first, for example, a picture or a video in the "shot picture" is compressed or amplified according to a 1080 x 720 bit specification, and the fit without cutting is ensured, wherein the time length of the picture stay calls a time length parameter M seconds in the document. Then judging whether a text is present or not and whether a subtitle is checked or not, if the text is present or not, calling text information in the text, and overlapping the text information on the picture by using a 32-bit white Song body and a black background strip based on the total word number length. For the audio part, firstly, the "sound effect" parameter in the document can be called for filling (the empty content is silent), and the playing time is M seconds; and then according To whether the Text file parameter has the bypass or not, calling the Text information To convert the Text information into sound information through Text To Speech (TTS) technology, if the sound playing time is not longer than M seconds (namely, the time length parameter in the document), not processing, and if the sound playing time is longer than M seconds, accelerating the bypass sound To play.
As an example, the manner of linking between the shot segments is based on a "transition" parameter in the document, for example, when the "transition" parameter of a shot segment is N seconds of superimposition, the shot segment is extended by N seconds, the transparency of the shot segment is gradually reduced from 100 to 0 during this time, and the transparency of the next shot segment is gradually increased from 0 to 100, so that smooth process presentation is realized. If the transition parameter is empty, the processing is not performed, and the transition parameter is directly hard-cut.
In step S804, a preview video is displayed in response to the video preview operation.
In some embodiments, in fig. 7B, when the user triggers the video generation portal 703, the server may generate the preview video 704 according to the content filled in by the user in the video editing script template 702, where the size of the preview video 704 may default to 1080 x 720, and the user may modify the content in the video editing script template 702 according to the preview video 704 without the fit of the material of each shot. And, when the user triggers the video download portal 705, the preview video 704 may be downloaded.
According to the video generation method and device, the video can be generated rapidly through literal editing of the shot script, and the cost of generating the video can be reduced and the video generation efficiency can be improved through multi-user cooperation and real-time modification.
An exemplary structure of the video editing processing apparatus according to the embodiment of the present application implemented as a software module is described below with reference to fig. 3.
In some embodiments, as shown in fig. 3, the software modules stored in the video editing processing device 455 of the memory 450 may include: a display module 4551 for displaying a video editing script template in the document editing interface; an editing module 4552 for displaying, in response to a video editing operation, a plurality of lens pictures set and parameters corresponding to each lens picture in a video editing script template; a generating module 4553, configured to generate a preview video based on at least one shot picture and parameters corresponding to the at least one shot picture in response to a video preview operation received during video editing; the display module 4551 is further configured to display a preview video.
In the above scheme, the parameters include a clipping parameter and a linking parameter; the generating module 4553 is further configured to perform the following processing for each lens frame: according to the clipping parameters of the shot pictures, clipping the shot pictures to obtain shot fragments corresponding to the shot pictures; when the number of at least one shot picture is one, determining a shot segment of the corresponding shot picture as a preview video; when the number of at least one lens picture is a plurality of, combining the lens fragments corresponding to each lens picture according to the connection parameter of each lens picture to obtain a preview video.
In the above scheme, the clip parameter includes at least one of: a document, a duration and an audio effect; the generating module 4553 is further configured to clip the shot frame into a pre-processing segment, where a playing duration of the pre-processing segment is a duration in a clipping parameter of the shot frame; and adding the text in the clipping parameters of the shot picture into the preprocessing segment, and adding the sound effect in the clipping parameters into the preprocessing segment to obtain the shot segment of the corresponding shot picture.
In the above scheme, the linking parameter includes at least one of: lens sequence number, transition; the generating module 4553 is further configured to sort the plurality of lens segments according to the sequence of the lens sequence in the linking parameter to obtain a lens segment sequence; the following connection processing is performed in sequence for each shot segment in the shot segment sequence: connecting the lens segment with the adjacent lens segment according to transition in the connection parameters corresponding to the lens segment; and taking the shot fragment sequence after the connection processing as a preview video.
In the above-mentioned scheme, the editing module 4552 is further configured to display, in response to a video editing operation submitted in the video editing script template by at least one of the first account number and the second account number, a plurality of shot frames set by the video editing operation and parameters corresponding to each shot frame in the video editing script template; the first account is an account for logging in a document editing interface, and the second account is an account for editing a video editing script template in cooperation with the first account.
In the above-mentioned scheme, the editing module 4552 is further configured to query a status of a first shot, where the first shot is a shot that the video editing operation requests to edit in the video editing script template, or a shot that the video editing operation requests to edit in the video editing script template corresponds to a parameter; determining that a process responsive to the video editing operation is to be performed when both the first shot and the second shot related to the first shot are in an unedited state; when the first shot picture and the second shot picture related to the first shot picture are in an editing state, first prompt information is displayed, wherein the first prompt information is used for prompting that video editing operation cannot be responded immediately due to editing conflict.
In the above scheme, the first lens picture and the second lens picture meet at least one of the following association conditions: the first shot picture and the second shot picture are edited into the video editing script template by the same account number; the first shot and the second shot belong to the same scene.
In the above solution, the editing module 4552 is further configured to query a state of a first parameter, where the first parameter is a parameter corresponding to a third shot frame that is edited in the video editing script template by the video editing operation request; determining that processing responsive to the video editing operation is to be performed when the first parameter and the second parameter associated with the first parameter are both in an unedited state; when the first parameter and the second parameter related to the first parameter are both in the editing state, displaying second prompt information, wherein the second prompt information is used for prompting that the video editing operation cannot be responded immediately due to editing conflict.
In the above scheme, the first parameter and the second parameter satisfy at least one of the following association conditions: the first parameter and the second parameter are edited by the same account number into a video editing script template; the lens picture corresponding to the first parameter is the same as the lens picture corresponding to the second parameter; the first parameter is of the same type as the second parameter.
In the above scheme, the video editing operation includes a shot screen editing operation and a parameter editing operation; the editing module 4552 is further configured to display a plurality of shot pictures set by the shot picture editing operation in the video editing script template in response to the shot picture editing operation; in response to the parameter editing operation, parameters of the corresponding plurality of shot pictures set by the parameter editing operation are displayed in the video editing script template.
In the above scheme, the editing module 4552 is further configured to perform parameter identification processing on each lens frame to obtain parameters adapted to a plurality of lens frames, and display the parameters adapted to the plurality of lens frames as default parameters in the video editing script template; in response to a parameter editing operation for a plurality of shot pictures, default parameters of the plurality of shot pictures are displayed in a video editing script template and replaced by parameters set by the parameter editing operation.
In the above-described aspect, the editing module 4552 is further configured to perform, for each shot, the following processing: performing object recognition processing on the lens picture, and determining the recognized object as a document matched with the lens picture; wherein, the types of the objects include: scene, person, event.
In the above-described aspect, the editing module 4552 is further configured to perform, for each shot, one of the following processes: performing object recognition processing on the lens picture, and determining the time length matched with the lens picture according to the number of recognized objects, wherein the number of recognized objects is positively correlated with the time length; and identifying the similarity between the shot picture and the adjacent shot picture, and determining the time length matched with the shot picture according to the similarity, wherein the similarity is inversely related to the time length.
In the above-described aspect, the editing module 4552 is further configured to perform, for each shot, the following processing: performing object recognition processing on the lens picture, and inquiring sound effects with mapping relations between the recognized objects in a mapping table, wherein the mapping table comprises a plurality of objects and a plurality of sound effects which are in one-to-one correspondence with the plurality of objects; and determining the sound effect with the mapping relation with the identified object as the sound effect matched with the lens picture.
In the above-described aspect, the editing module 4552 is further configured to perform, for each shot, the following processing: identifying the similarity between the lens picture and the adjacent lens picture, and inquiring transition matched with the similarity; and determining the transition matched with the similarity as the transition matched with the lens picture.
In the above-described aspect, the editing module 4552 is further configured to perform, for each shot, the following processing: identifying a history parameter matched with the lens picture as a parameter matched with the lens picture; wherein the type of history parameter includes one of: historical parameters corresponding to the historical lens picture with highest similarity with the lens picture; setting a history parameter with highest frequency in the video editing process; historical parameters set in the video editing process which are most closely spaced from the current time point.
In the above solution, the editing module 4552 is further configured to display a collaborative account setting page of the first account in response to a collaborative editing trigger operation received in the document editing interface; the first account is an account logged in a document editing interface, and the collaborative account setting page comprises at least one candidate account; and in response to the account selection operation received on the collaborative account setting page, determining the selected at least one candidate account as a second account for collaborating with the first account to edit the video editing script template, and sending the video editing script template to the second account.
In the above solution, the editing module 4552 is further configured to display an editing permission setting entry on the collaborative account setting page; in response to a rights setting operation for an editing rights setting portal, acquiring the set rights, wherein the types of rights include: viewing rights, editing rights; a determination is made that a process of sending the video editing script template of the application rights to the second account is to be performed.
In the above-described aspect, the video editing processing apparatus 455 further includes: a modification module for displaying a video modification page in response to a modification operation for the preview video, wherein the video modification page includes at least one shot picture and a parameter corresponding to each shot picture; and responding to the parameter modification operation received in the video modification page, and updating the displayed preview video according to the modified parameter.
In the above solution, the display module 4551 is further configured to display a video type selection page in response to a video editing trigger operation, where the video type selection page includes a plurality of candidate video types; and responding to the video type selection operation received in the video type selection page, and displaying a video editing script template corresponding to the selected video type.
Embodiments of the present application provide a computer program product or computer program comprising computer instructions stored in a computer readable storage medium. The processor of the computer device reads the computer instructions from the computer-readable storage medium, and the processor executes the computer instructions, so that the computer device executes the video editing processing method according to the embodiment of the present application.
The embodiments of the present application provide a computer-readable storage medium storing computer-executable instructions, in which the computer-executable instructions are stored, which when executed by a processor, cause the processor to perform the video editing processing method provided by the embodiments of the present application, for example, the video editing processing method shown in fig. 4, 5, 6 and 8, and the computer includes various computing devices including a smart terminal and a server.
In some embodiments, the logic of the video editing processing method provided in the embodiments of the present application may be implemented in an intelligent contract, and the node (e.g., a server) generates a preview video by calling the intelligent contract, and stores the preview video in the blockchain network, so that the blockchain network responds to a preview request of a client for the preview video according to the stored preview video, thereby improving reliability of obtaining the preview video through the blockchain network.
In some embodiments, the computer readable storage medium may be FRAM, ROM, PROM, EP ROM, EEPROM, flash memory, magnetic surface memory, optical disk, or CD-ROM; but may be a variety of devices including one or any combination of the above memories.
In some embodiments, computer-executable instructions may be written in any form of programming language, including compiled or interpreted languages, or declarative or procedural languages, in the form of programs, software modules, scripts, or code, and they may be deployed in any form, including as stand-alone programs or as modules, components, subroutines, or other units suitable for use in a computing environment.
As an example, computer-executable instructions may, but need not, correspond to files in a file system, may be stored in a portion of a file that holds other programs or data, such as in one or more scripts in a hypertext markup language document, in a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub-programs, or portions of code).
As an example, computer-executable instructions may be deployed to be executed on one computing device or on multiple computing devices located at one site or, alternatively, distributed across multiple sites and interconnected by a communication network.
In summary, in the embodiment of the present application, the shot frame and the corresponding parameters in the video editing script template can be accurately and efficiently identified and a preview video can be generated on the document editing interface, so that a user can intuitively understand the visual effect of the video editing script through vision, thereby reducing the optimization times of the shot frame and the corresponding parameters, further saving editing resources, and improving the editing efficiency of the video.
The foregoing is merely exemplary embodiments of the present application and is not intended to limit the scope of the present application. Any modifications, equivalent substitutions, improvements, etc. that are within the spirit and scope of the present application are intended to be included within the scope of the present application.

Claims (14)

1. A video editing processing method, the method comprising:
displaying a video editing script template in a document editing interface;
in response to a shot screen editing operation in a video editing operation, displaying a plurality of shot screens set by the shot screen editing operation in the video editing script template;
Carrying out parameter identification processing on each lens picture to obtain parameters matched with the plurality of lens pictures, and displaying the parameters matched with the plurality of lens pictures in the video editing script template;
in response to a video preview operation received during video editing, generating a preview video based on at least one shot and parameters corresponding to the at least one shot, and displaying the preview video.
2. The method of claim 1, wherein the step of determining the position of the substrate comprises,
the parameters comprise a clipping parameter and a linking parameter;
the generating the preview video based on the at least one shot picture and the parameters corresponding to the at least one shot picture includes:
the following processing is performed for each of the lens pictures: according to the clipping parameters of the shot pictures, clipping the shot pictures to obtain shot fragments corresponding to the shot pictures;
when the number of the at least one shot picture is one, determining a shot segment corresponding to the shot picture as the preview video;
and when the number of the at least one lens picture is a plurality of, combining the lens segments corresponding to each lens picture according to the connection parameter of each lens picture to obtain the preview video.
3. The method of claim 2, wherein the clipping parameters comprise: the text, duration and sound effects;
the step of clipping the shot picture according to the clipping parameters of the shot picture to obtain a shot segment corresponding to the shot picture comprises the following steps:
clipping the shot picture into a preprocessing segment, wherein the playing time length of the preprocessing segment is the time length in clipping parameters of the shot picture;
and adding the text in the clipping parameters of the shot picture into the preprocessing segment, and adding the sound effect in the clipping parameters into the preprocessing segment to obtain the shot segment corresponding to the shot picture.
4. The method of claim 2, wherein the engagement parameters comprise: lens sequence number and transition;
combining the lens segments corresponding to each lens picture according to the connection parameters of each lens picture to obtain the preview video, wherein the method comprises the following steps:
sequencing the plurality of lens fragments according to the sequence of the lens sequence numbers in the connection parameters to obtain a lens fragment sequence;
the following connection processing is sequentially performed for each of the shot segments in the shot segment sequence: connecting the lens segment with the adjacent lens segment according to transition in the connection parameters corresponding to the lens segment;
And taking the shot fragment sequence after the connection processing as the preview video.
5. The method of claim 1, wherein displaying the set plurality of shot frames and the parameters corresponding to each of the shot frames in the video editing script template in response to a video editing operation comprises:
responding to video editing operation submitted in the video editing script template by at least one of a first account and a second account, and displaying a plurality of shot pictures set by the video editing operation and parameters corresponding to each shot picture in the video editing script template;
the first account is an account logged in the document editing interface, and the second account is an account cooperated with the first account to edit the video editing script template.
6. The method of claim 5, wherein prior to the video editing operations submitted in the video editing script template in response to at least one of the first account number and the second account number, the method further comprises:
inquiring the state of a first shot picture, wherein the first shot picture is a shot picture which is edited in the video editing script template by the video editing operation request or a shot picture which corresponds to parameters which are edited in the video editing script template by the video editing operation request;
Determining that a process responsive to the video editing operation is to be performed when the first shot and a second shot related to the first shot are both in an unedited state;
and when the first lens picture and the second lens picture related to the first lens picture are in an editing state, displaying first prompt information, wherein the first prompt information is used for prompting that the video editing operation cannot be responded immediately due to editing conflict.
7. The method of claim 6, wherein the step of providing the first layer comprises,
the first lens picture and the second lens picture satisfy at least one of the following association conditions:
the first lens picture and the second lens picture are edited into the video editing script template by the same account number;
the first shot picture and the second shot picture belong to the same scene.
8. The method of claim 5, wherein prior to the video editing operations submitted in the video editing script template in response to at least one of the first account number and the second account number, the method further comprises:
inquiring the state of a first parameter, wherein the first parameter is a parameter corresponding to a third shot picture edited in the video editing script template by the video editing operation request;
Determining that processing responsive to the video editing operation is to be performed when the first parameter and a second parameter related to the first parameter are both in an unedited state;
and when the first parameter and the second parameter related to the first parameter are both in an editing state, displaying second prompt information, wherein the second prompt information is used for prompting that the video editing operation cannot be responded immediately due to editing conflict.
9. The method of claim 8, wherein the step of determining the position of the first electrode is performed,
the first parameter and the second parameter satisfy at least one of the following associated conditions:
the first parameter and the second parameter are edited into the video editing script template by the same account;
the lens picture corresponding to the first parameter is the same as the lens picture corresponding to the second parameter;
the first parameter is of the same type as the second parameter.
10. The method of claim 1, wherein the video editing operation further comprises a parameter editing operation, wherein after displaying parameters adapted to the plurality of shot pictures in the video editing script template, the method further comprises:
In response to the parameter editing operation for the plurality of shot pictures, replacing parameters adapted to the plurality of shot pictures on the video editing script template with parameters set by the parameter editing operation;
the generating the preview video based on the at least one shot picture and the parameters corresponding to the at least one shot picture includes:
and generating a preview video based on the at least one shot picture and the replaced parameters corresponding to the at least one shot picture.
11. The method according to claim 1, wherein the performing the parameter identification process on each of the lens frames to obtain parameters adapted to the plurality of lens frames includes:
the following processing is performed for each of the lens pictures:
performing object identification processing on the lens picture, and determining the identified object as a document matched with the lens picture;
wherein the type of object comprises at least one of: scene, person, event.
12. A video editing processing apparatus, comprising:
the display module is used for displaying the video editing script template in the document editing interface;
an editing module for responding to a shot picture editing operation in a video editing operation, and displaying a plurality of shot pictures set by the shot picture editing operation in the video editing script template;
The editing module is further used for carrying out parameter identification processing on each lens picture to obtain parameters matched with the plurality of lens pictures, and displaying the parameters matched with the plurality of lens pictures in the video editing script template;
a generation module for generating a preview video based on at least one shot picture and parameters corresponding to the at least one shot picture in response to a video preview operation received in a video editing process;
and the display module is also used for displaying the preview video.
13. An electronic device, comprising:
a memory for storing computer executable instructions;
a processor for implementing the video editing processing method of any of claims 1 to 11 when executing computer executable instructions stored in said memory.
14. A computer-readable storage medium storing computer-executable instructions that, when executed, perform the video editing processing method of any one of claims 1 to 11.
CN202110371612.1A 2021-04-07 2021-04-07 Video editing processing method and device, electronic equipment and storage medium Active CN113709575B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110371612.1A CN113709575B (en) 2021-04-07 2021-04-07 Video editing processing method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110371612.1A CN113709575B (en) 2021-04-07 2021-04-07 Video editing processing method and device, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN113709575A CN113709575A (en) 2021-11-26
CN113709575B true CN113709575B (en) 2024-04-16

Family

ID=78647961

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110371612.1A Active CN113709575B (en) 2021-04-07 2021-04-07 Video editing processing method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN113709575B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114302174A (en) * 2021-12-31 2022-04-08 上海爱奇艺新媒体科技有限公司 Video editing method and device, computing equipment and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103839562A (en) * 2014-03-17 2014-06-04 杨雅 Video creation system
CN104965816A (en) * 2015-07-22 2015-10-07 网易(杭州)网络有限公司 Editing method and device for data sheet
CN111277905A (en) * 2020-03-09 2020-06-12 新华智云科技有限公司 Online collaborative video editing method and device
CN112422831A (en) * 2020-11-20 2021-02-26 广州太平洋电脑信息咨询有限公司 Video generation method and device, computer equipment and storage medium

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2948949A4 (en) * 2013-01-24 2016-09-21 Telesofia Medical Ltd System and method for flexible video construction

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103839562A (en) * 2014-03-17 2014-06-04 杨雅 Video creation system
CN104965816A (en) * 2015-07-22 2015-10-07 网易(杭州)网络有限公司 Editing method and device for data sheet
CN111277905A (en) * 2020-03-09 2020-06-12 新华智云科技有限公司 Online collaborative video editing method and device
CN112422831A (en) * 2020-11-20 2021-02-26 广州太平洋电脑信息咨询有限公司 Video generation method and device, computer equipment and storage medium

Also Published As

Publication number Publication date
CN113709575A (en) 2021-11-26

Similar Documents

Publication Publication Date Title
CN111260545B (en) Method and device for generating image
CN111294663B (en) Bullet screen processing method and device, electronic equipment and computer readable storage medium
US11321667B2 (en) System and method to extract and enrich slide presentations from multimodal content through cognitive computing
CN111914523B (en) Multimedia processing method and device based on artificial intelligence and electronic equipment
WO2022057722A1 (en) Program trial method, system and apparatus, device and medium
CN104735468A (en) Method and system for synthesizing images into new video based on semantic analysis
CN115082602B (en) Method for generating digital person, training method, training device, training equipment and training medium for model
CN109300179A (en) Animation method, device, terminal and medium
US20180143741A1 (en) Intelligent graphical feature generation for user content
CN114339285B (en) Knowledge point processing method, video processing method, device and electronic equipment
CN113709575B (en) Video editing processing method and device, electronic equipment and storage medium
CN113207039B (en) Video processing method and device, electronic equipment and storage medium
KR102040392B1 (en) Method for providing augmented reality contents service based on cloud
US11532111B1 (en) Systems and methods for generating comic books from video and images
CN113268232B (en) Page skin generation method and device and computer readable storage medium
CN114443022A (en) Method for generating page building block and electronic equipment
Shim et al. CAMEO-camera, audio and motion with emotion orchestration for immersive cinematography
CN113010129A (en) Virtual studio full-flow multi-terminal blackboard writing extraction method and device
CN113806596B (en) Operation data management method and related device
JP7153052B2 (en) Online Picture Book Content Acquisition Method, Apparatus, and Smart Screen Device
CN114722136B (en) System and method for customizing and displaying interaction of webpage text map
CN116366909B (en) Virtual article processing method and device, electronic equipment and storage medium
CN117435193A (en) Page generation method and device, electronic equipment and storage medium
JP2022524563A (en) Short video generation methods, platforms, electronics, storage media and programs
CN116980718A (en) Scenario recomposition method and device for video, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant