CN117528175A - Video processing method, device, equipment and medium - Google Patents

Video processing method, device, equipment and medium Download PDF

Info

Publication number
CN117528175A
CN117528175A CN202210910797.3A CN202210910797A CN117528175A CN 117528175 A CN117528175 A CN 117528175A CN 202210910797 A CN202210910797 A CN 202210910797A CN 117528175 A CN117528175 A CN 117528175A
Authority
CN
China
Prior art keywords
text
segment
mirror
sub
segments
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210910797.3A
Other languages
Chinese (zh)
Inventor
黄雅勤
曹翔
周天鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Bilibili Technology Co Ltd
Original Assignee
Shanghai Bilibili Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Bilibili Technology Co Ltd filed Critical Shanghai Bilibili Technology Co Ltd
Priority to CN202210910797.3A priority Critical patent/CN117528175A/en
Publication of CN117528175A publication Critical patent/CN117528175A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44016Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving splicing one content stream with another content stream, e.g. for substituting a video clip

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

The disclosure provides a video processing method, a video processing device, electronic equipment, a computer readable storage medium and a computer program product, and relates to the technical field of video processing. The implementation scheme is as follows: acquiring a plurality of sub-lens segments of a video to be processed and text segments corresponding to each sub-lens segment in the plurality of sub-lens segments; in response to receiving a first editing request for one of the sub-mirror segments, the sub-mirror segments are edited, and text segments corresponding to the sub-mirror segments are edited based on the edited sub-mirror segments.

Description

Video processing method, device, equipment and medium
Technical Field
The present disclosure relates to the field of video processing technologies, and in particular, to a video processing method, apparatus, device, medium, and program.
Background
In video processing, editing of both video content and text corresponding to the video content is often involved. For video content, it may be created, modified, and deleted by, for example, video editing software. And for the corresponding text content, it can be created, modified and deleted by, for example, text editing software.
The approaches described in this section are not necessarily approaches that have been previously conceived or pursued. Unless otherwise indicated, it should not be assumed that any of the approaches described in this section qualify as prior art merely by virtue of their inclusion in this section. Similarly, the problems mentioned in this section should not be considered as having been recognized in any prior art unless otherwise indicated.
Disclosure of Invention
The present disclosure provides a video processing method, apparatus, electronic device, computer readable storage medium, and computer program product.
According to an aspect of the present disclosure, there is provided a video processing method including: acquiring a plurality of sub-lens segments of a video to be processed and text segments corresponding to each sub-lens segment in the plurality of sub-lens segments; in response to receiving a first editing request for one of the sub-mirror segments, the sub-mirror segments are edited, and text segments corresponding to the sub-mirror segments are edited based on the edited sub-mirror segments.
According to an aspect of the present disclosure, there is provided a video processing apparatus including: an acquisition unit configured to acquire a plurality of minute mirror segments of a video to be processed and a text segment corresponding to each of the plurality of minute mirror segments; and an editing unit configured to edit the sub-mirror segments in response to receiving the first editing request for one of the sub-mirror segments, and edit the text segments corresponding to the sub-mirror segments based on the edited sub-mirror segments.
According to another aspect of the present disclosure, there is also provided an electronic apparatus including: at least one processor; and a memory communicatively coupled to the at least one processor; in which a memory stores a computer program which, when executed by at least one processor, implements a method according to the above.
According to another aspect of the present disclosure, there is also provided a non-transitory computer readable storage medium storing a computer program, wherein the computer program, when executed by a processor, implements a method according to the above.
According to another aspect of the present disclosure, there is also provided a computer program product comprising a computer program, wherein the computer program, when executed by a processor, implements a method according to the above.
According to one or more embodiments of the present disclosure, by associating a sub-mirror segment in a video form with a text segment in a text form, the problem of independence between editing of text content and editing of video content is solved, so that a user can rapidly associate a text segment with a sub-mirror segment to view and edit sub-mirror files and text files more clearly. Therefore, the video editing efficiency is improved, the legibility of the file in the editing process is improved, and the learning and cognition cost of the user is reduced.
It should be understood that the description in this section is not intended to identify key or critical features of the embodiments of the disclosure, nor is it intended to be used to limit the scope of the disclosure. Other features of the present disclosure will become apparent from the following specification.
Drawings
The accompanying drawings illustrate exemplary embodiments and, together with the description, serve to explain exemplary implementations of the embodiments. The illustrated embodiments are for exemplary purposes only and do not limit the scope of the claims. Throughout the drawings, identical reference numerals designate similar, but not necessarily identical, elements.
FIG. 1 shows a flow chart of a video processing method according to an embodiment of the present disclosure;
FIG. 2 shows a schematic diagram of a structural partitioning of video to be processed according to an embodiment of the present disclosure;
FIG. 3 shows a flow chart of a video processing method according to an embodiment of the present disclosure;
FIG. 4 illustrates a schematic diagram of a user interface for editing a section of a split mirror in a split mirror mode, according to an embodiment of the present disclosure;
FIG. 5 shows a schematic diagram of a user interface editing a text segment in text mode, according to an embodiment of the present disclosure;
fig. 6 shows a block diagram of a video processing apparatus according to an embodiment of the present disclosure;
fig. 7 shows a block diagram of an electronic device according to an embodiment of the disclosure.
Detailed Description
Exemplary embodiments of the present disclosure are described below in conjunction with the accompanying drawings, which include various details of the embodiments of the present disclosure to facilitate understanding, and should be considered as merely exemplary. Accordingly, one of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope of the present disclosure. Also, descriptions of well-known functions and constructions are omitted in the following description for clarity and conciseness.
In the present disclosure, the use of the terms "first," "second," and the like to describe various elements is not intended to limit the positional relationship, timing relationship, or importance relationship of the elements, unless otherwise indicated, and such terms are merely used to distinguish one element from another. In some examples, a first element and a second element may refer to the same instance of the element, and in some cases, they may also refer to different instances based on the description of the context.
The terminology used in the description of the various examples in this disclosure is for the purpose of describing particular examples only and is not intended to be limiting. Unless the context clearly indicates otherwise, the elements may be one or more if the number of the elements is not specifically limited. Furthermore, the term "and/or" as used in this disclosure encompasses any and all possible combinations of the listed items.
The inventors have noted that text content is required to describe or illustrate the video in the process of editing the video, for example, description of information such as scenes, tasks, types of mirrors, etc. shown in the video, or as scenario information, story-taking, etc. associated with the video content. The editing of the text content and the editing of the video content are independent from each other, and cannot form a one-to-one correspondence. For example, when a user edits, adds, and deletes video content, it is necessary to search text content to locate the position of the related text content and manually modify the corresponding text content. This results in the video editing process becoming cumbersome on the one hand and also in the clear presentation of the structure of the video and sharing on the other hand.
In order to solve the above technical problems, according to one or more embodiments of the present disclosure, a video processing method is proposed.
The method enables the sub-mirror segments in the video form to be in one-to-one correspondence with the text segments in the text form, for example, the sub-mirror segments can be respectively corresponding to the same structural division, so that the problem of mutual independence between editing of text content and editing of video content is solved, and a user can rapidly correspond the text segments with the sub-mirror segments so as to more clearly view and edit the text segments and the sub-mirror segments. In addition, the method further enables the user to clearly recognize the corresponding relation between the sub-mirror segment being edited and the text segment through defining two editing modes, namely the text mode and the sub-mirror mode, and through switching between the two modes, so that the cognitive cost of the user is further reduced.
In general, the video processing method provided by the disclosure realizes that the split-lens segments in the video form and the text segments in the text form are in one-to-one correspondence, and simultaneously, the split-lens segments and the text segments can be rapidly checked and edited, so that the video editing efficiency is improved, the readability of the file in the editing process is improved, and the learning and cognition cost of a user is reduced.
Fig. 1 shows a flowchart of a video processing method 100 according to an embodiment of the present disclosure.
As shown in fig. 1, the method includes: step S102, acquiring a plurality of sub-lens segments of a video to be processed and text segments corresponding to each sub-lens segment in the sub-lens segments; step S104, in response to receiving a first editing request for one of the sub-lens segments, editing the sub-lens segment, and editing a text segment corresponding to the sub-lens segment based on the edited sub-lens segment.
By establishing a one-to-one correspondence between the sub-mirror segments and the text segments, the intercommunication between the sub-mirror segments based on video content and the text segments based on text content can be realized. Editing of the video content (e.g., editing of the split mirror segments in step S104) may be synchronized into the corresponding text segments, thereby improving video editing efficiency. Meanwhile, the corresponding relation is more beneficial to the display of the sub-mirror fragments and the text fragments. The user can clearly know the information of each sub-mirror segment, and quickly locate and identify the sub-mirror segment and the text segment which are wanted to be edited or checked.
Wherein the corresponding sub-mirror segments and text segments may be represented as the same structural division. The form of the structural partitioning in some embodiments will be described in more detail below in conjunction with FIG. 2.
Fig. 2 shows a schematic diagram of a structural partitioning 200 of a video to be processed according to an embodiment of the present disclosure.
In this embodiment, the structural division 200 of the video to be processed includes: story level 201, which story level 201 includes story 1, story 1 may correspond to the entire video of the video to be processed, or may correspond to a portion of the video to be processed. A scene level 202, the scene level 202 comprising a plurality of scenes, e.g., scene 1, scene 2, scene 3, which together comprise story 1. In some embodiments, different scenes may correspond to different locations, or may correspond to different times at the same location. A shot level 203 comprising a plurality of shots, e.g. shot 1, shot 2, shot 3, which together constitute scene 1. In some embodiments, the different shots are divided by shot cuts in the video.
According to some embodiments, the video to be processed includes at least one scene, each of the at least one scene including at least one shot, each of the plurality of sub-mirror segments corresponding to one shot. In some embodiments, each text segment in the text file may also correspond to a shot in shot level 203. The text segment and the split lens segment that correspond to each other correspond to the same shot. Thus, a scene/shot structure division (also referred to as a "draft structure") of the video to be processed can be defined, which refines the structure framework in the video story. The sub-mirror segments and the text segments can be effectively corresponding to the same structural division, so that the sub-mirror segments and the text segments are convenient for users to browse and edit.
In some embodiments, the structural partition (draft structure) is a storable video script file that a user can store, browse, edit, and share with other users.
It will be appreciated that scene 2 and scene 3 shown in fig. 2 may also include one or more shots. And story level 201 in the structural partition 200 may include more stories, scene level 202 may include more scenes, and shot level 203 may include more shots. In practical use, the user may modify the number of stories, scenes, shots, as desired by himself, the specific number is not limited herein.
According to some embodiments, the text segment includes at least one of lens text description information of a sub-lens segment corresponding to the text segment, information of a lens corresponding to the sub-lens segment, and information of a scene including the lens. The information of the lens may include shooting technique information, for example: distant, medium, close-up, etc., and may also include summaries of related objects and/or content in the corresponding sub-mirror segments, such as "character dance", "audience appreciation", etc. The information of the scene may include the location involved in the corresponding one or more sub-mirror segments, e.g. "stage", and may also include the time involved, the object, the action or event occurring in the corresponding one or more sub-mirror segments, etc. And the shot text description information may include a text description of a corresponding one of the sub-mirror segments, such as a scenario description, transcript information, or any text information generated from video content. Therefore, text fragments of multiple levels such as a lens level, a scene level and the like can be determined according to the content related to each lens segment, a multi-level structured text file is realized, and a user can conveniently and quickly find and determine the text fragment corresponding to each lens segment.
Fig. 3 shows a flowchart of a video processing method 100 according to an embodiment of the present disclosure, wherein the first editing request includes a modification request, step S104, in response to receiving the first editing request for one of the sub-mirror segments, editing the sub-mirror segment, and based on the edited sub-mirror segment, editing the text segment corresponding to the sub-mirror segment further includes: step S1041, modifying a sub-mirror segment in response to receiving a modification request for one of the sub-mirror segments; step S1042, obtaining a sub-mirror picture in the modified sub-mirror segment; step S1043, obtaining a preset database, wherein the preset database comprises preset picture data and preset text data corresponding to the preset picture data; step S1044, calculating the matching degree between the preset picture data in the preset database and the sub-mirror picture; and step S1045, modifying the text segment corresponding to the split lens segment based on the preset text data corresponding to the preset picture data with the match degree of the split lens picture meeting the preset standard.
When a split lens segment is modified, the modified split lens segment may relate to a different scene, shot, or even story than the split lens segment before modification. In step S1042, a sub-mirror picture in the modified sub-mirror segment is obtained, and the sub-mirror picture may be the first frame picture in the sub-mirror segment or the picture with the most discrimination in the sub-mirror segment. The preset picture data in the preset database in step S1043 may be picture data composed of a plurality of video materials such as movie materials, animation materials, image materials, etc., or may be other types of preset picture data. The preset text data may be shot information, scene information, shot text description information, etc. of the preset picture data. By calculating the matching degree of the preset picture data and the sub-mirror picture in step S1044, the preset picture data with the highest matching degree with the sub-mirror picture or the preset picture data with the matching degree with the sub-mirror picture exceeding the threshold value can be found. In step S1045, preset text data corresponding to the preset picture data satisfying the preset criteria may be obtained, and added to the text segment corresponding to the split lens segment, or the original text segment may be replaced, or various options may be provided for the user to modify the original text segment. Thus, synchronous modification between the split lens segment and the text segment can be achieved. When a certain sub-mirror segment is modified, the corresponding text segment does not need to be searched again and modified, and therefore the efficiency of the video editing process is improved.
It is described above that the editing of the sub-mirror segment and the text segment (e.g., step S104 and step S106 in the method 100) can be performed in response to receiving the first editing request and the second editing request, respectively. According to some embodiments, the method 100 involves two editing modes: a mirror mode and a text mode. Wherein the first edit request for one of the segments is an edit request in the segment mirror mode and the second edit request for one of the text segments is an edit request in the text mode, the method 100 may further include: in response to receiving the switch request, switching between the mirror mode and the text mode. Thus, free switching of both the split-lens segment and the text segment modes can be achieved, maximizing the utilization of display resources of the device implementing method 100, for example, and improving video editing efficiency.
The minute mirror mode and the text mode according to the embodiment of the present disclosure will be described in detail with reference to fig. 4 and 5.
Fig. 4 shows a schematic diagram of a user interface 400 for editing a split-mirror segment in a split-mirror mode, according to an embodiment of the present disclosure.
The user interface 400 includes: a first switching component 401 for receiving a switching request, a first component 402 for presenting a currently selected sub-mirror segment, and components (e.g., a second component 403a, a third component 403 b) presenting a plurality of sub-mirror segments.
As shown in fig. 4, components exhibiting multiple split lens segments may be displayed in groupings of scenes. Wherein the split lens segments shown by the second component 403a and the third component 403b belong to a first scene (e.g., "28 dance-sole" in fig. 4). In addition, there are also components of a plurality of other sub-mirror segments in the user interface 400 that exhibit sub-mirror segments belonging to a second scene (e.g., "29 stage-principal angle" in FIG. 4) and a third scene (e.g., "30 dance-independent" in FIG. 4). It will be appreciated that the user interface 400 may also include more or fewer scenes, stories, or other structural divisions, and that each scene, story, or other structural division may include more or fewer sub-mirror segments, the specific number of which is not limited herein.
Fig. 5 shows a schematic diagram of a user interface 500 for editing a text passage in text mode according to an embodiment of the present disclosure.
The user interface 500 includes: a second switching component 501 for receiving a switching request, a fourth component 502 for presenting a sub-mirror segment corresponding to a currently selected text segment, and components (e.g., a fifth component 503a, a sixth component 503 b) for presenting a plurality of text segments.
As shown in fig. 5, a component that presents multiple text segments may be displayed in groups of scenes. Wherein the text segments shown by the fifth component 503a and the sixth component 503b belong to a first scene (e.g., "dancing-single" in fig. 5). In addition, there are also components in the user interface 500 of a plurality of other text segments whose presented text segments belong to a second scene (e.g., "stage-principal angle" in fig. 5). It is to be appreciated that the user interface 500 can also include more or fewer scenes, stories, or other structural divisions, as can more or fewer pieces of text per scene, story, or other structural division, the specific number not being limited herein.
According to some embodiments, in response to receiving the switch request, switching between the mirror mode and the text mode includes: in response to receiving a switch request in the split mirror mode: switching to a text mode; selecting a text segment corresponding to the selected split lens segment in the split lens mode; in response to receiving the switch request in text mode: switching to a split mirror mode; and selecting a sub-mirror segment corresponding to the selected text segment in the text mode.
Specifically, for example, in user interface 400 in a split mirror mode, second component 403a is in a selected state, showing a split lens segment that is a selected split lens segment. At this point, the specific content of the split lens segment may be presented in the first component 402, e.g., playing the split lens segment, etc. When the first switching component 401 receives the switching request, it switches from the mirror mode to the text mode, for example, from the user interface 400 in the mirror mode to the user interface 500 in the text mode. At this time, the fifth component 503a is in the selected state, and the text segment shown by the fifth component 503a corresponds to the minute mirror segment shown by the second component 403 a. Meanwhile, in the fourth component 502, the specific content of the split mirror segment displayed by the second component 403a is displayed.
Similarly, for example, in the user interface 500 in text mode, the fifth component 503a is in a selected state, which presents a text segment that is a selected text segment. At this time, the specific content of the split lens segment corresponding to the text segment displayed by the fifth component 503a may be displayed in the fourth component 502, for example, the split lens segment is played. When the second switching component 501 receives the switching request, it switches from the text mode to the mirror mode, e.g., from the user interface 500 in the text mode to the user interface 400 in the mirror mode. At this time, the second component 403a is in the selected state, and the split lens segment shown by the second component 403a corresponds to the text segment shown by the fifth component 503 a. Meanwhile, the specific content of the split mirror segment displayed by the second component 403a is displayed in the first component 402.
Therefore, the video editing method can be freely switched between two display and editing modes, and the selected states are synchronized between the two modes, so that a user can be helped to quickly position the currently-viewed sub-mirror segment and the corresponding text segment, and the overall efficiency of the video editing process is improved.
According to some embodiments, the mirror segments are shown in a mirror mode. The presentation of the sub-mirror segments may include a static presentation (e.g., a thumbnail) or a dynamic presentation (e.g., a dynamic presentation of a portion of the sub-segments of the sub-mirror segments) of the plurality of sub-mirror segments. The presentation of the split lens segments may also include a specific dynamic presentation (e.g., a dynamic presentation of the entire split lens segment) for some of the split lens segments (e.g., selected split lens segments). According to some embodiments, a text segment and a mirror segment corresponding to a currently selected text segment are presented in text mode. The display of the sub-mirror segment corresponding to the currently selected text segment may be a static display, a dynamic display, or a specific dynamic display of the sub-mirror segment. Therefore, the sub-mirror segments of the selected object (sub-mirror segments or text segments) can be displayed in two modes, so that a user can clearly locate the currently selected sub-mirror segments, the learning and cognition cost of the user is reduced, and the editing efficiency is improved.
According to some embodiments, in the split mirror mode, at least a portion of the text segment corresponding to the split lens segment is also shown. At least a part of the text segment may be, for example, scene information, shot information, or the like, or may be a feature tag of the split lens segment. At least one part of the text segment can help a user to understand key information of the split lens segment or summary of corresponding information and the like, and efficiency of editing the video by the user is improved.
Returning to fig. 1, the editing request for a minute section and the editing request for a text section may also include more types of editing requests.
According to some embodiments, the edit request for the mirror segment (e.g., the first edit request in step S104) includes a delete request. Step S104, in response to receiving the first editing request for one of the sub-mirror segments, editing the sub-mirror segment, and based on the edited sub-mirror segment, editing the text segment corresponding to the sub-mirror segment includes: deleting the sub-mirror segments in response to receiving a delete request for one of the sub-mirror segments; and deleting the text segment corresponding to the split lens segment. When a portion of the sub-lens segments of the plurality of sub-lens segments are deemed unnecessary to be deleted, the text segment corresponding to the sub-lens segment will also lose association. At this time, also deleting the corresponding text segment can simplify the video editing flow. The user does not have to delete the corresponding text segment again, but only needs one deletion operation.
According to some embodiments, the edit request for the split mirror segment (e.g., the first edit request in step S104) includes an add request. Step S104, in response to receiving the first editing request for one of the sub-mirror segments, editing the sub-mirror segment, and based on the edited sub-mirror segment, editing the text segment corresponding to the sub-mirror segment includes: adding a split lens segment in response to receiving an addition request for one of the split lens segments; and adding text segments corresponding to the split lens segments. When it is desired to add a new segment to the plurality of segments, a text segment corresponding to the newly added segment should also be generated correspondingly. At this time, the corresponding text segment is directly added, so that the video editing process can be simplified. The user does not have to manually locate the position of the newly added split-mirror segment in the plurality of split-mirror segments, nor retrieve the position of the corresponding text segment to be added in the text segment, but only needs one addition operation of the split-mirror segment. This also makes the addition process of the score mirror segments and text segments more straightforward and efficient.
The text segments corresponding to the newly added split lens segments may be created by preset rules (e.g., creating a default initial text segment, an initial text segment according to multiple preset types, etc.). According to some embodiments, adding text segments corresponding to the split lens segments includes: obtaining a sub-mirror picture in the sub-mirror segment; acquiring a preset database, wherein the preset database comprises preset picture data and preset text data corresponding to the preset picture data; calculating the matching degree of preset picture data and the sub-mirror pictures in a preset database; and adding preset text data corresponding to preset picture data with the matching degree of the split lens pictures meeting preset standards as text fragments corresponding to the split lens segments.
Similar to what is described above in connection with fig. 3, the mirror image may be the first frame of the picture in the mirror segment, or the most differentiated picture in the mirror segment. The preset picture data in the preset database may be picture data composed of a plurality of video materials such as movie materials, animation materials, image materials, and the like, or may be other types of preset picture data. The preset text data may be shot information, scene information, shot text description information, etc. of the preset picture data. The preset standard may be preset picture data with the highest matching degree with the mirror image, preset picture data with the matching degree exceeding a threshold value with the mirror image, or the like. Thus, a corresponding at least base text segment may be created for the newly added segment, which may reflect the characteristics of the newly added segment. The user can distinguish between different sub-mirror segments via the added text segment and locate the different sub-mirror segments.
As described above in connection with fig. 4, text segments may also be edited (e.g., in text mode) and the split lens segments modified accordingly to simplify the editing process. Specifically, according to some embodiments, in response to receiving a second editing request for a plurality of text segments corresponding to a plurality of split lens segments, editing the plurality of text segments; and editing the plurality of sub-mirror segments based on the edited plurality of text segments. The method can simplify the operation quantity required by the user in the video editing process and improve the video editing efficiency.
Specifically, according to some embodiments, the second editing request for the text segment includes a delete request including an identification of the text segment to be deleted. Responsive to receiving a second editing request for the plurality of text segments, editing the plurality of text segments includes: deleting the text segment corresponding to the identifier in response to receiving a delete request for the plurality of text segments; and editing the plurality of split lens segments based on the edited plurality of text segments comprises: in response to identifying that the corresponding text segment is deleted, the sub-mirror segment corresponding to the text segment is deleted. When a part of the text segments of the plurality of text segments is considered to be no longer needed to be deleted, the partial mirror segments corresponding to the text segments will also lose the association relationship. At this time, deleting the corresponding sub-mirror segment can simplify the video editing process. The user needs only one deletion operation.
Further, according to some embodiments, the second editing request for the text segment includes an add request including the text segment to be inserted and an insertion position for indicating a position of inserting the text segment to be inserted in the plurality of text segments. Responsive to receiving a second editing request for the plurality of text segments, editing the plurality of text segments includes: in response to receiving an add request for a plurality of text segments, adding a text segment to be inserted at an insertion location; and editing the plurality of split lens segments based on the edited plurality of text segments comprises: an editable blank sub-lens segment is added to a position corresponding to the insertion position in the plurality of sub-lens segments. When a user wishes to create a new text segment, the user also often wishes to create a new sub-mirror segment to correspond to the text segment. Therefore, a blank split lens segment corresponding to the text segment can be directly added to simplify the editing flow. The blank sub-lens segment is an editable sub-lens segment into which the user can add video content that he or she wishes to add. In some embodiments, features (e.g., tags) in the newly added text segment may be identified, and based on the features, a feature-matching sub-mirror segment may be found from a preset sub-mirror segment library to replace the blank sub-mirror segment as a corresponding sub-mirror segment.
Fig. 6 shows a block diagram of a video processing apparatus 600 according to an embodiment of the disclosure, the apparatus 600 may include: an acquiring unit 601 configured to acquire a plurality of minute mirror segments of a video to be processed and a text segment corresponding to each of the plurality of minute mirror segments; an editing unit 602, configured to edit a minute lens segment in response to receiving a first editing request for one of the minute lens segments, and edit a text segment corresponding to the minute lens segment based on the edited minute lens segment.
The video processing apparatus 600 may be adapted to perform similar operations as the video processing methods 100, 300 described above, and will not be described in detail herein.
According to another aspect of the present disclosure, there is also provided an electronic apparatus including: at least one processor; and a memory communicatively coupled to the at least one processor; in which a memory stores a computer program which, when executed by at least one processor, implements a method according to the above.
According to another aspect of the present disclosure, there is also provided a non-transitory computer readable storage medium storing a computer program, wherein the computer program, when executed by a processor, implements a method according to the above.
According to another aspect of the present disclosure, there is also provided a computer program product comprising a computer program, wherein the computer program, when executed by a processor, implements a method according to the above.
Referring to fig. 7, a block diagram of an electronic device 700 that may be a server or a client of the present disclosure, which is an example of a hardware device that may be applied to aspects of the present disclosure, will now be described. The electronic devices may be different types of computer devices, such as laptop computers, desktop computers, workstations, personal digital assistants, servers, blade servers, mainframe computers, and other suitable computers. The electronic device may also represent various forms of mobile devices, such as personal digital processing, cellular telephones, smartphones, wearable devices, and other similar computing devices. The components shown herein, their connections and relationships, and their functions, are meant to be exemplary only, and are not meant to limit implementations of the disclosure described and/or claimed herein.
As shown in fig. 7, the electronic device 700 may include at least one processor 701, a working memory 702, an input unit 704, a display unit 705, a speaker 706, a storage unit 707, a communication unit 708, and other output units 709 that can communicate with each other through a system bus 703.
The processor 701 may be a single processing unit or multiple processing units, all of which may include a single or multiple computing units or multiple cores. The processor 701 may be implemented as one or more microprocessors, microcomputers, microcontrollers, digital signal processors, central processing units, state machines, logic circuitries, and/or any devices that manipulate signals based on operational instructions. The processor 701 may be configured to obtain and execute computer readable instructions stored in the working memory 702, the storage unit 707, or other computer readable media, such as program code of the operating system 702a, program code of the application programs 702b, and the like.
The working memory 702 and the storage unit 707 are examples of computer-readable storage media for storing instructions that are executed by the processor 701 to implement the various functions described previously. The working memory 702 may include both volatile memory and nonvolatile memory (e.g., RAM, ROM, etc.). In addition, the storage unit 707 may include hard drives, solid state drives, removable media, including external and removable drives, memory cards, flash memory, floppy disks, optical disks (e.g., CD, DVD), storage arrays, network attached storage, storage area networks, and the like. The working memory 702 and the storage unit 707 may both be referred to herein collectively as memory or computer-readable storage medium, and may be non-transitory media capable of storing computer-readable, processor-executable program instructions as computer program code that may be executed by the processor 701 as a particular machine configured to implement the operations and functions described in the examples herein.
The input unit 706 may be any type of device capable of inputting information to the electronic device 700, the input unit 706 may receive input numeric or character information and generate key signal inputs related to user settings and/or function control of the electronic device, and may include, but is not limited to, a mouse, a keyboard, a touch screen, a trackpad, a trackball, a joystick, a microphone, and/or a remote control. The output unit may be any type of device capable of presenting information and may include, but is not limited to, a display unit 705, a speaker 706, and other output units 709, which other output units 709 may include, but are not limited to, a video/audio output terminal, a vibrator, and/or a printer. The communication unit 708 allows the electronic device 700 to exchange information/data with other devices via a computer network, such as the internet, and/or various telecommunications networks, and may include, but is not limited to, modems, network cards, infrared communication devices, wireless communication transceivers and/or chipsets, such as bluetooth (TM) devices, 1302.11 devices, wiFi devices, wiMax devices, cellular communication devices, and/or the like.
The application 702b in the working register 702 may be loaded to perform the various methods and processes described above, e.g., step S201-step S203 in fig. 2. For example, in some embodiments, the sample adaptive compensation method may be implemented as a computer software program tangibly embodied on a machine-readable medium, such as the storage unit 707. In some embodiments, part or all of the computer program may be loaded and/or installed onto the electronic device 700 via the storage unit 707 and/or the communication unit 708. When the computer program is loaded and executed by the processor 701, one or more steps of the sample adaptive compensation method described above may be performed. Alternatively, in other embodiments, the processor 701 may be configured to perform the sample adaptive compensation method in any other suitable manner (e.g., by means of firmware).
Various implementations of the systems and techniques described here above may be implemented in digital electronic circuitry, integrated circuit systems, field Programmable Gate Arrays (FPGAs), application Specific Integrated Circuits (ASICs), application Specific Standard Products (ASSPs), systems On Chip (SOCs), load programmable logic devices (CPLDs), computer hardware, firmware, software, and/or combinations thereof. These various embodiments may include: implemented in one or more computer programs, the one or more computer programs may be executed and/or interpreted on a programmable system including at least one programmable processor, which may be a special purpose or general-purpose programmable processor, that may receive data and instructions from, and transmit data and instructions to, a storage system, at least one input device, and at least one output device.
Program code for carrying out methods of the present disclosure may be written in any combination of one or more programming languages. These program code may be provided to a processor or controller of a general purpose computer, special purpose computer, or other programmable data processing apparatus such that the program code, when executed by the processor or controller, causes the functions/operations specified in the flowchart and/or block diagram to be implemented. The program code may execute entirely on the machine, partly on the machine, as a stand-alone software package, partly on the machine and partly on a remote machine or entirely on the remote machine or server.
In the context of this disclosure, a machine-readable medium may be a tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. The machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. The machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a machine-readable storage medium would include an electrical connection based on one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
To provide for interaction with a user, the systems and techniques described here can be implemented on a computer having: a display device (e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) for displaying information to a user; and a keyboard and pointing device (e.g., a mouse or trackball) by which a user can provide input to the computer. Other kinds of devices may also be used to provide for interaction with a user; for example, feedback provided to the user may be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and input from the user may be received in any form, including acoustic input, speech input, or tactile input.
The systems and techniques described here can be implemented in a computing system that includes a background component (e.g., as a data server), or that includes a middleware component (e.g., an application server), or that includes a front-end component (e.g., a user computer having a graphical user interface or a web browser through which a user can interact with an implementation of the systems and techniques described here), or any combination of such background, middleware, or front-end components. The components of the system can be interconnected by any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include: local Area Networks (LANs), wide Area Networks (WANs), and the internet.
The computer system may include a client and a server. The client and server are typically remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
It should be appreciated that various forms of the flows shown above may be used to reorder, add, or delete steps. For example, the steps recited in the present disclosure may be performed in parallel, sequentially or in a different order, provided that the desired results of the disclosed aspects are achieved, and are not limited herein.
Although embodiments or examples of the present disclosure have been described with reference to the accompanying drawings, it is to be understood that the foregoing methods, systems, and apparatus are merely exemplary embodiments or examples, and that the scope of the present invention is not limited by these embodiments or examples but only by the claims following the grant and their equivalents. Various elements of the embodiments or examples may be omitted or replaced with equivalent elements thereof. Furthermore, the steps may be performed in a different order than described in the present disclosure. Further, various elements of the embodiments or examples may be combined in various ways. It is important that as technology evolves, many of the elements described herein may be replaced by equivalent elements that appear after the disclosure.

Claims (18)

1. A video processing method, comprising:
acquiring a plurality of sub-lens segments of a video to be processed and text segments corresponding to each sub-lens segment in the plurality of sub-lens segments;
in response to receiving a first editing request for one of the sub-mirror segments, the sub-mirror segment is edited, and a text segment corresponding to the sub-mirror segment is edited based on the edited sub-mirror segment.
2. The method of claim 1, wherein the video to be processed comprises at least one scene, each of the at least one scene comprising at least one shot, each of the plurality of sub-lens segments corresponding to a shot.
3. The method of claim 2, wherein the text segment includes at least one of lens text description information of a sub-mirror segment corresponding to the text segment, information of a lens corresponding to the sub-mirror segment, and information of a scene including the lens.
4. A method according to any one of claims 1 to 3, wherein the first editing request comprises a modification request, editing one of the sub-mirror segments in response to receiving the first editing request for the sub-mirror segment, and editing the respective text segment of the sub-mirror segment based on the edited sub-mirror segment comprises:
modifying one of the sub-mirror segments in response to receiving the modification request for the sub-mirror segment;
obtaining a sub-mirror picture in the modified sub-mirror segment;
acquiring a preset database, wherein the preset database comprises preset picture data and preset text data corresponding to the preset picture data;
calculating the matching degree of the preset picture data and the sub-mirror picture in the preset database; and
and modifying the text segment corresponding to the split lens segment based on preset text data corresponding to preset picture data of which the matching degree of the split lens picture meets a preset standard.
5. The method of any of claims 1-4, wherein the first edit request for one of the mirror segments is an edit request in mirror mode and the second edit request for one of the text segments is an edit request in text mode, the method further comprising:
in response to receiving a switch request, switching between the mirror mode and the text mode.
6. The method of claim 5, wherein switching between the mirror mode and the text mode in response to receiving the switch request comprises:
in response to receiving the switch request in the split mirror mode:
switching to the text mode;
selecting a text segment corresponding to the selected sub-lens segment in the sub-lens mode;
in response to receiving the switch request in the text mode:
switching to the split mirror mode;
and selecting a sub-mirror segment corresponding to the selected text segment in the text mode.
7. The method of claim 5, further comprising:
in the split mirror mode, displaying the split mirror segments; and
and in the text mode, displaying the text segment and the sub-mirror segment corresponding to the currently selected text segment.
8. The method of claim 7, wherein in the split mirror mode, at least a portion of a text segment corresponding to the split mirror segment is also displayed.
9. The method of any of claims 1-8, wherein the first editing request comprises a delete request, editing one of the split lens segments in response to receiving the first editing request for the split lens segment, and editing the respective text segment of the split lens segment based on the edited split lens segment comprises:
deleting one of the sub-mirror segments in response to receiving the delete request for the sub-mirror segment; and
and deleting the text segment corresponding to the split lens segment.
10. The method of any of claims 1-9, wherein the first editing request comprises an add request, editing one of the split lens segments in response to receiving the first editing request for the split lens segment, and editing the respective text segment of the split lens segment based on the edited split lens segment comprises:
adding a split mirror segment in response to receiving the addition request for one of the split mirror segments; and
And adding text fragments corresponding to the split lens segments.
11. The method of claim 10, wherein adding text segments corresponding to the split lens segments comprises: obtaining a sub-mirror picture in the sub-mirror segment;
acquiring a preset database, wherein the preset database comprises preset picture data and preset text data corresponding to the preset picture data;
calculating the matching degree of the preset picture data and the sub-mirror picture in the preset database;
and adding preset text data corresponding to preset picture data with the matching degree of the split lens pictures meeting a preset standard as text fragments corresponding to the split lens fragments.
12. The method of any of claims 1-11, further comprising:
responsive to receiving a second editing request for a plurality of text segments corresponding to the plurality of split lens segments, editing the plurality of text segments; and
and editing the plurality of sub-mirror fragments based on the edited plurality of text fragments.
13. The method of claim 12, wherein the second editing request comprises a delete request including an identification of a text segment to be deleted, and editing the plurality of text segments in response to receiving the second editing request for the plurality of text segments comprises:
Deleting the text segment corresponding to the identification in response to receiving the deletion request for the plurality of text segments;
and editing the plurality of split lens segments based on the plurality of edited text segments comprises:
and deleting the sub-mirror segments corresponding to the text segments in response to the text segments corresponding to the identifications being deleted.
14. The method of any of claims 12-13, wherein the second editing request comprises an add request comprising a text segment to be inserted and an insertion location for indicating a location of inserting the text segment to be inserted in the plurality of text segments,
responsive to receiving a second editing request for the plurality of text segments, editing the plurality of text segments includes:
in response to receiving the add request for the plurality of text segments, adding the text segment to be inserted at the insertion location;
and editing the plurality of split lens segments based on the plurality of edited text segments comprises:
and adding an editable blank sub-lens segment at a position corresponding to the insertion position in the plurality of sub-lens segments.
15. A video processing apparatus comprising:
an acquisition unit configured to acquire a plurality of minute mirror segments of a video to be processed and a text segment corresponding to each minute mirror segment of the plurality of minute mirror segments;
and an editing unit configured to edit one of the sub-mirror segments in response to receiving a first editing request for the sub-mirror segment, and edit a text segment corresponding to the sub-mirror segment based on the edited sub-mirror segment.
16. An electronic device, comprising:
at least one processor; and
a memory communicatively coupled to the at least one processor; wherein the method comprises the steps of
The memory stores a computer program which, when executed by the at least one processor, implements the method according to any one of claims 1-14.
17. A non-transitory computer readable storage medium storing a computer program, wherein the computer program when executed by a processor implements the method according to any one of claims 1-14.
18. A computer program product comprising a computer program, wherein the computer program, when executed by a processor, implements the method according to any of claims 1-14.
CN202210910797.3A 2022-07-29 2022-07-29 Video processing method, device, equipment and medium Pending CN117528175A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210910797.3A CN117528175A (en) 2022-07-29 2022-07-29 Video processing method, device, equipment and medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210910797.3A CN117528175A (en) 2022-07-29 2022-07-29 Video processing method, device, equipment and medium

Publications (1)

Publication Number Publication Date
CN117528175A true CN117528175A (en) 2024-02-06

Family

ID=89761325

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210910797.3A Pending CN117528175A (en) 2022-07-29 2022-07-29 Video processing method, device, equipment and medium

Country Status (1)

Country Link
CN (1) CN117528175A (en)

Similar Documents

Publication Publication Date Title
US20210294855A1 (en) Method for processing information, electronic device and storage medium
US20220417417A1 (en) Content Operation Method and Device, Terminal, and Storage Medium
US10978108B2 (en) Apparatus, method, and program for creating a video work
CN111263186A (en) Video generation, playing, searching and processing method, device and storage medium
US10356359B2 (en) Information processing apparatus, method for controlling the information processing apparatus, and recording medium
CN111309200B (en) Method, device, equipment and storage medium for determining extended reading content
US20220284377A1 (en) Method and apparatus for task group positioning
CN112000911A (en) Page management method and device, electronic equipment and storage medium
CN111382289A (en) Picture display method and device, computer equipment and storage medium
CN112256645A (en) Data processing method, device, terminal and storage medium
CN109697242B (en) Photographing question searching method and device, storage medium and computing equipment
CN109116718B (en) Method and device for setting alarm clock
CN112559913B (en) Data processing method, device, computing equipment and readable storage medium
CN105868168A (en) A document operation method and a terminal
CN105204718B (en) Information processing method and electronic equipment
CN117528175A (en) Video processing method, device, equipment and medium
CN108255486B (en) View conversion method and device for form design and electronic equipment
CN113542888B (en) Video processing method and device, electronic equipment and storage medium
CN114090060A (en) Webpage configuration method and device, terminal equipment and readable storage medium
US11303464B2 (en) Associating content items with images captured of meeting content
CN110413603B (en) Method and device for determining repeated data, electronic equipment and computer storage medium
JP2015146105A (en) Display control device, operation method of display control device, and computer program
CN109445757B (en) New design drawing generation method and device and terminal equipment
CN113706209B (en) Operation data processing method and related device
CN111459887B (en) Resource screening method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination