CN114222165A - Video playing method, device, equipment and computer storage medium - Google Patents

Video playing method, device, equipment and computer storage medium Download PDF

Info

Publication number
CN114222165A
CN114222165A CN202111677497.7A CN202111677497A CN114222165A CN 114222165 A CN114222165 A CN 114222165A CN 202111677497 A CN202111677497 A CN 202111677497A CN 114222165 A CN114222165 A CN 114222165A
Authority
CN
China
Prior art keywords
video
action
processed
playback
determining
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111677497.7A
Other languages
Chinese (zh)
Other versions
CN114222165B (en
Inventor
李荣生
李晓耕
陶嘉伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Mobile Communications Group Co Ltd
MIGU Video Technology Co Ltd
MIGU Culture Technology Co Ltd
Original Assignee
China Mobile Communications Group Co Ltd
MIGU Video Technology Co Ltd
MIGU Culture Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Mobile Communications Group Co Ltd, MIGU Video Technology Co Ltd, MIGU Culture Technology Co Ltd filed Critical China Mobile Communications Group Co Ltd
Priority to CN202111677497.7A priority Critical patent/CN114222165B/en
Publication of CN114222165A publication Critical patent/CN114222165A/en
Application granted granted Critical
Publication of CN114222165B publication Critical patent/CN114222165B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234381Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by altering the temporal resolution, e.g. decreasing the frame rate by frame skipping
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/432Content retrieval operation from a local storage medium, e.g. hard-disk
    • H04N21/4325Content retrieval operation from a local storage medium, e.g. hard-disk by playing back content from the storage medium
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • H04N21/440281Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display by altering the temporal resolution, e.g. by frame skipping
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The embodiment of the invention relates to the technical field of multimedia playing and discloses a method, which comprises the following steps: acquiring a video to be processed; the video to be processed comprises a plurality of actions; identifying a video to be processed to obtain action types of each action; respectively determining the playback duration of each action according to the action type; processing each action according to the playback duration to obtain a processed video; and playing the processed video. Through the mode, the video watching experience of the user is improved.

Description

Video playing method, device, equipment and computer storage medium
Technical Field
The embodiment of the invention relates to the technical field of multimedia playing, in particular to a video playing method, a video playing device, video playing equipment and a computer storage medium.
Background
In video playing, especially for videos with fast picture change speed, users often have a need for slow motion playback. To achieve slow motion playback of a video, the video may be frame-interpolated.
The inventor finds that the current video frame interpolation method has the problem of poor video watching experience of a user in the process of implementing the embodiment of the application.
Disclosure of Invention
In view of the foregoing problems, embodiments of the present invention provide a video playing method, apparatus, device and computer storage medium, which are used to solve the problem in the prior art that the video viewing experience of a user is not good.
According to an aspect of an embodiment of the present invention, there is provided a video playing method, including:
acquiring a video to be processed; the video to be processed comprises a plurality of actions;
identifying the video to be processed to obtain the action type of each action;
respectively determining the playback duration of each action according to the action type;
processing each action according to the playback duration to obtain a processed video;
and playing the processed video.
In an optional mode, the video to be processed comprises a plurality of video frames; one said action corresponds to at least one said video frame; the method further comprises the following steps:
extracting the characteristics of the video to be processed to obtain the image characteristics corresponding to each video frame;
clustering the video frames according to the image characteristics to obtain a clustering result;
and determining each action and the action type corresponding to the action according to the clustering result.
In an optional manner, the method further comprises:
determining type characteristic information of each action type;
determining information entropy corresponding to each action type according to all the type characteristic information;
and determining the playback time length according to the information entropy.
In an alternative mode, the type feature information includes at least one of an action complexity, a popularity and an action difficulty of the action type; the method further comprises the following steps:
respectively determining a playing weight value corresponding to each action type according to at least one of action complexity, popularity and action difficulty of the action type;
and respectively determining the information entropy corresponding to each action type according to the playing weight value.
In an optional manner, the method further comprises:
determining the playback proportion of each action type according to the information entropy;
determining the total playback time length corresponding to the video to be processed;
and determining the playback time length according to the playback proportion and the total playback time length.
In an optional manner, the method further comprises:
determining timestamp information and current playing parameters of each action;
determining frame insertion information corresponding to each action according to the playback duration, the timestamp information and the current playing parameter;
and carrying out frame interpolation processing on each action according to the frame interpolation information to obtain the processed video. .
In an optional manner, the method further comprises:
marking a frame interval of the action subjected to frame interpolation processing in the video to be processed to obtain a marked segment;
and when the operation aiming at the marked segment is detected, playing the marked segment.
According to another aspect of the embodiments of the present invention, there is provided a video playing apparatus, including:
the acquisition module is used for acquiring a video to be processed; the video to be processed comprises a plurality of actions;
the identification module is used for identifying the video to be processed to obtain the action type of each action;
the determining module is used for respectively determining the playback duration of each action according to the action type;
the processing module is used for respectively processing each action according to the playback time length to obtain a processed video;
and the playing module is used for playing the processed video.
According to another aspect of the embodiments of the present invention, there is provided a video playback device including: the system comprises a processor, a memory, a communication interface and a communication bus, wherein the processor, the memory and the communication interface complete mutual communication through the communication bus;
the memory is used for storing at least one executable instruction, and the executable instruction causes the processor to execute the operation of the video playing method.
According to another aspect of the embodiments of the present invention, there is provided a computer-readable storage medium having at least one executable instruction stored therein, where the executable instruction causes a video playing device to execute the operations of the video playing method.
The embodiment of the invention obtains the video to be processed; the video to be processed comprises a plurality of actions; identifying a video to be processed to obtain action types of each action; respectively determining the playback duration of each action according to the action type; processing each action according to the playback duration to obtain a processed video; and finally, playing the processed video, wherein the problem that the playing effect is poor due to the fact that overall playback is adopted in the prior art is solved, the playback duration can be determined according to the action type of the action included in the video to be processed, so that the playback with different durations can be adaptively performed according to the characteristics of different actions, a user can obtain the action detail information of which the number is matched with the action type, and the video watching experience of the user can be improved.
The foregoing description is only an overview of the technical solutions of the embodiments of the present invention, and the embodiments of the present invention can be implemented according to the content of the description in order to make the technical means of the embodiments of the present invention more clearly understood, and the detailed description of the present invention is provided below in order to make the foregoing and other objects, features, and advantages of the embodiments of the present invention more clearly understandable.
Drawings
The drawings are only for purposes of illustrating embodiments and are not to be construed as limiting the invention. Also, like reference numerals are used to refer to like parts throughout the drawings. In the drawings:
fig. 1 is a schematic flow chart illustrating a video playing method according to an embodiment of the present invention;
fig. 2 is a schematic diagram illustrating a mark of a to-be-processed frame interval according to an embodiment of the present invention;
FIG. 3 is a schematic diagram illustrating the playing of a processed video according to an embodiment of the present invention;
fig. 4 is a schematic structural diagram of a video playing apparatus according to an embodiment of the present invention;
fig. 5 shows a schematic structural diagram of a video playing device according to an embodiment of the present invention.
Detailed Description
Exemplary embodiments of the present invention will be described in more detail below with reference to the accompanying drawings. While exemplary embodiments of the invention are shown in the drawings, it should be understood that the invention can be embodied in various forms and should not be limited to the embodiments set forth herein.
Fig. 1 shows a flowchart of a video playing method provided by an embodiment of the present invention, which is executed by a computer processing device. The computer processing device may include a cell phone, a notebook computer, etc. As shown in fig. 1, the method comprises the steps of:
step 10: acquiring a video to be processed; the video to be processed comprises a plurality of actions.
In an embodiment of the present invention, the video to be processed may include one or more moving objects, and the multiple actions may correspond to the same moving object or different moving objects. Optionally, the moving object has high-speed motion in the video to be processed, such as a player in a competitive sports or a ball in a ball game.
In yet another embodiment of the present invention, the video to be processed may also be a highlight video segment obtained by pre-screening, such as a video segment that is pre-screened and clipped from the video stream according to the highlight or the association degree of the action. The action accuracy can be determined according to the continuity of the action, the action difficulty, the completion degree of the action and the like, the association degree can be the association degree of the video clip and the currently played scene information, and the currently played scene information comprises the currently played video information, the currently played playing time, the currently played geographic position, the user characteristic information and the like.
Step 20: and identifying the video to be processed to obtain the action type of each action.
In one embodiment of the invention, a video to be processed is divided into a plurality of continuous video frames, image feature recognition is carried out on each video frame, and clustering analysis is carried out according to recognition results to obtain action types corresponding to each video frame. The action type may be one of selectable types that are pre-marked with respect to a video type of the video to be processed, and the video type may be a motion type included in the video or a type of a motion target. When the type of motion as a video is a ski game, the selectable types may include skating, take-off, air flight, and landing. For video types with table tennis as the sport target, the optional types include landing, landing on a table, landing on a racket, net wiping, air sports, and the like.
In an embodiment of the present invention, the video to be processed includes a plurality of video frames; one said action corresponds to at least one said video frame; step 20 further comprises:
step 201: and identifying the video to be processed to obtain the image characteristics corresponding to each video frame.
In an embodiment of the present invention, image feature extraction is performed on each video frame in a video to be processed, and the extraction manner may be through a convolutional neural network, where the convolutional neural network includes a convolutional layer, a pooling layer, and a connection layer. The image features may include RGB features of the video frame, and grayscale features, among others.
Step 202: and clustering the video frames according to the image characteristics to obtain a clustering result.
In an embodiment of the invention, clustering can be performed according to a K-means algorithm, c groups are preset, corresponding to c action types, and action types corresponding to all video frames are obtained as clustering results by clustering the video frames under the corresponding groups.
Step 203: and determining each action and the action type corresponding to the action according to the clustering result.
In an embodiment of the present invention, video frames with the same motion type and occurring continuously may be determined as a frame interval corresponding to one motion, so as to obtain each motion, and then the motion type corresponding to each motion is determined according to the motion type of the video frame in the frame interval.
In yet another embodiment of the present invention, the video to be processed may be detected according to a preset motion detection algorithm to obtain each motion and a corresponding video frame, and then the motion type of the video frame with the largest proportion is determined as the motion type of the motion according to the proportion of the motion type of the corresponding video frame.
Step 30: and respectively determining the playback time length of each action according to the action type.
In one embodiment of the invention, the playback duration is used for representing the degree of slow motion of the motion, and the longer the playback duration is, the more time the user enjoys the motion is, so that more detailed information of the motion can be acquired. Considering that the viewing interestingness of the user for different types of actions is different, for actions with more complex actions, more difficult actions or more wonderful actions, the user tends to see more details and watch for a longer time, and therefore, the play weight value of each action type can be determined according to the type feature information of the action type, and the play weight value is used for representing the proportion of the playback time corresponding to the action of the type in the total playback time.
When determining the playing weight value according to the type feature information, an information entropy corresponding to the action type can be determined according to the type feature information, and then the playing weight value is determined according to the information entropy, wherein the information entropy is used for representing information that can be provided by the action type, specifically, if the information source symbol has n values: u1 … Ui … Un, the corresponding probability is: p1 … Pi … Pn, and the various symbols occur independently of each other. The average uncertainty of the source should be the statistical average (E) of the single symbol uncertainties-logPi, which can be referred to as the information entropy. The greater the entropy of the information, the more information is provided, the greater the uncertainty that can be eliminated, and thus the greater the viewing weight of the action type for the user.
Thus, in a further embodiment of the present invention, step 30 further comprises: step 301: determining type feature information of each action type.
In an embodiment of the present invention, a plurality of evaluation dimensions corresponding to an action type may be used as corresponding feature dimensions, so as to obtain type feature information, where the type feature information includes at least one of action complexity, popularity, and action difficulty of the action type. The evaluation dimensions may include motion complexity, motion attractiveness, and motion difficulty.
The motion complexity is used for representing the number of frames required for completing the motion, and the more the number of frames required by the motion is, the more steps for representing the motion are, the more details are contained, and the more the number of frames needs to be distributed for slow playing. The action attractiveness characterizes the user's interest level in the action, with a higher interest level indicating a greater desire of the user to see more detail and playback of the action. The action difficulty represents the difficulty level of the implementation of the action, and the more difficult action is, the more visible the action is, the enough number of frames should be allocated for slow playing.
In yet another embodiment of the present invention, the play weight values corresponding to each action type can be obtained by performing weighted summation on the dimension values under multiple evaluation dimensions.
Step 302: and determining the information entropy corresponding to each action type according to all the type characteristic information.
In an embodiment of the present invention, if the source symbol has n values, according to the following definition of the information entropy: u1 … Ui … Un, the corresponding probability is: p1 … Pi … Pn, and the various symbols occur independently of each other. The average uncertainty of the source should be the statistical average (E) of the single symbol uncertainties-logPi, which can be referred to as the information entropy.
Specifically, the information entropy of each action type can be determined according to the dimension values of all the action types under each type feature dimension and the number of actions under each action type. Wherein the probability of each action type is determined according to the proportion of the number of actions to the total number of all actions.
In yet another embodiment of the present invention, step 302 further comprises:
step 3021: and respectively determining the playing weight value corresponding to each action type according to at least one item of action complexity, popularity and action difficulty of the action type.
In one embodiment of the invention, the sum of at least one characteristic dimension of the action complexity, the popularity and the action difficulty is carried out according to all the action types, and then the playing weight value is determined according to the proportion of the dimension value of each action type in the sum of the dimension values under the characteristic dimension of at least one item.
Step 3022: and respectively determining the information entropy corresponding to each action type according to the playing weight value.
In an embodiment of the present invention, according to the definition of the information entropy, an action type is determined as a value of an information source symbol, a play weight value is used as a corresponding probability of the information source symbol corresponding to the action type, and the information entropy corresponding to each action type is obtained by calculation.
Step 303: and determining the playback time length according to the information entropy.
In one embodiment of the invention, the playback proportion is determined according to the information entropy, and the playback time length corresponding to each action is determined according to the playback proportion and the total playback time length corresponding to the video to be processed. Wherein the determining of the playback ratio according to the information entropy may be adjusting on a reference value of the underlying playback ratio according to the information entropy. And determining the playback time length corresponding to each action according to the product of the playback proportion corresponding to each action type and the total playback time length.
Thus, in yet another embodiment of the present invention, step 303 further comprises:
step 3031: and determining the playback proportion of each action type according to the information entropy.
In an embodiment of the present invention, a reference value of the playback scale may be set, and the playback scale of each action type may be obtained by adjusting the reference value according to the ratio of the information entropy of each action type.
Step 3032: and determining the total playback time length corresponding to the video to be processed.
In one embodiment of the present invention, the total playback time period may be adaptively determined according to the length of the video to be processed.
Step 3033: and determining the playback time length according to the playback proportion and the total playback time length.
In one embodiment of the invention, the playback time period may be determined by the following formula:
corresponding information entropy H (U) to different playback time tnAnd calculating the playback time corresponding to different types of actions. Setting the classifiable action type as N, corresponding different information entropies H (U) to different action types, calculating the information entropy H (U) N for each action value, setting the information entropies corresponding to all the classes of actions as H (U) all,
Figure BDA0003452552840000081
let the total playback time of all action types be T, then the playback time T corresponding to different actionsnIs composed of
Figure BDA0003452552840000082
In yet another embodiment of the present invention, the relationship between the playback time length and the frame insertion information may be determined as follows:
tn=Sn*(fen-fsn)/FPS;
wherein, tnPlayback duration of action n, SnFor playback scale of action n, fsnIs the starting frame number of action n, fenThe end frame number of action n and FPS is the current frame rate.
Step 40: and processing each action according to the playback time length to obtain a processed video.
In an embodiment of the present invention, frame insertion processing may be performed on each action, and the playing time length of each action is adjusted to be the playback time length, wherein, during the frame insertion processing, frame insertion information corresponding to each action is determined according to the playback time length, and the current playing parameter and the timestamp information of each action.
Specifically, the frame interpolation information refers to parameters for performing frame interpolation processing on the video to be processed, and may include, for example, a frame interval to be processed, a total number of frame interpolation frames, and a number of frame interpolation frames corresponding to each action. The target frame rate is used for adjusting the number of video frames included in the motion, and the larger the target frame rate is, the more picture frames can be received by the user in unit time, so that the user can enjoy more details of the motion.
Thus, in a further embodiment of the present invention, step 40 further comprises: step 401: and determining the timestamp information and the current playing parameter of each action.
In one embodiment of the invention, the timestamp information of the action comprises the sequence number of the corresponding start frame and the sequence number of the end frame of the action in the video to be processed. Specifically, the timestamp information may be determined according to a maximum value and a minimum value of all frame numbers corresponding to each action. The current playing parameters comprise the current frame rate of each action. The current frame rate may be a current playing frame rate of the video to be processed.
Step 402: and determining frame insertion information corresponding to each action according to the playback duration, the timestamp information and the current playing parameter.
In an embodiment of the present invention, the frame insertion information includes a frame insertion number corresponding to each action, and is determined according to the following formula:
hn=tn*Pn-(fen-fsn);
wherein h isnFor the number of inserted frames, P, corresponding to action nnTarget frame rate for action n, fsnIs the starting frame number of action n, fenThe ending frame number for action n.
For example, there are three actions M1, M2, and M3, the frame number intervals are (2000,2050), (2051, 2075), (2076, 2100), the corresponding action types are C1, C2, and C3, the corresponding playback ratios are S, respectively1=3,S2=2.5,S32, the target frame rates are respectively P1=50,P2=30,P 330, the corresponding current frame rate is FPS 25, and according to the foregoing steps, the playback time lengths of M1, M2, and M3 are 6s, 2.5s, and 2s, respectively, and the corresponding numbers of frame insertions are 250, 50, and 35, respectively.
Step 403: and carrying out frame interpolation processing on each action according to the frame interpolation information to obtain the processed video.
In an embodiment of the present invention, frame interpolation processing is performed on a video to be processed according to the number of frame interpolation, so that the time length of a video segment corresponding to each action after the frame interpolation is the corresponding playback time length.
In view of the fact that the playing of the video is blocked during slow playing and the viewing experience of the user is affected by adopting a global slow playing processing mode for the video to be processed in the prior art, in yet another embodiment of the invention, the video to be processed can be subjected to wonderful frame screening at first, the screened video is subjected to slow playing processing, and the frame interpolation information further comprises a frame interval to be processed; the action is located in the pending frame interval. Before step 403, the method further comprises:
step 4031: and performing wonderful action recognition on the video to be processed to obtain the frame interval to be processed.
In one embodiment of the invention, image features of a video to be processed are extracted, moving target identification is carried out according to the extracted image features, action fineness is determined according to action feature information of the moving target, and a frame interval to be processed is determined from the video to be processed according to the action fineness of each action. The image characteristics may include RGB characteristics of each video frame, grayscale characteristics, displacement characteristics of a pixel point, and the like, and the motion characteristic information may include information such as a motion type, a motion height, a motion speed, a rotation number, and the like.
Step 50: and playing the processed video.
In an embodiment of the invention, in order to remind a user that a to-be-processed video includes an action which can be slowly played after frame interpolation processing, the position of a to-be-processed frame interval for representing a highlight action can be marked, and a corresponding control is generated, so that the user can display the slowly played video corresponding to the action when operating the control.
Thus, in a further embodiment of the present invention, step 50 further comprises:
step 501: and marking the frame interval of the action subjected to the frame interpolation processing in the video to be processed to obtain a marked segment.
In an embodiment of the present invention, the marking manner may be to mark a frame interval in which the action after the frame insertion processing is located on a time axis, or generate a corresponding control to be displayed in the frame interval in which the action after the frame insertion processing is located, so as to prompt a user to click and trigger playback of the action. Wherein, the generated control can refer to fig. 2. The control may be a display that is started when it is detected that the video to be processed starts playing on the user side.
Step 502: and when the operation aiming at the marked segment is detected, playing the marked segment.
In one embodiment of the invention, the detected operation on the marked segment can be that the user clicks the time axis of the marked segment or clicks a function key corresponding to a playback function in a control displayed on the marked segment. Wherein the playing manner of the marked segment may be in the form of a small window as shown in fig. 3.
The video playing method provided by the embodiment of the invention obtains the video to be processed; the video to be processed comprises a plurality of actions; identifying a video to be processed to obtain action types of each action; respectively determining the playback duration of each action according to the action type; processing each action according to the playback duration to obtain a processed video; and finally, playing the processed video, wherein the problem that the playing effect is poor due to the fact that overall playback is adopted in the prior art is different from the problem that the playing effect is poor during playback in the prior art, the video playing method provided by the embodiment of the invention can determine the playback time according to the action type of the action included in the video to be processed, so that the playback in different time lengths can be carried out according to the characteristics of different actions, a user can obtain action detail information matched with the action type, and the video watching experience of the user can be improved.
Fig. 4 shows a schematic structural diagram of a video playing apparatus according to an embodiment of the present invention. As shown in fig. 4, the apparatus 600 includes: an acquisition module 601, an identification module 602, a determination module 603, a processing module 604, and a playing module 605.
The acquiring module 601 is configured to acquire a video to be processed; the video to be processed comprises a plurality of actions;
the identification module 602 is configured to identify the video to be processed to obtain an action type of each action;
a determining module 603, configured to determine playback durations of the actions according to the action types;
the processing module 604 is configured to process each action according to the playback time length to obtain a processed video;
a playing module 605, configured to play the processed video.
In an optional mode, the video to be processed comprises a plurality of video frames; one said action corresponds to at least one said video frame; the identification module 602 is further configured to:
identifying the video to be processed to obtain image characteristics corresponding to each video frame;
clustering the video frames according to the image characteristics to obtain a clustering result;
determining each action and the action type corresponding to the action according to the clustering result;
and determining the timestamp information of each action according to the frame number of the video frame corresponding to each action.
In an optional manner, the determining module 603 is further configured to:
determining type characteristic information of each action type;
determining information entropy corresponding to each action type according to all the type characteristic information;
and determining the playback time length according to the information entropy.
In an alternative mode, the type feature information includes at least one of an action complexity, a popularity and an action difficulty of the action type; the determining module 603 is further configured to:
respectively determining a playing weight value corresponding to each action type according to at least one of action complexity, popularity and action difficulty of the action type;
and respectively determining the information entropy corresponding to each action type according to the playing weight value.
In an optional manner, the determining module 603 is further configured to:
determining the playback proportion of each action type according to the information entropy;
determining the total playback time length corresponding to the video to be processed;
and determining the playback time length according to the playback proportion and the total playback time length.
In an alternative manner, the identification module 602 is further configured to:
determining timestamp information and current playing parameters of each action;
determining frame insertion information corresponding to each action according to the playback duration, the timestamp information and the current playing parameter;
and carrying out frame interpolation processing on each action according to the frame interpolation information to obtain the processed video.
In an optional manner, the playing module 605 is further configured to:
marking a frame interval of the action subjected to frame interpolation processing in the video to be processed to obtain a marked segment;
and when the operation aiming at the marked segment is detected, playing the marked segment.
The video playing device provided by the embodiment of the invention obtains the video to be processed; the video to be processed comprises a plurality of actions; identifying a video to be processed to obtain action types of each action; respectively determining the playback duration of each action according to the action type; processing each action according to the playback duration to obtain a processed video; and finally, playing the processed video, wherein the problem that the playing effect is poor due to the fact that overall playback is adopted in the prior art is solved, the video playing device provided by the embodiment of the invention can determine the playback time according to the action type of the action included in the video to be processed, so that the playback in different time lengths can be carried out according to the characteristics of different actions, a user can obtain action detail information matched with the action type, and the video watching experience of the user can be improved.
Fig. 5 is a schematic structural diagram of a video playing device according to an embodiment of the present invention, and the specific embodiment of the present invention does not limit the specific implementation of the video playing device.
As shown in fig. 5, the video playback device may include: a processor (processor)702, a Communications Interface 704, a memory 706, and a communication bus 708.
Wherein: the processor 702, communication interface 704, and memory 706 communicate with each other via a communication bus 708. A communication interface 704 for communicating with network elements of other devices, such as clients or other servers. The processor 702 is configured to execute the program 710, and may specifically execute the relevant steps in the above-described embodiment of the video playing method.
In particular, the program 710 may include program code comprising computer-executable instructions.
The processor 702 may be a central processing unit CPU, or an Application Specific Integrated Circuit (ASIC), or one or more Integrated circuits configured to implement an embodiment of the present invention. The video playing device comprises one or more processors, which can be the same type of processor, such as one or more CPUs; or may be different types of processors such as one or more CPUs and one or more ASICs.
The memory 706 stores a program 710. The memory 706 may comprise high-speed RAM memory, and may also include non-volatile memory (non-volatile memory), such as at least one disk memory.
Specifically, the program 710 can be invoked by the processor 702 to cause the video playback device to perform the following operations:
acquiring a video to be processed; the video to be processed comprises a plurality of actions;
identifying the video to be processed to obtain the action type of each action;
respectively determining the playback duration of each action according to the action type;
processing each action according to the playback duration to obtain a processed video;
and playing the processed video.
In an optional mode, the video to be processed comprises a plurality of video frames; one said action corresponds to at least one said video frame; the program 710 is invoked by the processor 702 to cause the video playback device to:
extracting the characteristics of the video to be processed to obtain the image characteristics corresponding to each video frame;
clustering the video frames according to the image characteristics to obtain a clustering result;
and determining each action and the action type corresponding to the action according to the clustering result. In an alternative manner, the program 710 is invoked by the processor 702 to cause a video playback device to:
determining type characteristic information of each action type;
determining information entropy corresponding to each action type according to all the type characteristic information;
and determining the playback time length according to the information entropy.
In an alternative mode, the type feature information includes at least one of an action complexity, a popularity and an action difficulty of the action type; the program 710 is invoked by the processor 702 to cause the video playback device to:
respectively determining a playing weight value corresponding to each action type according to at least one of action complexity, popularity and action difficulty of the action type;
and respectively determining the information entropy corresponding to each action type according to the playing weight value.
In an alternative manner, the program 710 is invoked by the processor 702 to cause a video playback device to:
determining the playback proportion of each action type according to the information entropy;
determining the total playback time length corresponding to the video to be processed;
and determining the playback time length according to the playback proportion and the total playback time length.
In an alternative manner, the program 710 is invoked by the processor 702 to cause a video playback device to:
determining timestamp information and current playing parameters of each action;
determining frame insertion information corresponding to each action according to the playback duration, the timestamp information and the current playing parameter;
and carrying out frame interpolation processing on each action according to the frame interpolation information to obtain the processed video. In an alternative manner, the program 710 is invoked by the processor 702 to cause a video playback device to:
marking a frame interval of the action subjected to frame interpolation processing in the video to be processed to obtain a marked segment;
and when the operation aiming at the marked segment is detected, playing the marked segment.
The video playing device provided by the embodiment of the invention obtains the video to be processed; the video to be processed comprises a plurality of actions; identifying a video to be processed to obtain action types of each action; respectively determining the playback duration of each action according to the action type; processing each action according to the playback duration to obtain a processed video; and finally, playing the processed video, wherein the problem that the playing effect is poor due to the fact that overall playback is adopted in the prior art is solved, and the video playing device provided by the embodiment of the invention can determine the playback time length according to the action type of the action included in the video to be processed, so that the playback in different time lengths can be carried out according to the characteristics of different actions, a user can obtain action detail information matched with the action type, and the video watching experience of the user can be improved.
An embodiment of the present invention provides a computer-readable storage medium, where the storage medium stores at least one executable instruction, and when the executable instruction is executed on a video playing device, the video playing device is enabled to execute a video playing method in any method embodiment described above.
The executable instructions may be specifically configured to cause the video playback device to perform the following operations:
acquiring a video to be processed; the video to be processed comprises a plurality of actions;
identifying the video to be processed to obtain the action type of each action;
respectively determining the playback duration of each action according to the action type;
processing each action according to the playback duration to obtain a processed video;
and playing the processed video.
In an optional mode, the video to be processed comprises a plurality of video frames; one said action corresponds to at least one said video frame; the executable instructions cause the video playback device to:
extracting the characteristics of the video to be processed to obtain the image characteristics corresponding to each video frame;
clustering the video frames according to the image characteristics to obtain a clustering result;
and determining each action and the action type corresponding to the action according to the clustering result.
In an alternative, the executable instructions cause the video playback device to:
determining type characteristic information of each action type;
determining information entropy corresponding to each action type according to all the type characteristic information;
and determining the playback time length according to the information entropy.
In an alternative mode, the type feature information includes at least one of an action complexity, a popularity and an action difficulty of the action type; the executable instructions cause the video playback device to:
respectively determining a playing weight value corresponding to each action type according to at least one of action complexity, popularity and action difficulty of the action type;
and respectively determining the information entropy corresponding to each action type according to the playing weight value.
In an alternative, the executable instructions cause the video playback device to:
determining the playback proportion of each action type according to the information entropy;
determining the total playback time length corresponding to the video to be processed;
and determining the playback time length according to the playback proportion and the total playback time length.
In an alternative, the executable instructions cause the video playback device to:
determining timestamp information and current playing parameters of each action;
determining frame insertion information corresponding to each action according to the playback duration, the timestamp information and the current playing parameter;
and carrying out frame interpolation processing on each action according to the frame interpolation information to obtain the processed video.
In an alternative, the executable instructions cause the video playback device to:
marking a frame interval of the action subjected to frame interpolation processing in the video to be processed to obtain a marked segment;
and when the operation aiming at the marked segment is detected, playing the marked segment.
The computer storage medium provided by the embodiment of the invention obtains the video to be processed; the video to be processed comprises a plurality of actions; identifying a video to be processed to obtain action types of each action; respectively determining the playback duration of each action according to the action type; processing each action according to the playback duration to obtain a processed video; and finally, playing the processed video, wherein the problem that the playing effect is poor due to the fact that overall playback is adopted in the prior art is solved, and the computer storage medium provided by the embodiment of the invention can determine the playback time according to the action type of the action included in the video to be processed, so that the playback with different times can be performed according to the characteristics of different actions, a user can obtain action detail information matched with the action type, and the video watching experience of the user can be improved.
The embodiment of the invention provides a video playing device, which is used for executing the video playing method.
Embodiments of the present invention provide a computer program, where the computer program can be called by a processor to enable a video playing device to execute a video playing method in any of the above method embodiments.
Embodiments of the present invention provide a computer program product, which includes a computer program stored on a computer-readable storage medium, where the computer program includes program instructions, and when the program instructions are run on a computer, the computer is caused to execute a video playing method in any of the above method embodiments.
The algorithms or displays presented herein are not inherently related to any particular computer, virtual system, or other apparatus. Various general purpose systems may also be used with the teachings herein. The required structure for constructing such a system will be apparent from the description above. In addition, embodiments of the present invention are not directed to any particular programming language. It is appreciated that a variety of programming languages may be used to implement the teachings of the present invention as described herein, and any descriptions of specific languages are provided above to disclose the best mode of the invention.
In the description provided herein, numerous specific details are set forth. It is understood, however, that embodiments of the invention may be practiced without these specific details. In some instances, well-known methods, structures and techniques have not been shown in detail in order not to obscure an understanding of this description.
Similarly, it should be appreciated that in the foregoing description of exemplary embodiments of the invention, various features of the embodiments of the invention are sometimes grouped together in a single embodiment, figure, or description thereof for the purpose of streamlining the invention and aiding in the understanding of one or more of the various inventive aspects. However, the disclosed method should not be interpreted as reflecting an intention that: that the invention as claimed requires more features than are expressly recited in each claim.
Those skilled in the art will appreciate that the modules in the device in an embodiment may be adaptively changed and disposed in one or more devices different from the embodiment. The modules or units or components of the embodiments may be combined into one module or unit or component, and may be divided into a plurality of sub-modules or sub-units or sub-components. All of the features disclosed in this specification (including any accompanying claims, abstract and drawings), and all of the processes or elements of any method or apparatus so disclosed, may be combined in any combination, except combinations where at least some of such features and/or processes or elements are mutually exclusive. Each feature disclosed in this specification (including any accompanying claims, abstract and drawings) may be replaced by alternative features serving the same, equivalent or similar purpose, unless expressly stated otherwise.
It should be noted that the above-mentioned embodiments illustrate rather than limit the invention, and that those skilled in the art will be able to design alternative embodiments without departing from the scope of the appended claims. In the claims, any reference signs placed between parentheses shall not be construed as limiting the claim. The word "comprising" does not exclude the presence of elements or steps not listed in a claim. The word "a" or "an" preceding an element does not exclude the presence of a plurality of such elements. The invention may be implemented by means of hardware comprising several distinct elements, and by means of a suitably programmed computer. In the unit claims enumerating several means, several of these means may be embodied by one and the same item of hardware. The usage of the words first, second and third, etcetera do not indicate any ordering. These words may be interpreted as names. The steps in the above embodiments should not be construed as limiting the order of execution unless specified otherwise.

Claims (10)

1. A video playback method, the method comprising:
acquiring a video to be processed; the video to be processed comprises a plurality of actions;
identifying the video to be processed to obtain the action type of each action;
respectively determining the playback duration of each action according to the action type;
processing each action according to the playback duration to obtain a processed video;
and playing the processed video.
2. The method of claim 1, wherein the video to be processed comprises a plurality of video frames; one said action corresponds to at least one said video frame; the identifying the video to be processed to obtain the action type of each action comprises:
extracting the characteristics of the video to be processed to obtain the image characteristics corresponding to each video frame;
clustering the video frames according to the image characteristics to obtain a clustering result;
and determining each action and the action type corresponding to the action according to the clustering result.
3. The method of claim 1, wherein determining the playback duration of the action based on the action type comprises:
determining type characteristic information of each action type;
determining information entropy corresponding to each action type according to all the type characteristic information;
and determining the playback time length according to the information entropy.
4. The method of claim 3, wherein the type feature information comprises at least one of an action complexity, a popularity, and an action difficulty of the action type; the determining the information entropy corresponding to each action type according to all the type feature information includes:
respectively determining a playing weight value corresponding to each action type according to at least one of action complexity, popularity and action difficulty of the action type;
and respectively determining the information entropy corresponding to each action type according to the playing weight value.
5. The method of claim 3, wherein the determining the playback time duration from the information entropy determination comprises:
determining the playback proportion of each action type according to the information entropy;
determining the total playback time length corresponding to the video to be processed;
and determining the playback time length according to the playback proportion and the total playback time length.
6. The method according to claim 3, wherein the processing the video to be processed according to the playback time length to obtain a processed video comprises:
determining timestamp information and current playing parameters of each action;
determining frame insertion information corresponding to each action according to the playback duration, the timestamp information and the current playing parameter;
and carrying out frame interpolation processing on each action according to the frame interpolation information to obtain the processed video.
7. The method of claim 1, wherein said playing said processed video comprises:
marking a frame interval of the action subjected to frame interpolation processing in the video to be processed to obtain a marked segment;
and when the operation aiming at the marked segment is detected, playing the marked segment.
8. A video playback apparatus, comprising:
the acquisition module is used for acquiring a video to be processed; the video to be processed comprises a plurality of actions;
the identification module is used for identifying the video to be processed to obtain the action type of each action;
the determining module is used for respectively determining the playback duration of each action according to the action type;
the processing module is used for respectively processing each action according to the playback time length to obtain a processed video;
and the playing module is used for playing the processed video.
9. A video playback device, comprising: the system comprises a processor, a memory, a communication interface and a communication bus, wherein the processor, the memory and the communication interface complete mutual communication through the communication bus;
the memory is configured to store at least one executable instruction that causes the processor to perform the operations of the video playback method according to any one of claims 1 to 7.
10. A computer-readable storage medium having stored therein at least one executable instruction that, when executed on a video playback device, causes the video playback device to perform operations of the video playback method of any one of claims 1-7.
CN202111677497.7A 2021-12-31 2021-12-31 Video playing method, device, equipment and computer storage medium Active CN114222165B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111677497.7A CN114222165B (en) 2021-12-31 2021-12-31 Video playing method, device, equipment and computer storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111677497.7A CN114222165B (en) 2021-12-31 2021-12-31 Video playing method, device, equipment and computer storage medium

Publications (2)

Publication Number Publication Date
CN114222165A true CN114222165A (en) 2022-03-22
CN114222165B CN114222165B (en) 2023-11-10

Family

ID=80707494

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111677497.7A Active CN114222165B (en) 2021-12-31 2021-12-31 Video playing method, device, equipment and computer storage medium

Country Status (1)

Country Link
CN (1) CN114222165B (en)

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2004091195A1 (en) * 2003-04-11 2004-10-21 Id (Event Marketing) Monaco Sam Method of and apparatus for providing a visual presentation
US20160293209A1 (en) * 2015-03-31 2016-10-06 Xiaomi Inc. Method and device for controlling playback
US9661384B1 (en) * 2016-04-05 2017-05-23 Arris Enterprises Llc Trick play user activity reconstruction
CN109964275A (en) * 2016-10-14 2019-07-02 乐威指南公司 For providing the system and method for slow motion video stream simultaneously with normal speed video flowing when detecting event
CN111147955A (en) * 2019-12-31 2020-05-12 咪咕视讯科技有限公司 Video playing method, server and computer readable storage medium
CN112153462A (en) * 2019-06-26 2020-12-29 腾讯科技(深圳)有限公司 Video processing method, device, terminal and storage medium
CN112422863A (en) * 2019-08-22 2021-02-26 华为技术有限公司 Intelligent video recording method and device
CN113014995A (en) * 2021-02-26 2021-06-22 北京蜂巢世纪科技有限公司 Video playback method, medium, glasses, and program product
CN113515997A (en) * 2020-12-28 2021-10-19 腾讯科技(深圳)有限公司 Video data processing method and device and readable storage medium

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2004091195A1 (en) * 2003-04-11 2004-10-21 Id (Event Marketing) Monaco Sam Method of and apparatus for providing a visual presentation
US20160293209A1 (en) * 2015-03-31 2016-10-06 Xiaomi Inc. Method and device for controlling playback
US9661384B1 (en) * 2016-04-05 2017-05-23 Arris Enterprises Llc Trick play user activity reconstruction
CN109964275A (en) * 2016-10-14 2019-07-02 乐威指南公司 For providing the system and method for slow motion video stream simultaneously with normal speed video flowing when detecting event
CN112153462A (en) * 2019-06-26 2020-12-29 腾讯科技(深圳)有限公司 Video processing method, device, terminal and storage medium
CN112422863A (en) * 2019-08-22 2021-02-26 华为技术有限公司 Intelligent video recording method and device
CN111147955A (en) * 2019-12-31 2020-05-12 咪咕视讯科技有限公司 Video playing method, server and computer readable storage medium
CN113515997A (en) * 2020-12-28 2021-10-19 腾讯科技(深圳)有限公司 Video data processing method and device and readable storage medium
CN113014995A (en) * 2021-02-26 2021-06-22 北京蜂巢世纪科技有限公司 Video playback method, medium, glasses, and program product

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
H. PAN等: ""Detection of slow-motion replay segments in sports video for highlights generation"", 《2001 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING. PROCEEDINGS》 *
周国昌: "MOOC在线学习行为建模与分析", 《中国硕士学位论文全文库》 *
董乐乐: ""基于关键帧提取的视频智能快进技术"", 《中国优秀硕士学位论文全文数据库》 *

Also Published As

Publication number Publication date
CN114222165B (en) 2023-11-10

Similar Documents

Publication Publication Date Title
CN109145840B (en) Video scene classification method, device, equipment and storage medium
US9600717B1 (en) Real-time single-view action recognition based on key pose analysis for sports videos
EP2800057B1 (en) Mobile determination of properties of a trajectory of a ball
CN111988638B (en) Method and device for acquiring spliced video, electronic equipment and storage medium
CN110505519A (en) A kind of video clipping method, electronic equipment and storage medium
CN110225402B (en) Method and device for intelligently keeping interesting target time display in panoramic video
JP4886707B2 (en) Object trajectory identification device, object trajectory identification method, and object trajectory identification program
CN112183588A (en) Video processing method and device, electronic equipment and storage medium
CN114584680A (en) Motion data display method and device, computer equipment and storage medium
CN112837350A (en) Target moving object identification method and device, electronic equipment and storage medium
US11917264B2 (en) Systems and methods for providing video enhancement for sporting events
CN110287934B (en) Object detection method and device, client and server
CN113992976B (en) Video playing method, device, equipment and computer storage medium
CN112150464A (en) Image detection method and device, electronic equipment and storage medium
CN114222165B (en) Video playing method, device, equipment and computer storage medium
CN111476056A (en) Target object identification method and device, terminal equipment and computer storage medium
CN110490064B (en) Sports video data processing method and device, computer equipment and computer storage medium
US20230377335A1 (en) Key person recognition in immersive video
EP4149115A1 (en) Video stream processing method and apparatus, and computer device and medium
CN113971693A (en) Live broadcast picture generation method, system and device and electronic equipment
US11749309B2 (en) Information processor, information processing method, and program
CN108121963B (en) Video data processing method and device and computing equipment
CN113992975B (en) Video playing method, device, equipment and computer storage medium
CN108632639B (en) Video type judgment method and server
Huang et al. A novel algorithm for shuttlecock tracking

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant