CN117061813A - Media playback method and related media playback device - Google Patents

Media playback method and related media playback device Download PDF

Info

Publication number
CN117061813A
CN117061813A CN202210489074.0A CN202210489074A CN117061813A CN 117061813 A CN117061813 A CN 117061813A CN 202210489074 A CN202210489074 A CN 202210489074A CN 117061813 A CN117061813 A CN 117061813A
Authority
CN
China
Prior art keywords
frame
data
index table
playback
intra
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210489074.0A
Other languages
Chinese (zh)
Inventor
陈峻儀
杨景尧
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Realtek Semiconductor Corp
Original Assignee
Realtek Semiconductor Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Realtek Semiconductor Corp filed Critical Realtek Semiconductor Corp
Priority to CN202210489074.0A priority Critical patent/CN117061813A/en
Publication of CN117061813A publication Critical patent/CN117061813A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44008Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • H04N21/4394Processing of audio elementary streams involving operations for analysing the audio stream, e.g. detecting features or characteristics in audio streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47217End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for controlling playback functions for recorded or on-demand content, e.g. using progress bars, mode or play-point indicators or bookmarks

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Human Computer Interaction (AREA)
  • Signal Processing For Digital Recording And Reproducing (AREA)

Abstract

Various embodiments of the present invention relate to a media playback method and related media playback apparatus. A method of media playback on a media playback device, comprising: receiving a data stream, so as to buffer the data stream in a buffer unit or record the data stream in a storage unit; performing an analysis operation on the cached or recorded data stream to obtain frame time and data offset corresponding to one or more intra-frame pictures of at least one audiovisual content contained in the data stream; constructing a frame index table according to the frame time and the data offset corresponding to the one or more intra-frame pictures respectively; responding to a playback operation, and referring to the frame index table to determine a data offset corresponding to the playback operation; acquiring one or more data units in the cache unit or the storage unit according to the data offset; and decoding the one or more data units for media playback.

Description

Media playback method and related media playback device
Technical Field
The present invention relates to media playback, and more particularly, to a media playback method and related media playback apparatus for accelerating playback response by parsing a data stream in advance.
Background
Generally, index information is included in the multimedia package (Multimedia container) that can indicate an association between a specific playback time of the packaged multimedia content and the data structure. However, the multimedia playback apparatus must parse the index information from the start point of the multimedia content until the target playback time is approached, and cannot parse out in which data unit the specific multimedia content corresponding to the target playback time is stored. However, in today's video coding techniques, groups of Pictures (Group of Pictures, GOP) are often based on IBP arrangement architecture (i.e., I-frames, B-frames, and P-frames), and the multimedia playback device cannot fully decode the video content until no data units containing Intra Pictures (i.e., I-frames) are obtained. The multimedia playback device must wait until a data unit containing the complete intra picture is obtained before decoding the video content. This greatly limits the response speed of the multimedia playback apparatus to the user's search, skip, fast forward or rewind, etc. playback operations. In this way, the user is easily made aware of the delay in the transition of the audiovisual content. Accordingly, there is a need in the art for a technique that addresses playback response delays, thereby enhancing user experience.
Disclosure of Invention
In view of the above, the present invention provides a media playing method and a related media playback device. In the invention, after receiving the data stream containing the audio-visual content, the media playback device pre-analyzes the cached data stream, and analyzes the frame time and data offset information corresponding to the intra-frame pictures of the audio-visual content in the data stream. Based on this information, the present invention further constructs a frame index table. After that, if the user wants to play back the audio-visual content, the invention loads the constructed frame index table, and searches the frame index table to quickly determine the storage position of the data unit of the picture in the frame when the user performs specific playback control operation, thereby obtaining the data unit for decoding and playing back. The invention effectively improves the response speed of the media playback device to the playback control operation of the user through the frame index table.
The embodiment of the invention provides a method for playing back media on a media playback device, which comprises the following steps: receiving a data stream on the media playback device to buffer the data stream in a buffer unit in the media playback device or a storage unit recorded in the media playback device; and carrying out an analysis operation on the cached or recorded data stream to obtain frame time and data offset corresponding to one or more intra-frame pictures of at least one audiovisual content contained in the data stream. Constructing a frame index table corresponding to the audio-visual content according to the frame time and the data offset corresponding to the one or more intra-frame pictures respectively; determining a data offset corresponding to a playback operation of the audiovisual content by referring to the frame index table; acquiring one or more data units in the cache unit or the storage unit according to the data offset; and decoding the one or more data units for media playback.
The embodiment of the invention provides a media playback device. The media playback apparatus includes: a receiving and buffering circuit, a parsing and decoding circuit, an index table constructing circuit and a playback control circuit. The receiving and buffering circuit is used for receiving a data stream and buffering the data stream in a buffering unit. The parsing and decoding circuit is coupled to the receiving and buffering circuit, and is configured to perform a parsing operation on the buffered data stream or the data stream recorded in a storage unit, so as to obtain frame time and data offset corresponding to one or more intra pictures of at least one audiovisual content included in the data stream. The index table constructing circuit is coupled to the parsing and decoding circuit and is configured to construct a frame index table corresponding to the audio-visual content according to the frame time and the data offset corresponding to the one or more intra-frame pictures. The playback control circuit is used for responding to a playback operation of the audio-visual content, referring to the frame index table to determine a data offset corresponding to the playback operation, and acquiring one or more data units in the buffer unit or the storage unit according to the data offset. Wherein the parsing and decoding circuit is further configured to decode the one or more data units for media playback.
Drawings
Fig. 1 is a schematic diagram of a media playback apparatus according to an embodiment of the present invention.
Fig. 2 is a block diagram of a receiving and buffering circuit in a media playback apparatus according to an embodiment of the present invention.
Fig. 3A and 3B show a frame index table according to an embodiment of the invention.
Fig. 4 shows the association between the data unit to which the frame belongs and the data offset in an embodiment of the present invention.
Fig. 5 is a flowchart of a media playback method according to an embodiment of the present invention.
FIG. 6 illustrates a device frame pattern for implementing one or more circuit elements or modules in an embodiment of the invention.
Detailed Description
In the following, numerous specific details of embodiments of the invention are described to provide a thorough understanding of the invention. One skilled in the relevant art will recognize, however, how to implement the invention without one or more of the specific details, or with other methods or components or materials, etc. In some instances, well-known structures, materials, or operations are not shown or described in detail to avoid obscuring the core concepts of the invention.
Reference in the specification to "one embodiment" means that a particular feature, structure, or characteristic described in connection with the embodiment may be included in at least one embodiment of the invention. Thus, the appearances of the phrase "in one embodiment" in various places in the specification are not necessarily all referring to the same embodiment. Furthermore, the particular features, structures, or characteristics described above may be combined in any suitable manner in one or more embodiments.
Fig. 1 is a schematic diagram of an embodiment of a media playback apparatus according to the present invention. As shown, the media playback apparatus 100 is configured to receive a data stream from a media source 10 for media playback. The media source may be, among other things, a digital video broadcast (Digital Video Broadcasting, DVB) service provider, an internet protocol television (Internet Protocol Television, IPTV) service provider, or an OTT service (Over-the-top media services) provider, the internet, etc. Also, the media playback apparatus 100 may be wired or wireless through a specific image/sound transmission interface such as: a high-definition multimedia interface (High Definition Multimedia Interface, HDMI) or DisplayPort (DP), etc., outputs video/audio signals to the display device 200 and/or the acoustic device 300, thereby reproducing audio-visual contents through the devices. The media playback apparatus 100 includes: the receiving and buffering circuit 110, the parsing and decoding circuit 120 (including the video parsing and decoding unit 122 and the audio parsing and decoding unit 124), the index table construction circuit 130, the storage unit 140, the playback control circuit 150 and the recording circuit 160.
When a user wants to watch a certain audiovisual content (e.g., music, movies, broadcast programs, etc.), the media playback apparatus 100 obtains the audiovisual content from the media source 10, for example, locks the frequency point to which the audiovisual content belongs (when the media source 10 is a DVB service), or makes a request to a server (when the media source 10 is an IPTV service, OTT service, internet, etc.). The audiovisual content may be transmitted to the media playback device 100 in the form of a data stream. The receive and buffer circuit 110 functions to receive a data stream from the media source 10 and buffer the data stream. Wherein the data stream may be transmitted by the media source 10 to the media playback apparatus 100 in wireless or wired form and through one or more relay devices. In one embodiment, the data stream may include compressed Video content based on Video coding standards such as the moving picture expert group (Moving Picture Experts Group, MPEG) standard, the advanced Video coding (Advanced Video Coding, AVC) standard, high efficiency Video coding (High Efficiency Video Coding, HEVC), or AOMedia Video 1 (AV 1). In addition, the data stream may further include Audio content compressed based on Advanced coding 3 (ac-3), MPEG-1 or MPEG-2Audio Layer III (MPEG-1 Audio Layer III or MPEG-2Audio Layer III, abbreviated as MP 3), advanced Audio coding (Advanced Audio Coding, AAC), MPEG, and other Audio coding standards. It should be noted here that the particular video/audio coding standard employed for the audiovisual content in the data stream is not a limitation of the present invention. Video/audio coding standards other than those described above may be selected under different design requirements.
Fig. 2 is a schematic diagram of a receiving and buffering circuit 110 in a media playback apparatus according to an embodiment of the present invention. The receiving and buffering circuit 110 includes a network receiving device 111, a frequency modulation device 112, a descrambling device 113, a demultiplexer 114 and a buffering unit 115. Wherein, when the media source 10 is an IPTV service, OTT service, internet, or the like, the network receiving device 111 is configured to receive a data stream; and when the media source 10 is a DVB service, the frequency modulation device 112 is configured to receive a data stream. In addition, when a data sheet in the data stream is subjected to scrambling (scrambling), the descrambling means 113 may perform descrambling (unscramble) thereon, thereby obtaining an unscrambled data unit. The demultiplexer 114 is used for demultiplexing the data units in the data stream. The data units processed by the descrambling means 113 and the demultiplexer 114 are buffered in the buffering unit 115. For more details on the principle and architecture change of the receiving and buffering circuit 110, reference is made to the description in taiwan patent application (application number 110100272).
After the data units of the data stream are stored in the buffer unit 115, the parsing and decoding circuit 120 performs parsing (buffering) on the buffered data units in advance. The video parsing and decoding unit 122 and the audio parsing and decoding unit 124 determine the video/audio content included in the data unit and various auxiliary information related to the video/audio content according to the header (header) of the data unit. In addition, if the media playback apparatus 100 receives the data stream and records the data stream in the storage unit 140 through the recording circuit 160, the video parsing and decoding unit 122 and the audio parsing and decoding unit 124 may parse the data unit in the storage unit 140.
After parsing, the relevant information of each audio-visual content in the data stream can be obtained. For example: the Program identification code (Program ID) of the audio-visual content may also be obtained as frame-related information for each audio-visual content, for example: frame type, data offset, encoder format, frame rate, etc. The index table construction circuit 130 constructs the information into a frame index table as shown in fig. 3A and 3B.
As shown in fig. 3A and 3B, the frame index table includes a plurality of fields for recording information such as a frame type, a data offset, a frame time, an encoder format, and a frame rate. The field for recording the frame type further indicates, in addition to the frame type, the program identification code (i.e., video program identification code (VPID) and Audio program identification code (APID)) of the audiovisual content to which the frame belongs. In one embodiment, when the value of the byte used to represent the frame type is 1, this frame is meant to be an Intra picture (Intra pictures), such as an I frame (I frames) in the h.264 coding format or HEVC coding format; when the value of this byte is 2, this means that this frame is a predicted picture (Predicted pictures), such as a P frame (P frames) in h.264 coding format or HEVC coding format; and when the value of this byte is 3, this means that this frame is a Bi-predictive picture (Bi-predictive pictures or Bi-directional pictures), such as a B frame (B frames) in the h.264 coding format or HEVC coding format. Note that in different embodiments of the present invention, the frame type may be represented using values different from those listed in the above embodiments.
In the frame index table, a field in which data offset information is recorded is used to mainly indicate the data offset (offset) of the data unit to which each frame belongs with respect to the data unit to which the starting frame (i.e., frame 1) belongs. For example, the data offset information of frame 3 is 500, representing that the data unit to which frame 3 belongs is located 500 data offset units (e.g., bytes) after the data unit to which frame 1 belongs; the data offset information of frame 5 is 1200, representing that the data unit to which frame 5 belongs is located 1200 bytes after the data unit to which frame 1 belongs. It should be noted that, in the embodiments of the present invention, the data offset may be recorded in data units of different sizes, and although the above description is given by taking bytes as an example, those skilled in the art should have sufficient knowledge to record the data offset information in other data units after knowing the essential content of the present invention. Furthermore, a data unit may contain one or more frames depending on the size of the data unit.
The field for recording frame time information is mainly used to indicate the time interval of each frame relative to the starting frame (i.e., frame 1). For example, the time information of frame 4 is 90, representing that the time interval between frame 4 and frame 1 is 90 time units (e.g., 1/30 seconds); the time information for frame 8 is 210, representing a time interval between frame 8 and frame 1 of 210 time units. Note that in different embodiments of the present invention, the time information may be recorded in different time units. Although the above description is given taking (1/30) second as an example, a person skilled in the art, having knowledge of the substance of the present specification, should have sufficient knowledge to record the time information of a frame in other time units. Such variations are intended to fall within the scope of the present invention.
A field for recording encoder format information may be used to indicate the video encoder format used for each frame and the audio encoder format of the audio content to which it corresponds. For example, the first byte recorded in this field may be used to represent a video encoder format and the second byte may be used to represent an audio encoder format. However, in other embodiments of the present invention, the first byte may be used to represent the audio encoder format and the second byte may be used to represent the video encoder format. In one embodiment, when the value of the byte used to represent the video encoder format is 1, this means that the video encoder is in MPG format; when the value of this byte is 2, it means that the video encoder is in AVC format; when the value of this byte is 3, it means that the video encoder is HEVC or in h.265 format; and when the value of this byte is 4, it means that the video encoder is in AV1 format. Note that in various embodiments of the present invention, the video encoder format may be represented using values different from those listed in the above embodiments.
In an embodiment, when the value of the byte used to represent the audio encoder format is 1, this means that the audio encoder is in AC-3 format; when the value of this byte is 2, this means that the audio encoder is in MP3 format; when the value of this byte is 3, this means that the audio encoder is in AAC format; when the value of this byte is 4, this means that the audio encoder is in MPEG format. Note that in various embodiments of the present invention, the audio encoder format may be represented using values different from those listed in the above embodiments.
The field of the frame rate information is recorded, and is mainly used for representing the frame rate of the audiovisual content to which each frame belongs. For example, frame 1 has a frame rate of 30, representing that frame 1 is one frame of audiovisual content having a frame rate of 30; frame 10 has a frame rate of 24 and represents one frame of audiovisual content having a frame rate of 24 for frame 10. It should be noted that frame 1 may belong to regular content of one channel/program, while frame 10 may belong to commercial content of the same channel/program, so that both have different frame rates, different video/audio encoder formats, and different video program identifiers and audio program identifiers.
In addition, although in the embodiments shown in fig. 3A and 3B, the frame index table records index information of consecutive frames in the audio-visual content, in some embodiments of the present invention, the frame index table may only record index information of intra pictures/I frames, but not record index information of predicted pictures/P frames or bi-directional predicted pictures/B frames. In addition, in the embodiment shown in fig. 3A and 3B, the index information for the predicted image/P frame, and the bi-predicted image/B frame contains only the frame time and data offset, because the rest of the information (e.g., video program identification code, audio program identification code, video/audio encoder format, frame rate) for these frames is identical to the previous intra picture/I frame.
Since the frame index table contains index information of the intra pictures, the media playback apparatus 100 can more quickly locate the data units containing the intra pictures in the buffer unit 115 or the storage unit 140 according to the frame index table, thereby accelerating the response to the user playback control operation. Further, in an embodiment, the index table construction circuit 130 may name the frame index table using the same master file name as the audio-visual content recorded in the storage unit 140, and store the frame index table in an IDX file format.
Thereafter, when the user again selects and plays back a specific audio-visual content that has been buffered in the buffer unit 115 or that has been recorded in the storage unit 140, the playback control circuit 150 in the media playback apparatus 100 loads a frame index table corresponding to the specific audio-visual content. During playback, if the user inputs an instruction to request the media playback device 100 to perform a search, skip, fast forward or fast backward operation, the playback control circuit 150 determines a target playback time according to the search, skip, fast forward or fast backward operation requested by the user. Then, the playback control circuit 150 searches the frame index table (binary search may be used) using the target playback time, thereby determining a nearest intra picture. The frame time of the nearest intra picture is closest to the target playback time than the frame times of other intra pictures. Then, according to the frame index table, the playback control circuit 150 may obtain the data offset corresponding to the picture in the nearest frame. After obtaining the data offset of the nearest intra picture, the data unit containing the nearest intra picture may be found from the buffer unit 115 or the storage unit 140 according to the data offset corresponding to the nearest intra picture. Playback control circuit 150 may obtain one or more data units starting from the data unit. The one or more data units are decoded by the parsing and decoding circuit 120 and the nearest intra picture or the picture following the nearest intra picture (e.g., the predicted picture following the nearest intra picture or the bi-predicted picture) is restored.
In the following, the frame index table of fig. 3A and 3B will be taken as an example to explain how to use the frame index table in more detail. For example, when the search, skip, fast forward, or fast reverse playback operation entered by the user is converted to the target playback time 180 by the playback control circuit 150, the playback control circuit 150 may search the frame index table to obtain that the frame time 120 of the frame 5 (i.e., intra picture/I frame) is closest to the target playback time 180, so that the frame 5 is considered to be the nearest intra picture. As can be seen from the frame index table, the data offset corresponding to the frame 5 is 1200. Accordingly, the media playback apparatus 100 may obtain one or more data units including the frame 5 from the buffer unit 115 or the storage unit 140, and send the data units to the parsing and decoding circuit 120 to restore the audio-visual content. As shown in fig. 4, the playback control circuit 150 starts from the data Offset 0 of the frame 1, locates one or more data units located in the data Offset 1200, and delivers the one or more data units to the parsing and decoding circuit 120 for decoding.
Note that in the above example, the nearest intra picture is frame 5, but the image is not the actual frame corresponding to the target playback time 180. In fact, the frame time corresponding to frame 7 coincides with the target playback time 180, that is, frame 7 is the target frame. Thus, different implementations of the invention will have different processing modes for this case. In one precision display mode, the playback control circuit 150 does not output the nearest intra picture (frame 5) to the display device 200, but continues to obtain more data units from the buffer unit 115 or the storage unit 140, and waits until the parsing and decoding circuit 120 decodes a frame (i.e., frame 7) whose frame time is closer to the target playback time, then starts from this frame, outputs a subsequent frame to the display device 200, and outputs the relevant audio to the audio device 300. It should be noted that, in some embodiments of the present invention, the frame index table does not record the frame time and data offset associated with the predicted image/P frame or the bi-directional predicted image/B frame. Thus, in such embodiments, the playback control circuit 150 may obtain such unrecorded information via a difference calculation. Take the frame index table of fig. 3A and 3B as an example, and the target playback time 180. The neighboring intra picture earlier than the target playback time 180 is frame 5, and the frame time of frame 5 is 120 and its data offset is 1200. Further, the neighboring intra picture later than the target playback time 180 is frame 10, and the frame time of frame 10 is 270, and the data offset thereof is 1560. Plus the frame rate between frame 5 and frame 10 is 30, it can be estimated by the difference calculation that the frame corresponding to the target playback time 180 is two frames after frame 5 (i.e., frame 7) and the data offset thereof is approximately 1300. In this way, the playback control circuit 150 can determine that the frame 7 (i.e., the target frame) is to be output in response to the playback control operation of the user in the precise mode, and output the frame from the frame 7 to the display device 200 and output the relevant audio to the audio device 300.
In a fast display mode of the present invention, the parsing and decoding circuit 120 responds to the playback control operation of the user with the nearest intra picture (e.g., frame 5) obtained from the frame index table, and outputs the frame from frame 5 to the display device 200 and the associated audio to the audio device 300. In another embodiment of the fast display mode of the present invention, the parsing and decoding circuit 120 obtains the nearest intra picture (e.g. frame 5) and the predicted image/P frame or bi-predicted image/B frame after the nearest intra picture, and then the parsing and decoding circuit 120 starts outputting the frame from the predicted image/P frame or bi-predicted image/B frame to the display device 200 and outputs the relevant audio to the audio device 300.
Fig. 5 shows a flowchart of an image processing method of an embodiment of the present invention. As shown in the drawings, the image processing method of the present invention comprises the following steps:
s310: a buffer unit for receiving the data stream on the media playback device to buffer the data stream in the media playback device, or a storage unit recorded in the media playback device;
s320: analyzing the cached or recorded data stream to obtain frame time and data offset corresponding to one or more intra-frame pictures of at least one audiovisual content contained in the data stream;
s330: constructing a frame index table corresponding to the audio-visual content according to frame time and data offset corresponding to one or more intra-frame pictures respectively;
s340: in response to a playback operation of the audiovisual content, determining a data offset corresponding to the playback operation with reference to the frame index table;
s350: acquiring one or more data units in a cache unit or a storage unit according to the data offset; and
s360: the one or more data units are decoded for media playback.
Since the principle of the above steps and the specific details have been described in the previous embodiments, a repeated description is not made here. It should be noted that the above-described process may be better implemented for media playback by adding other additional steps or making appropriate changes and adjustments, thereby further enhancing the user experience. In addition, all or part of the operations, circuit elements, and modules in the foregoing embodiments of the present invention may be implemented by the apparatus 400 shown in fig. 6, for example, the index table construction circuit 130 and the playback control circuit 150 in the foregoing media playback apparatus 100. Wherein a storage unit 410 (e.g., non-volatile memory) in the device 400 may be used to store source code, instructions, variables, or data. The hardware processing unit 420 (e.g., a general-purpose processor) in the apparatus 400 may execute the source codes and instructions stored in the memory unit 410, and perform all the operations described in the foregoing embodiments with reference to the variables or data therein.
Embodiments of the invention may be implemented using hardware, software, firmware, and related combinations thereof. Embodiments of the invention may be implemented using software or firmware stored in a memory and a corresponding instruction execution processor via an appropriate instruction execution system. In terms of hardware, this can be accomplished using any one of the following techniques or a combination thereof: an individual operation logic with logic gates that perform logic functions based on data signals, an application specific integrated circuit (application specific integrated circuit, ASIC), programmable gate array (programmable gate array, PGA), or a field programmable gate array (field programmable gate array, FPGA) with appropriate combinational logic gates, or the like.
The flowcharts and blocks in the flowchart illustrations within the specification illustrate the architecture, functionality, and operation of implementations of systems, methods and computer software products according to various embodiments of the present invention. In this regard, each block in the flowchart or functional block diagrams may represent a module, segment, or portion of source code, which comprises one or more executable instructions for implementing the specified logical function(s). Additionally, each block of the functional block diagrams and/or flowchart illustration, and combinations of blocks in the functional block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer program instructions. These computer program instructions may also be stored in a computer-readable medium that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable medium implement the function/act specified in the flowchart and/or block diagram block or blocks.
The foregoing description is only of the preferred embodiments of the invention, and all changes and modifications that come within the meaning and range of equivalency of the claims are therefore intended to be embraced therein.
[ symbolic description ]
10. Media source
100. Media playback apparatus
200. Display apparatus
300. Sound equipment
110. Receiving and buffering circuit
111. Network receiving device
112. Frequency modulation device
113. Descrambling device
114. Demultiplexer
115. Cache unit
120. Parsing and decoding circuit
122. Video parsing and decoding unit
124. Audio parsing and decoding unit
130. Index table construction circuit
140. Memory cell
150. Playback control circuit
160. Recording circuit
S310 to S360 steps
400. Device and method for controlling the same
410. Memory cell
420. Hardware processing unit

Claims (10)

1. A method for media playback on a media playback device, comprising:
receiving a data stream on the media playback device, a buffer unit for buffering the data stream in the media playback device, or a storage unit recorded in the media playback device;
performing an analysis operation on the cached or recorded data stream to obtain frame time and data offset corresponding to one or more intra-frame pictures of at least one audiovisual content contained in the data stream;
constructing a frame index table corresponding to the at least one audio-visual content according to the frame time and the data offset corresponding to the one or more intra-frame pictures respectively;
determining a data offset corresponding to a playback operation of the at least one audiovisual content by referring to the frame index table in response to the playback operation;
acquiring one or more data units in the cache unit or the storage unit according to the data offset; and
the one or more data units are decoded for media playback.
2. The method of claim 1, further comprising:
determining a target playback time according to the playback operation;
searching the frame index table according to the target playback time to determine a nearest intra-frame picture, wherein the difference between the frame time corresponding to the nearest intra-frame picture and the target playback time is smaller than the difference between the frame time corresponding to all other intra-frame pictures in the frame index table and the target playback time;
obtaining a data offset corresponding to the picture in the nearest frame from the frame index table; and
and acquiring the one or more data units from the cache unit or the storage unit according to the data offset.
3. The method of claim 1, further comprising:
determining a target playback time according to the playback operation;
searching the frame index table, and determining a first adjacent intra-frame picture with a frame time earlier than the target playback time;
determining a target frame corresponding to the target playback time according to the frame rate of the picture in the first adjacent frame in the frame index table; and
decoding the one or more data units until the target frame is obtained, and not starting outputting the frame to a display device.
4. A media playback device, comprising:
a receiving and buffering circuit for receiving a data stream to buffer the data stream in a buffering unit;
the analysis and decoding circuit is coupled with the receiving and buffering circuit and is used for carrying out analysis operation on the buffered data stream or the data stream recorded in a storage unit so as to obtain frame time and data offset corresponding to one or more intra pictures of at least one audiovisual content contained in the data stream;
an index table constructing circuit, coupled to the parsing and decoding circuit, for constructing a frame index table corresponding to the at least one audiovisual content according to the frame time and the data offset corresponding to the one or more intra-frame pictures, respectively; and
a playback control circuit, configured to respond to a playback operation of the at least one audiovisual content, determine a data offset corresponding to the playback operation by referring to the frame index table, and obtain one or more data units in the buffer unit or the storage unit according to the data offset;
wherein the parsing and decoding circuitry is further for decoding the one or more data units for media playback.
5. The media playback device of claim 4, wherein the index table construction circuit is further configured to construct the frame index table based on one or more of a video program identification code, an audio program identification code, a video encoder format, an audio encoder format, and a frame rate, respectively, to which the one or more intra pictures correspond.
6. The media playback device of claim 4, wherein the parsing and decoding circuitry is further configured to perform parsing operations to obtain one or more predicted images of the at least one audiovisual content and/or frame times and data offsets corresponding to the one or more bi-directional predicted images, respectively; the index table constructing circuit is further configured to construct the frame index table according to the one or more predicted images and/or frame time and data offset corresponding to the one or more bi-directional predicted images.
7. The media playback device of claim 4, wherein the playback control circuit is to:
determining a target playback time according to the playback operation;
searching the frame index table according to the target playback time to determine a nearest intra-frame picture, wherein the difference between the frame time corresponding to the nearest intra-frame picture and the target playback time is smaller than the difference between the frame time corresponding to all other intra-frame pictures in the frame index table and the target playback time;
obtaining a data offset corresponding to the picture in the nearest frame from the frame index table; and
and acquiring the one or more data units from the cache unit or the storage unit according to the data offset.
8. The media playback apparatus of claim 7, wherein the parsing and decoding circuitry is to decode the one or more data units to obtain the nearest intra picture and output frames from the nearest intra picture to a display device.
9. The media playback apparatus of claim 7, wherein the parsing and decoding circuit is configured to decode the one or more data units to obtain the nearest intra picture and a predicted picture or a bi-predicted picture following the nearest intra picture, and to output frames from the predicted picture or the bi-predicted picture to a display device.
10. The media playback device of claim 4, wherein the playback control circuit is to determine a target playback time according to the playback operation, search the frame index table to determine a first adjacent intra-frame picture having a frame time earlier than the target playback time, and determine a target frame corresponding to the target playback time according to a frame rate of the frame index table with respect to the first adjacent intra-frame picture; and the parsing and decoding circuit is used for decoding the one or more data units and not starting outputting the frame to a display device until the target frame is obtained.
CN202210489074.0A 2022-05-06 2022-05-06 Media playback method and related media playback device Pending CN117061813A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210489074.0A CN117061813A (en) 2022-05-06 2022-05-06 Media playback method and related media playback device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210489074.0A CN117061813A (en) 2022-05-06 2022-05-06 Media playback method and related media playback device

Publications (1)

Publication Number Publication Date
CN117061813A true CN117061813A (en) 2023-11-14

Family

ID=88659577

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210489074.0A Pending CN117061813A (en) 2022-05-06 2022-05-06 Media playback method and related media playback device

Country Status (1)

Country Link
CN (1) CN117061813A (en)

Similar Documents

Publication Publication Date Title
JP4769717B2 (en) Image decoding method
EP2186340B1 (en) A video data reproduction system
US8918533B2 (en) Video switching for streaming video data
US6628890B1 (en) Digital recording/reproduction apparatus
KR20060047952A (en) Reverse presentation of digital media streams
US5739862A (en) Reverse playback of MPEG video
EP2642764B1 (en) Transcoding a video stream to facilitate accurate display
US20110081133A1 (en) Method and system for a fast channel change in 3d video
US20110135286A1 (en) Apparatus and method for extracting key frames and apparatus and method for recording broadcast signals using the same
US6891547B2 (en) Multimedia data decoding apparatus and method capable of varying capacity of buffers therein
US20110080948A1 (en) Method and system for 3d video decoding using a tier system framework
KR101142379B1 (en) Method and Apparatus of playing digital broadcasting and Method of recording digital broadcasting
JP2001346165A (en) Image processing method and image processing unit utilizing this method and television receiver
US20090136204A1 (en) System and method for remote live pause
US6373905B1 (en) Decoding apparatus and decoding method
CN117061813A (en) Media playback method and related media playback device
US20130287361A1 (en) Methods for storage and access of video data while recording
TWI819580B (en) Media playback method for improving playback response based on pre-parsing operation and related media playback device
JP2008072182A (en) Moving picture decoding device, moving picture decoding method, moving picture decoding program, moving picture encoding device, moving picture encoding method, moving picture encoding program, and moving picture encoding and decoding device
JP2008136211A (en) Encoding device, decoding device, recording device, and audio/video data transmission system
JP2003052020A (en) Picture processing method and picture processor capable of using the same and television receiver
KR100956821B1 (en) Method for playing Personal Video Recorder
JP2001238167A (en) Image reproducing method, image reproducing apparatus capable of using the method and television receiver
KR20150018108A (en) Stream recycling method for waiting time shortening of screen refreshing in IPTV broadcasting, and computer-readable recording medium for the same
US20100178039A1 (en) Video and audio playback apparatus and video and audio playback method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination