GB2602101A - Method and apparatus for encapsulating image data in a file for progressive rendering - Google Patents

Method and apparatus for encapsulating image data in a file for progressive rendering Download PDF

Info

Publication number
GB2602101A
GB2602101A GB2020066.3A GB202020066A GB2602101A GB 2602101 A GB2602101 A GB 2602101A GB 202020066 A GB202020066 A GB 202020066A GB 2602101 A GB2602101 A GB 2602101A
Authority
GB
United Kingdom
Prior art keywords
progressive
image
item
sub
version
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
GB2020066.3A
Other versions
GB202020066D0 (en
Inventor
Ruellan Hervé
Denoual Franck
Maze Frédéric
Ouedraogo Naël
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Canon Inc
Original Assignee
Canon Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Canon Inc filed Critical Canon Inc
Priority to GB2020066.3A priority Critical patent/GB2602101A/en
Publication of GB202020066D0 publication Critical patent/GB202020066D0/en
Priority to GB2105269.1A priority patent/GB2602169A/en
Priority to GB2105852.4A priority patent/GB2602170B/en
Priority to GB2304517.2A priority patent/GB2617913B/en
Priority to EP21839925.1A priority patent/EP4264949A1/en
Priority to US18/257,236 priority patent/US20240107129A1/en
Priority to JP2023508604A priority patent/JP2023552029A/en
Priority to PCT/EP2021/086002 priority patent/WO2022129235A1/en
Publication of GB2602101A publication Critical patent/GB2602101A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234318Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by decomposing into objects, e.g. MPEG-4 objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234327Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by decomposing into layers, e.g. base layer and one or more enhancement layers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234363Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by altering the spatial resolution, e.g. for clients with a lower screen resolution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/23439Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements for generating different versions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/236Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/236Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
    • H04N21/23605Creation or processing of packetized elementary streams [PES]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/236Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
    • H04N21/23614Multiplexing of additional data and video streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/434Disassembling of a multiplex stream, e.g. demultiplexing audio and video streams, extraction of additional data from a video stream; Remultiplexing of multiplex streams; Extraction or processing of SI; Disassembling of packetised elementary stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/434Disassembling of a multiplex stream, e.g. demultiplexing audio and video streams, extraction of additional data from a video stream; Remultiplexing of multiplex streams; Extraction or processing of SI; Disassembling of packetised elementary stream
    • H04N21/4343Extraction or processing of packetized elementary streams [PES]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • H04N21/440263Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display by altering the spatial resolution, e.g. for displaying on a connected PDA
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • H04N21/44029Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display for generating different versions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8146Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics
    • H04N21/8153Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics comprising still images, e.g. texture, background image
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/84Generation or processing of descriptive data, e.g. content descriptors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/85406Content authoring involving a specific file format, e.g. MP4 format

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Security & Cryptography (AREA)
  • Computer Graphics (AREA)
  • Image Processing (AREA)
  • Editing Of Facsimile Originals (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

A method of encapsulating image data in a media file, the image data being related to a main image to be generated based on a plurality of input images, the method comprising: obtaining the plurality of input images, at least one input image being associated with different versions of the input image; generating descriptive metadata for describing the main image; encapsulating the plurality of input images and the descriptive metadata in the media file; where the method further comprises: generating progressive information (e.g. an indication of a progressive refinement pattern, indicative of a scanning order of input images, e.g. an index in a list of predefined progressive refinement patterns) defining a set of consecutive progressive steps for generating a version of the main image, each progressive step being associated with a set of input image versions required to generate the version of the main image; and embedding the progressive information in the descriptive metadata.

Description

METHOD AND APPARATUS FOR ENCAPSULATING IMAGE DATA IN A FILE FOR PROGRESSIVE RENDERING
FIELD OF THE INVENTION
The present disclosure concerns a method and a device for encapsulating image data in a file. It concerns more particularly a method of encapsulation allowing a progressive rendering of an image.
BACKGROUND OF INVENTION
Images displayed to a user can be loaded from a local disk, a network disk or from a remote server. To be loaded, the images are typically encapsulated within a file comprising the image data and metadata. The metadata are used to describe the organisation of the image data in the file, the type of the content and any useful information for the rendering or the image data.
The images are typically encoded to reduce the size of data on the storage device. Many encoding standards may be used, like JPEG, AV1, or the more recent HEVC or VVC standards.
The HEVC and VVC standards define profiles for the encoding of still images and describe specific tools for compressing single still images or bursts of still images. An extension of the ISO Base Media File Format (ISOBMFF) used for such kind of image data has been proposed for inclusion into the ISO/IEC 23008 standard, in Part 12, under the name: "H El F" or "High Efficiency Image File Format".
HEIF (High Efficiency Image File Format) is a standard developed by the Moving Picture Experts Group (MPEG) for storage and sharing of images and image sequences.
The MIAF (Multi-Image Application Format) is a standard developed by MPEG into ISO/IEC 23000 standard part 22. The MIAF specification specifies a multimedia application format, the Multi-Image Application Format (MIAF), which enables precise interoperability points for the creation, reading, parsing and decoding of images embedded in the High Efficiency Image File (H El F) format. The MIAF specification fully conforms to the HEIF format and only defines additional constraints to ensure higher interoperability.
Depending on the location of the image, loading the image from the local or remote storage system may require a delay long enough to be noticeable by a user. Some formats, such as JPEG or JPEG-2000 enable a progressive rendering of images where a full but low-quality version of the image is contained in the beginning of the file.
This low-quality version of the image is refined, possibly in multiple passes by the rest of the file, providing a full-quality version with the whole file. A progressive format enables to quickly display a full-size preview version of the image without waiting for the full reception of the image. As such it increases the reactivity of the displaying of images that are loaded slowly and improves the user experience.
The HEIF format has no explicit support for progressive rendering.
HEIF defines derived images that are representations of images built by applying an operation on other images present in the same file. Two types of derived images in particular are the grid type and the overlay type. A grid is an array of smaller images that all have the same size. An overlay defines a derived image by overlaying one or more images in a given layering order within a larger canvas.
The HEIF and MIAF file formats do not provide a mechanism allowing a progressive rendering of the grid and overlay derived image items. In particular, there is no description on how to progressively build from their individual components and render as a whole, progressively, a derived image of the grid type or of the overlay type, or for any type of derived image based on a plurality of input images.
SUMMARY OF THE INVENTION
The present invention has been devised to address one or more of the foregoing concerns. It concerns a method and device for encapsulating image data in a file in view of the progressive rendering of the image. It concerns more particularly, images built based on a plurality of input images. It applies in particular to derived images of the type grid or overlay, but is not limited to these types of images.
According to a first aspect of the invention, there is provided a method of encapsulating image data in a media file, the image data being related to a main image to be generated based on a plurality of input images, the method comprising: obtaining the plurality of input images, at least one input image being associated with different versions of the input image; generating descriptive metadata for describing the main image; encapsulating the plurality of input images and the descriptive metadata in the media file; wherein the method further comprises: generating a progressive information defining a set of consecutive progressive steps for generating a version of the main image, each progressive step being associated with a set of input image versions required to generate the version of the main image; and embedding the progressive information in the descriptive metadata.
In an embodiment, the progressive information comprises an indication of a progressive refinement pattern, the progressive refinement pattern being indicative of a scanning order of the plurality of input images.
In an embodiment, the indication of a progressive refinement pattern is an index in a list of predefined progressive refinement patterns.
In an embodiment, the progressive information further comprises a parameter associated with the progressive refinement pattern.
In an embodiment, the indication of a progressive refinement pattern is based on an equation and associated parameters.
In an embodiment, the progressive information comprises, for each progressive step, a position in the file of the input image data associated with the progressive step.
In an embodiment, the position indicates the last byte of image data in the file associated with the progressive step.
In an embodiment, the position comprises an offset and a length to indicate the input image data associated with the progressive step.
In an embodiment, the position comprises a list of extents of the input image data associated with the progressive step.
In an embodiment, input image data being organized into one or more extent, each extent being composed of contiguous image data related to a version of one input image, the progressive information comprises, in descriptive metadata describing an extent, an information indicating that the last byte of the extent corresponds to a progressive step.
In an embodiment, input images being described as image items, the progressive information comprises for each progressive step a list of the image item identifiers identifying the image items associated with the progressive step.
In an embodiment, the progressive information comprises for each progressive step a number of input image version comprised in the progressive step.
In an embodiment, at least one input image being composed of a plurality of layers, each layer being associated with a version of the input image, the progressive information further comprises a layer identifier associated with the image item identifier of the input image.
In an embodiment, generating a progressive information comprises generating a progressive refinement data structure comprising data for determining the location of reconstruction points in the media file, each reconstruction point indicating that a reconstruction of the main image is possible using image data associated to input images previously received.
In an embodiment, the progressive refinement data structure further comprises a number of reconstruction points.
In an embodiment, the progressive refinement pattern defines the location of reconstruction points in the media file, each reconstruction point indicating that a reconstruction of the main image is possible using image data associated to input images previously received.
In an embodiment, the progressive information characterizes a construction of the main image so that its quality is gradually improved.
According to another aspect of the invention, there is provided a method of generating a main image to be generated based on a plurality of input images from an image data file, wherein the method comprises: obtaining from the image data file descriptive metadata describing the main image; obtaining from the descriptive metadata a progressive information defining a set of consecutive progressive steps for generating a version of the main image, each progressive step being associated with a set of input image versions required to generate the version of the main image; obtaining image data corresponding to the input images from the image data file; and generating at least two versions of the main image corresponding to respective progressive steps, each version of the main image being generated when the set of input image versions associated with the respective progressive step is obtained from the image data file.
According to another aspect of the invention, there is provided a computer program product for a programmable apparatus, the computer program product comprising a sequence of instructions for implementing a method according to the invention, when loaded into and executed by the programmable apparatus.
According to another aspect of the invention, there is provided a computer-readable storage medium storing instructions of a computer program for implementing a method according to the invention.
According to another aspect of the invention, there is provided a computer program which upon execution causes the method of the invention to be performed.
According to another aspect of the invention, there is provided a device for encapsulating image data in a media file, the image data being related to a main image to be generated based on a plurality of input images, the device comprising a processor configured for: obtaining the plurality of input images, at least one input image being associated with different versions of the input image; generating descriptive metadata for describing the main image; - encapsulating the plurality of input images and the descriptive metadata in the media file; wherein the method further comprises: - generating a progressive information defining a set of consecutive progressive steps for generating a version of the main image, each progressive step being associated with a set of input image versions required to generate the version of the main image; and - embedding the progressive information in the descriptive metadata.
According to another aspect of the invention, there is provided a device for generating a main image to be generated based on a plurality of input images from an image data file, wherein the device comprises a processor configured for: - obtaining, from the image data file, descriptive metadata describing the main image; obtaining, from the descriptive metadata, a progressive information defining a set of consecutive progressive steps for generating a version of the main image, each progressive step being associated with a set of input image versions required to generate the version of the main image; - obtaining image data corresponding to the input images from the image data file; and generating at least two versions of the main image corresponding to respective progressive steps, each version of the main image being generated when the set of input image versions associated with the respective progressive step is obtained from the image data file.
At least parts of the methods according to the invention may be computer implemented. Accordingly, the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a "circuit", "module" or "system".
Furthermore, the present invention may take the form of a computer program product embodied in any tangible medium of expression having computer usable program code embodied in the medium.
Since the present invention can be implemented in software, the present invention can be embodied as computer readable code for provision to a programmable apparatus on any suitable carrier medium. A tangible, non-transitory carrier medium may comprise a storage medium such as a floppy disk, a CD-ROM, a hard disk drive, a magnetic tape device or a solid state memory device and the like. A transient carrier medium may include a signal such as an electrical signal, an electronic signal, an optical signal, an acoustic signal, a magnetic signal or an electromagnetic signal, e.g., a microwave or RE signal
BRIEF DESCRIPTION OF THE DRAWINGS
Embodiments of the invention will now be described, by way of example only, and with reference to the following drawings in which: Figure 1 illustrates an example of a HEIF file that contains media data like one or more still images and possibly one or more video and/or one or more sequences of images; Figure 2 illustrates a first example of file structure for enabling a progressive rendering of a grid item; Figure 3 illustrates another file structure for enabling a progressive rendering of a similar grid item; Figure 4a to 4c illustrate different ordering of the image data in an image file; Figure 4d illustrates another example where the same ordering as Figure 4b is used in relation with four different progressive steps; Figure 4e represents another example where a different ordering than Figure 4c is used in relation with four different progressive steps; Figure 4f illustrates the rendering of the grid image item after the second progressive step corresponding to Figure 4d; Figure 4g illustrates the rendering of the grid image item after the second progressive step corresponding to Figure 4e; Figure 5 illustrates the main steps of generating an HEIF file according to embodiments of the invention; Figure 6 illustrates the main steps for a progressive rendering of an HEIF file generated according to embodiments of the invention; Figure 7 illustrates the main steps for a progressive rendering of an HEIF file generated according to other embodiments of the invention; and Figure 8 is a schematic block diagram of a computing device for implementation of one or more embodiments of the invention.
DETAILED DESCRIPTION OF THE INVENTION
The HEVC or VVC standard defines a profile for the encoding of still images and describes specific tools for compressing single still images or bursts of still images. An extension of the ISO Base Media File Format (ISOBMFF) used for such kind of image data has been proposed for inclusion into the ISO/IEC 23008 standard, in Part 12, under the name: "H El F" or "High Efficiency Image File Format".
The HEIF and MIAF standards cover two forms of storage corresponding to different use cases: * the storage of image sequences, which can be indicated to be displayed as a timed sequence or by other means, and in which the images may be dependent on other images, and * the storage of a single coded image or a collection of independently coded images, possibly with derived images.
In the first case, the encapsulation is close to the encapsulation of the video tracks in the ISO Base Media File Format (see document « Information technology -Coding of audiovisual objects -Part 12: ISO base media file format», w18855, ISO/IEC 14496-12, Sixth edition, October 2019), and similar tools and concepts are used, such as the Irak' boxes and the sample grouping for description of groups of samples. The trak' box is a file format box that contains sub boxes for describing a track, that is to say, a timed sequence of related samples.
Boxes, also called containers, are hierarchical data structures provided to describe the data in the files. Boxes are object-oriented building blocks defined by a unique type identifier (typically a four-character code, also noted FourCC or 4CC) and a length. All data in a file (media data and metadata describing the media data) is contained in boxes. There is no other data within the file. File-level boxes are boxes that are not contained in other boxes.
In the second case, a set of ISOBMFF boxes, the 'meta' boxes are used. These boxes and their hierarchy offer fewer description tools than the 'track-related' boxes ('trak' box hierarchy) and relate to "information items" or "items" instead of related samples. It is to be noted that the wording 'box' and the wording 'container' may be both used with the same meaning to refer to data structures that contain metadata describing the organization or/and properties of the image data in the file.
Figure 1 illustrates an example of a HEIF file 101 that contains media data like one or more still images and possibly one or more video and/or one or more sequences of images. This file contains a 'ftyp' box (FileTypeBox) 111 that contains an identifier of the type of file (typically a set of four-character codes). This file contains a box called 'meta' (MetaBox) 102 that is used to contain general untimed metadata including metadata structures describing the one or more still images. This 'meta' box 102 contains an ciinf box (ItemInfoBox) 121 that describes several single images. Each single image is described by a metadata structure ItemInfoEntry also denoted items 1211 and 1212. Each item has a unique 16-bit or 32-bit identifier item_l D. The media data corresponding to these items is stored in a container for media data, e.g., the 'mdaf box 104. The media data may also be stored in an cidat or in an tridat box, in an 'imda' box or in another file. An 'Hoc' box (ItemLocafionBox) 122 provides for each item the offset and length of its associated media data in the rndat, idat, or 'imda' box 104. The media data for an item may be fragmented into extents. In this case, the 'Hoc' box 122 provides the number of extents for the item and for each extent its offset and length in the tridal, cidar, or 'imda' box 104. An 'iref box (ItemReferenceBox) 123 may also be defined to describe the association of one item with other items via typed references.
Optionally, for describing the storage of image sequences or video, the HEIF file 101 may contain a box called crnoov' (MovieBox) 103 that describes one or more image sequences or video tracks 131 and 132. Typically, the track 131 may be an image sequence ('pict') track designed to describe a set of images for which the temporal information is not necessarily meaningful and 132 may be a video ('vide') track designed to describe video content. Both tracks describe a series of image samples, an image sample being a set of pixels captured at the same time, for example a frame of a video sequence. The main difference between the two tracks is that in 'pict' tracks the timing information is not necessarily meaningful whereas for 'vide' tracks the timing information is intended to constraint the timing of the display of the samples. The data corresponding to these samples is stored in the container for media data, the mdaf box 104.
The 'mdat' container 104 stores the untimed encoded images corresponding to items as represented by the data portions 141 and 142 and optionally the timed encoded images corresponding to samples as represented by the data portion 143 when a video track is also present in the HEIF file.
An HEIF file 101 offers different alternatives to store multiple images. For instance, it may store the multiple images either as items or as a track of samples that can be a 'pict' track or a 'vide' track. The actual choice is typically made by the application or device generating the file according to the type of images and the contemplated usage of the file.
A HEIF item is a derived image item, when it has a 'climg' item reference to one or more other image items, which are inputs to the derivation.
An item with an item_type value of ciovl' defines a derived image item by overlaying one or more input images in a given layering order within a larger canvas. The input images are listed in the order they are layered in the ding' item reference for this derived image item. The data of the overlay image item specifies the location of each input image within the larger canvas.
An item with an item_type value of grid' defines a derived image item whose reconstructed image is formed from one or more input images in a given grid order within a larger canvas. The input images are listed in the grid order in the ding' item reference for this derived image item. The data of the grid image item specifies the number of rows and columns in the grid and the size of the larger canvas.
The ISO Base Media File Format specifies several alternatives to group samples or items depending on the container that holds the samples or items to group. These alternatives can be considered as grouping data structures or grouping mechanism, i.e., boxes or data structures providing metadata describing a grouping criterion and/or group properties and/or group entities.
A first grouping mechanism represented by an EntityToGroupBox is adapted for the grouping of items or tracks. In this mechanism, the wording 'entity' is used to refer to items or tracks or other EntityToGroupBoxes. This mechanism specifies the grouping of entities. An EntityToGroupBox is defined according to the following syntax: aligned(8) class EntityToGroupBox(grouping_type, version, flags) extends FullBox(grouping_type, version, flags) { unsigned int(32) group_id; unsigned int(32) num_entities_in_group; for(i=0; i<num_entities_in_group; i++) unsigned int(32) entity_id; // the remaining data may be specified for a particular grouping_type The grouping_type is used to specify the type of the group. Several values for the grouping_type are specified in HEIF. The group_id provides an identifier for the group of entities. The enfity_id represents the identifier of entities that compose the group, i.e., either a track_ID for a track, an item_ID for an item or another group_id for an entity group. In Figure 1, the groups of entities inheriting from the EnfityToGroup box 1241 and 1242 are comprised in the container 124 identified by the four characters code 'gprl' for Groups ListBox.
Entity grouping consists in associating a grouping type which identifies the reason of the grouping of a set of items, tracks or other entity groups. In this document, it is referred to Grouping Information as information in one of the EntityToGroup Boxes which convey information to group a set of images.
A grouping_type value of 'alt r' indicates that the items and tracks mapped to this grouping are alternatives to each other, and only one of them should be rendered.
A player should select the first entity in the grouping that it can process and that suits the application needs.
ISOBMFF provides a mechanism to describe and associate properties with items. These properties are called item properties. The ItemPropertiesBox iprp' 125 enables the association of any item with an ordered set of item properties. The ItemPropertiesBox consists of two parts: an item property container box ipco' 1251 that contains an implicitly indexed list of item properties 1253, and an item property association box ipma' 1252 that contains one or more entries. Each entry in the item property association box associates an item with its item properties. The HEIF standard extends this mechanism to enable the association of item properties with items and/or entity groups. Note that in the description, for genericity, we generally use item properties to designate both properties of an item or properties of an entity group. An item property associated with an entity group applies to the entity group as a whole and not individually to each entity within the group.
The associated syntax is as follows: aligned(8) class ItemProperty(property_type) extends Box(property_type) aligned(8) class ItemFullProperty(property_type, version, flags) extends FullBox(property_type, version, flags) aligned(8) class ItemPropertyContainerBox extends Box( ipco') properties ItemProperty() []; If boxes derived from // ItemProperty or ItemFullProperty, to fill box aligned(8) class ItemPropertyAssociation extends FullBox('ipma', version, flags) unsigned int(32) entry_count; for(i = 0; i < entry_count; i++) { if (version < 1) unsigned int(16) item_ID; else unsigned int(32) item_ID; unsigned int(8) association_count; for (i=0; i<association_count; i++) { bit(1) essential; if (flags & 1) unsigned int(15) property_index; else unsigned int(7) property_index; aligned(8) class ItemPropertiesBox extends Box(riprp') { ItemPropertyContainerBox property container; ItemPropertyAssociation association[]; The ItemProperty and ItemFullProperty boxes are designed for the description of an item property. ItemFullProperty allows defining several versions of the syntax of the box and may contain one or more parameters whose presence is conditioned by either the version or the flags parameter.
The ItemPropertyContainerBox is designed for describing a set of item properties as an array of ItemProperty boxes or ItemFullProperty boxes.
The ItemPropertyAssociation box is designed to describe the association between items and/or entity groups and their item properties. It provides the description of a list of item identifiers and/or entity group identifiers, each identifier (item_ID) being associated with a list of item property index referring to an item property in the ItemPropertyContainerBox.
The goal of the invention is to provide a fine control over the granularity of a progressive rendering in HEIF of an image item that is built from several image components. Such an image item may be a grid derived image item (also denoted grid item), an overlay derived image item (also denoted overlay item) or any other type of derived image based on a plurality of input images. These are built from several input image items. Such an input image item may also be a coded image items encoded using HEVC files or layers or using VVC sub-pictures or layers. The rest of the description is based on grid items for clarity, but it could also apply to other image items built from several image components (also denoted input images).
In the following description, the target image of a progressive rendering is called the main image. The components or input images used to build this main image are called the sub-images. A progressive step refers to a reconstructed version of this main image. The initial progressive step corresponds to an empty image, while the final progressive step corresponds to the main image. Intermediate progressive steps may have a lower resolution and/or a lower quality than the main image. Note that in the following description, the initial progressive step is sometimes ignored for simplicity.
Possibly the progressive rendering functionality may be used for other purposes than displaying the image. For example, a process for detecting people on an image may use a fast first check on the low-quality version of the images for quickly rejecting images not containing people.
The progressive rendering of the main image is realized by using different coded versions of the sub-images. A reconstructed version of the main image may consist in decoding all the input images and assembling these decoded sub-images according to the type of the main image item (e.g., grid, overlay, sub-picture composition...) A sub-image version may be the sub-image itself, a thumbnail of the sub-image, a lower quality and/or lower-resolution coded version of the sub-image. The different coded versions may also be contained in one or more layers of a multi-layer image. When the sub-image is encoded with several quality or resolution levels, the progressive rendering of the main image may use a subset of these quality or resolution levels. The content of a progressive step refers to the set of sub-image versions that are used during the reconstruction of the version of the main image corresponding to this progressive step. The refinements of a progressive step refer to the subset of sub-image versions that are part of the content of the progressive step and that were not part of the content of the previous progressive step.
Each progressive step may be associated with a reconstruction point. A reconstruction point is a location in the file corresponding to a point in the file where all the versions of input images required to generate the version of the main image corresponding to the progressive step have been run through. It means that, when reading the file, a reconstruction point is the point in the file where all the required image data for a given progressive step have been read.
When a sub-image is a multi-layer image, i.e., encoded with several quality or resolution layers, the sub-image may be represented by a single item whose media data contain all the encoded layers. Possibly, the media data corresponding to the different layers may be fragmented and described in the Item Location Box Hoc' using extents. For example, the media data for a sub-image encoded as a multi-layer image with two layers may be located using two extents, each extent indicating the location of the data corresponding to a layer when this sub-image is described by one image item.
The multi-layer image may also be represented by several image items corresponding to the different layers, or different subsets of layers. These image items may share the common parts of the media data. For example, two different image items may correspond to a sub-image encoded as a multi-layer image with two layers. The first image item would correspond to the first layer and its media data would be located using an extent indicating the location of the data corresponding to the first layer. The second image item would correspond to the combination of both layers and its media data would be located using two extents indicating respectively the location of the data corresponding to the two layers. A description of the layers may also be available in an image property to indicate which layers are contained in an image item. Thus, each image item representing a sub-image encoded as a multi-layer image may be considered as a sub-image version, like different coded versions of a non-multi-layer image item. When the main image is an HEVC image encoded with tiles or a VVC image encoded with sub-pictures, a sub-image may be an image item corresponding to a tile (e.g. chvtl item) or to a sub-picture (e.g. vvst or vvc1' item). The main image may be an HEVC item, also called here HEVC base item, reconstructed from the tile sub-images that are associated to it using an item reference of type 'tbas' from each file sub-image item to the base item, the positions of the tile sub-images being provided by an 'rloc' property. The main image may be a VVC base item reconstructed from the sub-picture sub-images that are associated to it using an item reference of type 'subp' from the base item to the sub-picture sub-image items. A sub-image version may be a tile or a sub-picture encoded at a given quality or resolution levels. A tile or a sub-picture may also be encoded as a multi-layer image and a version of such sub-image may be obtained by decoding a subset of its quality or resolution layers.
Figure 2 illustrates a first example of file structure for enabling a progressive rendering of a grid item (the same would apply to an overlay or to an HEVC base item or to a WC base item: overlay also references sub-images through a 'dimg' item reference type, while an HEVC base item is referred to by tile sub-images via a 'tbas' item reference type and a VVC base item references sub-images through a 'subp' item reference type). The main or primary image item is the grid item 202, "high-quality grid", which is built using the nine sub-images "h1", referenced 220, to "h9", referenced 228. These sub-images are encoded with a high-quality. This main image is included in an entity group of type 'altr' 200 that also contains the grid item 201, "low-quality grid", which is built using the 9 sub-images "11", referenced 210, to "19", referenced 218. These sub-images are encoded with a low-quality. A progressive rendering of the main image 202 may be realized by first rendering the grid item 201 once all the sub-images of low-quality 210 to 218 are available and then rendering the grid item 202 once all the sub-images of high-quality 220 to 228 are available. However, this file structure only provides two refinement steps. In particular, if the main image 202 is large with a good quality, the quantity of data corresponding to the sub-images 220 to 228 may be large and take some time to load. Moreover, there are no guarantee that the file organizes the data for low-quality versions and high-quality versions of the sub-images in a way enabling progressive refinement.
Figure 3 illustrates another file structure for enabling a progressive rendering of a similar grid item. The main or primary image item is the grid item 300, composed of the nine sub-images "h1", referenced 330, to "h9", referenced 338, encoded in high-quality. Each sub-image is associated to a low-quality encoded alternative through an 'aft( entity group: the nine sub-images "h1" to "h9" are respectively associated to the nine sub-images "11", referenced 320, to "19", referenced 328. For example, the 'eV entity group associating "h1" with "11" is referenced 310. As such, the grid item may be reconstructed using either the low-quality version or the high-quality version of each sub-image, depending on which version is available for each sub-image. This file structure can provide up to ten different progressive steps: the first step consists of all the low-quality sub-images, the second step replaces one of these low-quality sub-images with a high-quality one, and so on until the tenth step where only the high-quality sub-images are used.
However, the actual progressive steps depend on the ordering of the sub-image data inside the HEIF file. Figure 4a represents a first ordering where almost no progressive rendering is possible: the low-quality and the high-quality sub-images are interleaved in the HEIF file. As well, for multi-layer sub-images, base layer and enhancement layers may not be organized to have all the base layers first, then followed by all enhancement layers. As a consequence, no complete reconstruction of the grid item is possible until the penultimate sub-image, "19", is received. The main image can be reconstructed soon after this when the last sub-image, "h9", is received.
Figure 4b represents a second ordering better suited to progressive rendering, where all the low-quality sub-images occur first in the HEIF file, followed by all the high-quality sub-images. With this ordering, ten progressive steps are possible, from using only low-quality sub-images to using only high-quality sub-images. A new progressive step is possible after the loading of each high-quality image. The same data ordering may be used for multi-layer images to provide first the base layers and then the enhancement layers. In this case, the sub-images used by the main image have dependencies to other sub-image items, the data for depended on sub-image items should be placed before the data for the sub-image item.
Figure 4c represents a third ordering also suited to progressive rendering with ten possible progressive steps, albeit using a different ordering for the high-quality sub-images. While the ordering of Figure 4b provides a progressive rendering row-by-row, the ordering of Figure 4c provides a progressive rendering column-by-column.
While the file structure described in relation to Figure 3 enables more progressive steps than the structure described in relation to Figure 2, it still has several drawbacks. First, a typical usage of a grid item may use a much larger number of input image items, resulting in many possible progressive steps. A larger number of progressive steps, while allowing for a smoother refinement of the image also requires more processing power, as the grid image has to be reconstructed at each step. Second, there is no guarantee that the ordering of sub-image versions in the HEIF file effectively enables a progressive rendering of the grid image. In addition, there is no indication on how this progressive rendering is organized, whether row by row, or column by column, or using another refinement pattern. An arbitrary refinement may lead to a bad user experience when looking at the image.
The invention proposes to extend HEIF by adding information about the organization of the file regarding progressive rendering. This information, hereafter named progressive information can take several forms that are detailed in the progressive embodiments described hereafter. This progressive information may include a progressive rendering strategy, or progressive strategy. Examples of progressive rendering strategy are the top-bottom ordering from Figure 4b or the left-right ordering from Figure 4c. This progressive information may also include the description of the progressive steps proposed by the file.
Figure 4d represents another example where the same ordering as in Figure 4b is used in relation with four different progressive steps: the initial progressive step containing only the low-quality sub-images, followed by three progressive steps, each adding a row of high-quality sub-images.
Figure 4f illustrates the rendering of the grid image item after the second progressive step, when three high-quality sub-images corresponding to the first row of the grid have been received. In this example, the content of the second progressive step is the three high-quality sub-images h1 to h3 and the six low-quality sub-images 14 to 19.
The refinements of this progressive step are the three high-quality sub-images h1 to h3.
Figure 4e represents another example where the same ordering as in Figure 4c is used in relation with four different progressive steps: the initial progressive step containing only the low-quality sub-images, followed by three progressive steps, each adding a column of high-quality sub-images.
Figure 4g illustrates the rendering of the grid image item after the second progressive step, when three high-quality sub-images corresponding to the first column of the grid have been received. In this example, the content of the second progressive step is the three high-quality sub-images hi, h4 and h7 and the six low-quality sub- 1 5 images 12, 13, 15, 16, 18 and 19. The refinements of this progressive step are the three high-quality sub-images h1, h4 and h7.
Figure 5 illustrates the main steps of generating an HEIF file according to embodiments the invention.
In a first step 500, the image data to be contained in the HEIF file is obtained. In the example of Figure 3, this image data consists of the nine low-quality sub-images and the nine high-quality sub-images. In addition, information about the structure of the HEIF file is obtained. This includes for example the type of the main image, its relations with the sub-images, and the relations between the sub-images. In the example of Figure 3, this information is that the main image is a grid item, composed of the nine sub-images h1 to h9, and that each of these sub-images h1 to h9 has an associated low-quality sub-image ranging from 11 to 19.
In the second step 510, the progressive specifications for organizing the file content for progressive rendering of the main image is obtained. These progressive specifications are dependent on the selected progressive embodiment. These progressive specifications may be obtained directly from a user input, or from a configuration file, or may be determined through some processing. For example, a software component may determine the type of content represented by the main image and select the most appropriate progressive rendering strategy. If the main image corresponds to a landscape picture, it may select a top-bottom progressive strategy, while if the main image corresponds to a person whose face is in the centre of the image, it may select a centre-to-border progressive strategy. As another example, when an object or region of interest is present, the progressive rendering strategy may start by refining first the area surrounding the object or region of interest and then progressing towards the image borders.
In the next step 520, the progressive information to be included in the H El F file is generated from the progressive specifications obtained at step 510. This generation depends on the selected progressive embodiment. These embodiments are described in detail below.
In the next step 530, the image data is ordered according to the progressive specifications. For example, if the progressive specifications indicate a top-bottom progressive rendering strategy as in the example of Figure 4b, then the image data will be ordered as shown in Figure 4b.
For this ordering step, the image data is organized into progressive data blocks.
These progressive data blocks are ordered according to the progressive specifications.
When the main image is a grid item, there is a progressive data block containing the data for the grid item itself. There is also one progressive data block for each version of each sub-image composing the grid. In the example of Figure 4b, there are nineteen progressive data blocks: the grid, the nine low-quality image data and the nine high-quality image data.
When the main image is an overlay item, there is a progressive data block containing the data for the overlay item itself, and one progressive data block for each version of each sub-image composing the overlay.
This may be transposed similarly for any kind of main image based on input sub-25 images.
When the main image is an HEVC image encoded with files or a VVC image encoded with sub-pictures, there may be one or more progressive data block for the image itself, and there may be one progressive data block for each version of each tile or sub-picture. When the sub-images are encoded as multi-layer images with several quality layers, for example a base layer and one or several enhancement layers, each layer corresponds to a progressive data block.
Possibly, a progressive data block may be further split into several sub-blocks. Preferably, the data block containing the data related to the main image structure is placed first. For example, the data for the grid item (or for an overlay) itself is preferably placed first.
This ordering may also cover data not directly related to the main image. For example, there may be a progressive data block corresponding to the thumbnail of the main image. This progressive data block is preferably ordered before any progressive data block for the main image.
There may be progressive data blocks corresponding to metadata associated to the main image. Preferably these data blocks are ordered after all the progressive data blocks for the main image. Possibly these data blocks may be inserted between two progressive steps, for example when it is useful to process or render some metadata information before having loaded the main image. Some progressive specifications may indicate where to place these progressive data blocks.
There may be progressive data blocks corresponding to region items associated to the main image. Preferably these data blocks are ordered after all the progressive data blocks for the main image. Possibly these data blocks may be inserted between two progressive steps, for example when it is useful to process or render some region-related information before having loaded the main image. Some progressive specifications may indicate where to place these progressive data blocks.
In the next step 540, the metadata for the HEIF file are written. These metadata include the progressive information generated at step 520.
In the last step 550, all the progressive data blocks are written, in the order determined at step 530, in the media part of the file, like for example an mdat box.
Possibly the ordering of steps 520 and 530 may be switched. This may be useful if the generating of the progressive information depends on the ordering of the progressive data blocks.
Figure 6 illustrates the main steps for a progressive rendering of an HEIF file generated according to an example of the invention. These steps are intended to be executed while loading the HEIF file. The HEIF file may be loaded from a local storage, from a network storage, or received from a network server, for example using the HTTP protocol.
In the first step 600, the start of the HEIF file is loaded. This start of the HEIF file includes all the metadata composing the 'meta' box of the HEIF file.
In step 610, the 'meta' box is parsed to obtain the structure of the HEIF file, including for example the number of items, the type of these items, their relations, the properties associated with these items and so on. This 'meta' box may contain an 'idat' box providing the data defining the overlay item or grid item structure. It may also contain a thumbnail.
In the next step 620, the progressive information is extracted from the result of the parsing of step 610. The refinements of the progressive information depends on the selected progressive embodiment. As a part of this extraction, the progressive steps contained in the HEIF file are obtained. This obtaining depends on the selected progressive embodiment.
In the next step 630, a part of the data present in the file is loaded.
At step 640, it is checked whether the data corresponding to a new progressive step has been loaded. This check depends on the selected progressive embodiment.
If the result of the check is positive, the next step is step 650, where a new rendering of the version of the main image corresponding to this progressive step is realized. This rendering may include computing a reconstructed image from the sub-image versions corresponding to the progressive step reached. It may also include some post-processing computed on the version of the reconstructed image, such as rescaling it for fitting the display area, or applying a filter selected by the user. It may also include applying some transformafive properties to the version of the reconstructed image, such as a rotation, a scaling, a mirroring or a crop.
Note that in some embodiments, a further check is realized based on the progressive information and the progressive step reached to determine if the version of the reconstructed image should be rendered. For example, if only part of the version of the main image is displayed, this further check may verify whether the progressive step reached provides refinements to the displayed part of the version of the reconstructed image.
If the result of the check at step 640 is negative and after step 650, then next step is step 660. This step checks whether the file has been fully received or not.
If the result of the check is negative, the next step is step 630, waiting for an additional part of the file to be received.
If the result of the check is positive, the next step is step 670 where the algorithm ends.
Note that at step 630, data corresponding to several refinement steps may be received. In this case, the check at step 640 will be positive and at step 650, the reconstructed image is rendered using the sub-images corresponding to the last progressive step reached.
Note that all these steps are intended for receiving the HEIF file in sequential order. However, they can easily be adapted to receiving the HEIF file as chunks in an arbitrary order. This adaptation mostly concerns steps 640 and 660. In these steps, the checks must take into account that the parts of the HEIF file may be received in an arbitrary order, verifying explicitly that all the data corresponding to a progressive step or to the whole file has been received and not taking advantage of a sequential reception of the HEIF file.
Different embodiments for encoding the progressive information in an HEIF file are now described.
Refinement pattern embodiment In a first embodiment, hereinafter called the refinement pattern embodiment, the progressive information contained in an HEIF file is based on a progressive refinement pattern that defines the progressive refinement strategy used in the HEIF file. A progressive refinement pattern may be for example the top-bottom progressive refinement of Figures 4d and 4f or the left-right progressive refinement of Figures 4e and 4g. The progressive refinement pattern is indicative of a scanning order of the input images. A progressive step is defined when a set of versions of the input images corresponding to a step in the scanning order of the input image has been loaded allowing the rendering of a version of the main image. In other words, the progressive refinement pattern is indicative of the ordering of the refinements for the different spatial regions of the main image. For example, in the case of a grid item, the progressive refinement pattern is indicative of the ordering of the refinements for the different sub-images that compose the grid.
This first embodiment advantageously provides a compact signalization for the progressive information inside the HEIF file.
The progressive refinement pattern used inside an HEIF file may be exposed at the application level, allowing an application to select an HEIF file among several based on its progressive refinement pattern. For example, for the same main image, a first HEIF file may use a top-bottom progressive refinement and a second HEIF file may use a left-right progressive refinement. An application may select between these two files the one that is best suited to its needs.
Predefined refinement variant In a first variant of this first embodiment, hereinafter called the predefined refinement pattern variant, a list of progressive refinement pattern is predefined. The progressive information comprises an indication of one of the predefined progressive refinement patterns.
This variant advantageously allows a simple signalling of the progressive refinement strategy used in an HEIF file.
A progressive refinement pattern may be represented as an item property associated with the image item to which it applies. This progressive refinement pattern item property may be described by the following structure: aligned(8) class ProgressivePatternProperty { extends ItemFullProperty(ppat., version = 0, flags = 0) unsigned int(8) pattern_type; unsigned int(1) reverse flag; unsigned int(1) bidireciional_flag; unsigned int(1) parameter_flag; unsigned int(5) reserved; unsigned int(8) scale; if (parameter_flag == 1) { unsigned int(32) parameter; This progressive refinement pattern item property may be identified, for example, by the cppatffi 4cc. Another 4cc may be used.
In this structure, the pattern_type specifies the type of the refinement pattern used. Several refinement pattern types may be defined, including: top-bottom, left-right, top-left to bottom-right diagonal, bottom-left to top-right diagonal, centre to border, clockwise... In addition, specific refinement pattern types for an overlay may be defined, as for example front to back or largest to smallest.
As an example, the following values may be defined for the pattern_type: * When equal to 0, the refinement is done from top to bottom on a row basis; This guarantees that data for the input images are ordered on a row basis.
* When equal to 1, the refinement is done from left to right on a column basis; * When equal to 2, the refinement is done from top-left to bottom-right on a diagonal basis * When equal to 3, the refinement is done from centre to border; * When equal to 4, the refinement is done on an input image item basis in a counter clockwise order; * When equal to 5, the refinement is done on an input image item basis from front to back (i.e., according to the layering order of an overlay); * When equal to 6, the refinement is done on an input image item basis from largest item to smallest item.
The reverseflag indicates that the ordering defined by the refinement pattern is reversed. For example, the top-bottom refinement pattern becomes a bottom to top refinement pattern with the reverse flag. Possibly, this flag may be removed from the Progress ivePattern Property and be replaced with new refinement pattern types. For example, a bottom-top refinement pattern type could be added.
The bidirectional_flag indicates that the ordering starts from both ends defined by the refinement pattern. For example, the top-bottom refinement pattern becomes a refinement pattern where the progressive refinement starts from both the top and bottom rows and progress towards the centre row. This flag may be combined with the reverseflag, reversing the order obtained after applying the reordering associated with the bidirectional_flag. Possibly, this bidirectional_flag may be removed from the ProgressivePatternProperty and be replaced with new refinement pattern types. For example, a bottom and top to centre refinement pattern type could be added.
The parameterflag indicates whether a parameter value is present for the refinement pattern. If the value of this flag is 1, then the parameter value is present in the item property. The meaning of this parameter value depends on the refinement pattern type. For example, a clockwise refinement pattern may start by default from a position corresponding to midday on a clock continuing in a clockwise order. The parameter value for this refinement pattern may indicate a different starting position by indicating a different starting hour, or by giving a starting angle in degree. The parameter value may be present for all the refinement pattern types. It may also be absent for all the refinement pattern types. The presence of the parameter value may depend on the pattern type. The parameter value may be a set of values or a more complex structure.
The scale value indicates that the progressive steps defined by the refinement pattern type are grouped together into a single coarser step (denoted scaled progressive step). In such a case, only scaled progressive steps should be rendered by the reader and not each single progressive step. The scale value may indicate how many progressive steps are grouped together. For example, a scale value of 2 for a left-right refinement pattern type indicates that each scaled progressive step's refinements contain the refinement for two columns at once. The scale value may also be a fractional value. In this case, the scaled progressive step n for the refinement pattern specified by the item property corresponds to the progressive step number In x scale] as defined by the refinement pattern type, i.e., the rounded-down value of the index n of the progressive step multiplied by the scale value. In this formula, the initial step corresponding to the low-quality version of the sub-images is numbered 0. For example, with a scale value of 1.5, a top-bottom ordering alternates scaled progressive steps adding 1 or 2 rows of high-quality images. Possibly, this scale value may be removed from the ProgressivePatternProperty and be replaced with new refinement pattern types. For example, there may be a top-bottom by 2 rows refinement pattern type. In a variant, the scale value may be replaced by a list of numbers indicating how many progressive steps are grouped into each scaled progressive step. For example, if the list has the values [2, 1, 3] for a top-bottom refinement pattern, then the first scaled progressive step's refinements contain 2 rows of high-quality sub-images, the next one 1 row, and the last one 3 rows. If there are more progressive steps than the sum of the values contained in the list, the list may be repeated to continue grouping the progressive steps, the remaining progressive steps may be grouped using a default group size, or the remaining progressive steps may be kept alone.
In this embodiment, in step 510 of Figure 5, the obtained progressive specifications include the pattern type and may include a flag indicating whether a reverse ordering is used, a flag indicating whether a bidirectional ordering is used, a scaling value for the progressive steps, and a parameter value associated with the pattern type.
In step 520 of Figure 5, an item property structure is generated for representing all these specifications. This item property structure is written as a part of the 'meta' box of the HEIF file at step 540 and is associated with the image item corresponding to the main image in the HEIF file. For example, the item property structure is written as part of the item property container box I pco' and is associated with the image item corresponding to the main image via an item property association box (ipma'.
In step 530 of Figure 5, the ordering defined by the pattern type is retrieved and applied to the progressive data blocks. This ordering may be modified by the reverse flag and the bidirectional flag.
In this embodiment, in step 620 of Figure 6, the progressive information for an image item is extracted by parsing the progressive refinement pattern property item associated with this image item. Using the pattern type specified in the progressive refinement pattern property item and the different flags and values contained in this property item, the list of progressive steps is computed. Each progressive step is specified by the list of sub-image versions it adds to the previous progressive step. For example, in the case of Figure 4f, the pattern is a top-bottom pattern. There is three progressive steps: one for each row of the grid item. And each progressive step contains three sub-image versions, corresponding to the three columns of the grid item.
At step 640, it is checked whether all the sub-image versions contained in the list associated with a progressive step (or a scaled progressive step when the scale value is present and not equal to 1) have already been loaded.
Possibly, no progressive step is computed at step 620. In this case, the check of step 640 may be based on the number of sub-image versions received. For example, the check may be positive each time a new predefined number of new sub-image versions has been received, or when the number of new sub-image versions received is a predefined percentage of the total number of sub-image versions. Possibly low-quality sub-image versions may not be taken into account in these checks.
The check may also be based on the number of bytes received. For example, the check may be positive each time a new predefined number of bytes has been received, or when the number of newly received bytes is a predefined percentage of the total number of bytes of the HEIF file. Possibly, bytes corresponding to the content of low-quality sub-image versions and/or the content of the HEIF file structure, such as the 'meta' box may not be taken into account in these checks.
In a variant, the scale value may be replaced by a scalefraction value indicating how many progressive steps are grouped together into scaled progressive steps as a fraction of the total number of progressive steps.
In another variant, the scale value may be replaced by a scale_number value indicating that the progressive steps are grouped together into a number of scaled progressive steps equal to this scale_number value.
Analytic refinement patterns In a second variant of the first embodiment, hereinafter called the analytic refinement patterns, progressive refinement patterns are defined in an analytic way. In this embodiment, the progressive information comprises an indication of a scanning order of the input images based on an equation and associated parameters.
This variant advantageously enables a rich expressivity in the possible progressive refinement patterns used in HEIF files.
An analytic refinement patterns may be defined inside the progressive refinement pattern item property. The structure of this progressive refinement pattern item property may be: aligned(8) class ProgressivePatternProperty extends ItemFullProperty('ppat', version = 0, flags = 0) { unsigned int(2) pattern_type; unsigned int(1) bidirectional_flag; unsigned int(5) reserved; 77 Linear if (pattern_type == 0) unsigned int(8) start_x; unsigned int(8) start_y; unsigned int(8) end_x; unsigned int(8) end_y; unsigned int(16) step_numerator; unsigned int(16) step_denominator; // Radial else if (pattern_type == 1) { unsigned int(8) center_x; unsigned int(8) center_y; int(16) step_numerator; unsigned int(16) step_denominator; // Circular else if (pattern_type == 2) { unsigned int(8) center_x; unsigned int(8) center_y; unsigned int(16) starting_angle; int(16) step_numerator; unsigned int(16) step_denominator; In this structure, the pattern_type indicates the general type of the refinement pattern, for example whether it is linear, radial or circular.
The bidirectional_flag indicates that the ordering starts from both ends defined by the refinement pattern.
If the pattern_type value is 0, then the refinement pattern is a linear refinement pattern. A linear refinement pattern may be for example a top-bottom refinement pattern, or a left-right refinement pattern.
The start_x and start_y values indicate the starting point of the refinement pattern, which is the centre of the grid item whose coordinates in the grid are given by these values.
Similarly, the end_x and end_y values indicate the ending point of the refinement pattern, which is the centre of the grid item whose coordinates in the grid are given by these values.
Last, the step_numerator and step_denominator values indicate the size (equal to step_numerator divided by step_denominator) of each progressive refinement step.
To determine to which progressive step the high-quality version of a sub-image of the grid item belongs to, the following computation is realized using grid coordinates. First, the centre of the grid item is orthogonally projected on the line defined by the starting and ending point. The distance of this projected point from the starting point is computed. This distance is considered positive if the projected point belongs to the half line also containing the ending point, and negative in the other case. This distance is then divided by the size of the progressive refinement steps, rounded up and increased by one to obtain the index of the progressive step to which this grid item belongs to. If the computed index is lower than one, then the grid item belongs to the first progressive step. If the computed index is greater than the index of the ending point, then the grid item belongs to the last progressive step.
The computation of the index of the progressive step of a grid item can be summarized by the following formula: is - PiG PiPf 11Prit-11 step_numerator + 1 step_denominator Where Pi is the starting point of the refinement pattern, 191 is the ending point of the refinement pattern, and G is the centre point of the grid item.
If the pattern_type value is 1, then the refinement pattern is a radial refinement pattern. A radial refinement pattern may be for example a centre-border refinement pattern.
The center_x and center_y values indicate the starting point of the radial refinement pattern, as the centre of the grid item at the given position in the grid.
Last, the step_numerator and step_denominator values indicate the size of each progressive refinement step. The step_numerator is a signed integer, allowing the size to be either positive or negative. A positive size indicates that the progressive steps go outwards from the centre of the refinement pattern while a negative size indicates that the progressive steps go inwards towards the centre of the refinement pattern.
Possibly, the step_numerator is an unsigned integer and a flag is added for signalling whether the progressive steps go outwards or inwards.
To determine to which progressive step a grid item belongs to, the following computation is realized using grid coordinates. The distance between the centre of the grid item and the centre of starting point of the radial refinement pattern is computed.
This distance is divided by the value of the size, rounded up and increased by one to obtain the relative index of the progressive step to which this grid item belongs to. Once all the relative indexes for the progressive step of all the grid items have been computed, the absolute indexes are computed by subtracting to the relative indexes the value of the smallest relative index.
The computation of the index of the progressive step of a grid item can be summarized by the following formulae 11C step_numerator +1 step_denominator ic = riG± mdnric Where C is the starting point of the radial refinement pattern and G is the center point of the grid item.
If the pattern_type value is 2, then the refinement pattern is a circular refinement pattern. A circular refinement pattern may be for example a clockwise refinement pattern.
The center_x and center_y values indicate the reference point of the circular refinement pattern, as the centre of the grid item at the given position in the grid.
The starting_angle value indicate the starting direction of the circular refinement pattern, where the value 0 corresponds to the direction towards the top of the image, and the starting_angle value is expressed in degrees with a clockwise 20 orientation.
Last, the step_numerator and step_denominator values indicate the size of each progressive refinement step. The step_numerator is a signed integer, allowing the size to be either positive or negative. A positive size indicates that the progressive steps go clockwise around the reference point, while a negative size indicates that the progressive steps go counter clockwise around the reference point.
Possibly, the step_numerator is an unsigned integer and a flag is added for signalling whether the progressive steps go clockwise or counter clockwise.
To determine to which progressive step a grid item belongs to, the following computation is realized using grid coordinates. The angle between the starting direction and the vector from the reference point to the centre point of the grid item is computed.
This angle is divided by the value of the size, rounded up and increased by one to obtain the relative index of the progressive step to which this grid item belongs to. Once all the relative indexes for the progressive step of all the grid items have been computed, the riG absolute indexes are computed by subtracting to the relative indexes the value of the smallest relative index.
The computation of the index of the progressive step of a grid item can be summarized by the following formulae: (d, RG) ric step_numerator +1 step_denominator iG = riG -npnriG Where a is a vector corresponding to the starting direction, R is the reference point of the circular refinement pattern and G is the center point of the grid item.
Other coordinates systems may be used instead of grid coordinates. For example, pixel coordinates may be used, coordinates linked to encoding blocks may be used, or coordinates corresponding to a predefined number of pixels may be used. Possibly the horizontal unit and vertical unit may correspond to a different number of pixels.
Possibly, the values contained in the progressive refinement pattern item property may be encoded on a larger number of bits. This may be useful for example when using pixel coordinates. For example, the start _x and start_y values may be encoded on 16 or 32 bits.
Possibly, the start_x, starts, end_x, end_y, center_x and centers values may be encoded as signed integers to allow these points to fall outside the grid itself. Possibly other refinement pattern types may be defined. For example, a pattern for providing a Venetian blind effect may be defined.
Possibly some flags may be added to the refinement pattern to define more precisely the position of the different points used. For example, some flags may be added to specify whether a point is in the middle of the corresponding grid item, or at one of its corner. Possibly a 2-bit flag may be used for enabling to position each point on one of the four locations of a grid item: top-left corner, centre, middle of the left border and middle of the top border.
Multiple-level variant In a variant of these previously described embodiments, sub-images have more than two different versions. For example, each sub-image may have a low-quality version, a medium-quality version and a high-quality version, or even more intermediate versions.
In this variant, the same refinement pattern may be used repeatedly for describing all the transitions between the different quality levels. For example, the same refinement pattern may be used to define the progressive steps between low-quality sub-images and medium-quality sub-images and also to define the progressive steps between medium-quality sub-images and high-quality sub-images.
For example, in Figure 4f, using a top-bottom progressive refinement pattern, there are three progressive steps corresponding to the three rows of the grid item. With three versions for each sub-image, this would result in seven progressive steps in a sequential progressive refinement of the rows: an initial progressive step with only the low-quality sub-images, followed by three progressive steps each adding the medium-quality sub-images for one row, followed by three other progressive steps each adding the high-quality sub-images for one row.
Possibly different refinement patterns may be used for describing the transitions between the different versions of the sub-images. For example, a first refinement pattern may be used to define the progressive steps between low-quality sub-images and medium-quality sub-images and a second refinement pattern may be used to define the progressive steps between medium-quality sub-images and high-quality sub-images.
For allowing several refinement patterns, the progressive refinement pattern item property may include the number of refinement patterns it contains and a loop for describing each of these patterns.
Possibly specific multi-level refinement patterns may be defined, to allow mixing the different versions of the sub-images. For example, in a centre-to-border refinement pattern, the high-quality version of the centre sub-images could be used in a progressive step before the medium-quality version of the border sub-images.
In the predefined refinement pattern variant, new multi-level refinement patterns may be defined.
Possibly, a step_offset parameter may be added to the progressive refinement pattern description, to indicate at which progressive step a new quality level for the versions of the sub-images is introduced. In the example of Figure 4f, with a step_off set parameter with a value of 2, the high-quality sub-images for the first row are contained in the same progressive step as the medium-quality sub-images for the third row: refining the first row with the high-quality sub-images is realized 2 progressive steps after refining this first row with the medium-quality sub-images.
Possibly, a version_splitilag parameter may be added to the progressive refinement pattern description, preferably in combination with the step_off set parameter. If the value of this flag is true, then the versions of the sub-images corresponding to the new quality level are contained in the refinements of a separate progressive step, before the progressive step whose refinements contain the versions of the sub-images corresponding to the current quality level. In the example of Figure 4f, with a step_off set parameter with a value of 2, after the progressive step whose refinements contain the medium-quality version of the second row, the next progressive step's refinements contain only the high-quality version of the first row, and is followed by another progressive step whose refinements contain the medium-quality version of the third row.
Possibly, a refinement pattern may be defined for the lowest-quality version of the sub-images. In the example of Figure 4f, the top-bottom refinement pattern may be applied to the low-quality version of the sub-images and the first progressive step may contain only the low-quality version of the sub-images in the first row of the grid. Incomplete Level Variant In a variant, some versions of the sub-images may be missing. In the example of Figure 4f, the high-quality version of the bottom-right sub-image, h9, may be missing, or the low-quality version of the bottom-left sub-image, 17, may be missing.
Possibly, these missing sub-images are just ignored. If a progressive step's refinements contain only missing sub-images, it is just skipped during the progressive rendering of the main image.
Possibly, these missing sub-images are indicated by adding a list of missing sub-images to the progressive refinement pattern item property.
Possibly, if the number of missing sub-image versions contained in the refinements of a progressive step is below a predefined threshold, then the progressive step is still considered as loaded at step 640. This threshold may be defined as a number of sub-image versions. It may also be defined as a percentage of the total number of sub-image versions contained in the refinements of the progressive step (for example, 5% or 10% of the sub-image versions). Possibly, a progressive step may be considered as loaded while some sub-image versions contained in its refinements are missing only when the elapsed time since the previous progressive step has been loaded is greater than or equal to a predefined threshold.
When generating an HEIF file using the steps of Figure 5, the progressive specifications may include a list of sub-images that are not to be included in the resulting HEIF file.
Multi-Level Pattern In a variant the progressive refinement order inside a progressive step may be further specified by another progressive refinement pattern. For example, a top-bottom progressive refinement pattern indicates that the progressive refinement of the main image is realized row by row. A second progressive refinement pattern may be used to indicate that inside each row, the progressive refinement starts from the centre of the row and goes towards its borders.
This second progressive refinement pattern may be specified inside the progressive refinement pattern item property, using for example the following structure: aligned(8) class ProgressivePatternProperty extends ItemFullProperty('ppat', version = 0, flags = 0) { unsigned int(8) main_pattern_type; unsigned int(1) main_reverse_flag; unsigned int(1) main_bidirectional_flag; unsigned int(1) main_parameter_flag; unsigned int(5) main_reserved; unsigned int(8) main_scale; if (parameter_flag == 1) f unsigned int(32) main_parameter; unsigned int(8) secondary_pattern_type; unsigned int(1) secondary_reverse_flag; unsigned int(1) secondary_bidirectional_flag; unsigned int(1) secondary_parameter_flag; unsigned int(5) secondary_reserved; unsigned int(8) secondary_scale; if (parameter_flag == 1) { unsigned int(32) secondary_parameter; Several second progressive refinement patterns may be specified for the different progressive steps.
The second progressive refinement patterns may have a different structure than the main progressive refinement patterns.
Offset embodiment In a second embodiment of the invention, hereinafter called the offset embodiment, the progressive steps are indicated using positions in the HEIF file.
Advantageously, this second embodiment provides an explicit and simple indication of the progressive steps for an HEIF file.
Some information about these progressive steps may be exposed at the application level, allowing an application to select an HEIF file among several based on its number of progressive steps.
Step location variant In a first variant of this second embodiment, the step location variant, for each progressive step, the location of the last byte in the HEIF file that is part of the last sub-image version contained in this progressive step's refinements is specified in the progressive information stored in the HEIF file.
The positions of the progressive steps for an image item inside the HEIF file may be indicated with an item property associated with this image item. This progressive step location item property may be described by the following structure: aligned(8) class ProgressiveStepLocationProperty extends ItemFullProperty('ploc', version 0, flags 0) { unsigned int(16) step_count; for (i=0; i < step_count; i++) { if (flags & 1) { unsigned int(32) item_index; else { unsigned int(16) item_index; unsigned int(16) extent_index; 1 Different sizes may be used for the fields of this progressive step location item property. Possibly some flags, either using the flags parameter for the ItemFullProperty or flag values directly encoded inside the progressive step location item property, may be used to specify the size of the different fields.
In this structure, the step_count indicates the number of progressive steps described inside the progressive step location item property.
For each progressive step, the item_index value indicates the index of a resource inside the array in the item location box, 'floc'. The extent_index value indicates the index of an extent for this resource. The location of the progressive step is the last byte of this indicated extent.
Possibly the extent_index is an optional field. If it is not present, then the location of the progressive step is the last byte of the last extent of the item indicated by the item_index value.
Possibly the progressive step location item property doesn't contain an
extent_index field.
In this embodiment, in step 510 of Figure 5, the progressive specifications include some information for defining the progressive steps. This information may include for example the list of sub-image versions contained in each progressive step refinements. In this embodiment the ordering of step 520 and step 530 is switched.
First, the ordering of the progressive data blocks is realized in step 530 using the definition of the progressive steps. Using this ordering, the structure of the HEIF file is created and the content of the floc' box is generated.
Then, step 520 is executed, generating a progressive step location item property based on the content of the floc' box and the definition of the progressive steps. This item property is written as a part of the 'meta' box of the HEIF file at step 540 and is associated with the image item corresponding to the main image in the HEIF file. For instance, the item property structure is written as part of the item property container box cipco' and is associated with the image item corresponding to the main image via an item property association box ipma'.
In this embodiment, in step 620 of Figure 6, the progressive information is extracted by parsing the progressive step location item property. This progressive information is used to generate the list of progressive steps for the main image. Each progressive step is specified by its location that is the last byte of the extent indicated in the progressive step location item property.
Possibly, if the extent_index field is not present, the location of the progressive step is the last byte of the last extent of the item indicated in the progressive step location item property.
At step 640, the location of the progressive steps is compared to the last loaded byte of the HEIF file. If the location of the progressive step is before the position of the last loaded byte, then the progressive step has been loaded.
If the progressive steps are ordered according to their location, only the first progressive step not already loaded needs to be checked at step 640.
Preferably, the progressive steps are ordered in the progressive step location item property according to their location order inside the HEIF file. If this is not the case, the progressive steps can be reordered according to their location after having been extracted from the progressive step location item property.
Possibly in this variant, an item may be identified by its identifier instead of its index inside the array in the item location box.
Step refinement variant In a second variant, the step refinement variant, each progressive step is specified by the list of the extents containing data for the sub-image versions contained in the progressive step's refinements.
The lists of extents composing the different progressive steps' refinements of an image item may be indicated with an item property associated with this image item. This progressive step refinement item property may be described by the following structure: aligned(8) class ProgressiveStepRefinementProperty extends ItemFullProperty('prco', version 0, flags = 0) { unsigned int(16) step_count; for (i=0; i < step_count; i++) { unsigned int(10 item_count; for (j=0; j < item_count; j++) { if (flags & 1) { unsigned int(32) item_index; else { unsigned int(16) item_index; 1 unsigned int(16) extent_count; for (k=0; k < extent_count; k++) { unsigned int(16) extent_index; 1 Different sizes may be used for the fields of this progressive step refinement item property. Possibly some flags, either using the flags parameter for the itemFullProperty or flag values directly encoded inside the progressive step refinement item property, may be used to specify the size of the different fields.
In this structure, the step_count indicates the number of progressive steps described inside the progressive step refinement item property.
For each progressive step, the item_count value indicates how many items are comprised in the progressive step' refinements. For each item, the item_index value indicates the index of a resource inside the array in the item location box, 'floc'.
For each item, the extent_count value indicates how many extents are comprised in the progressive step's refinements for this item. For each extent, the extent_index value indicates the index of an extent for this resource.
The generation of an HEIF file using this variant is similar to the one for the 'floc' references.
In this variant, in step 620 of Figure 6, each progressive step is specified by the list of extents that are part of its refinements.
At step 640, it is checked whether all the extents belonging to the refinements of the progressive step have been loaded.
This variant may be relevant for image items that are split into spatial parts like tiles or sub-pictures or even layers, where each spatial part has its data identified by an extent in the 'floc' box.
Possibly in this variant, an item may be identified by its identifier instead of its index inside the array in the item location box.
Possibly, the progressive step refinement item property is described by the following structure: aligned(8) class ProgressiveStepRefinementProperty extends ItemFullProperty('prco', version 0, flags = 0) unsigned int(16) step_count; for (i=0; i < step_count; i++) { unsigned int(16) item_count; for (j=0; j < item_count; j++) { if (flags & 1) { unsigned int(32) item_index; else { unsigned int(16) item_index; unsigned int(16) extent_count; In this variant, the extent_count value indicates how many extents for the item are comprised in the progressive step's refinements or in previous progressive steps' refinements. In this variant, for an item, extents are organized in decoding order. Possibly, the extents that are part of the progressive step's refinements are computed by comparing the list of extents specified for this progressive step in the progressive step refinement item property with the extents present in previous progressive steps' refinements. These extents are used at step 640 to check whether the progressive step has been loaded.
Possibly, the check of step 640 is modified to check whether all the extents specified for this progressive step in the progressive step refinement item property have been loaded.
Possibly, in this variant, the extent_count field is optional. If it is absent, all the extents for the item are comprised in the progressive step's refinements or in previous progressive steps' refinements.
In a variant, the specification of a progressive step may use pointers into the HEIF file. This variant may be used either with the step location variant or with the step refinement variant.
Possibly, these pointers may be absolute offsets into the HEIF file.
For example, with the step location variant, for each progressive step, the location of the last byte that is part of a sub-image version contained in this progressive step is specified directly as an offset into the HEIF file.
As another example, with the step refinement variant, for each progressive step, the locations of the sub-image versions contained in the progressive step refinements is specified as a list of offsets into the HEIF file, each offset being associated with a length.
Possibly, these pointers may be specified using a mechanism similar to the one used by the Item Location Box. Possibly, only part of the mechanism used by the Item Location Box may be used. For example, these pointers may be specified as offset into the ItemDataBox idat' or into the IdentifiedMediaDataBox lloc'-based Variant In a variant, the specification of a progressive step is realized in a new version of the Item Location Box.
In this variant, a flag is added to each extent to indicate whether the last byte of this extent is the last byte of a sub-image version contained in a progressive step refinements. In this way, the progressive steps may be reconstructed from the information contained in the Item Location Box. This flag may be named progressive_step_end.
Possibly, this flag may be added to each item instead of to each extent.
Possibly, a progressive step number field is added to each extent to indicate into which progressive step's refinements this extent is contained. The value 0 may be reserved to indicate that the extent is not contained in any progressive step's refinements. This progressive step number field may be named progressive_step_number. Possibly a flag may be used to indicate the presence of this field.
Possibly, a progressive step number end field is added to each extent to indicate that the last byte of this extent is the last byte of a sub-image version contained in the corresponding progressive step's refinements. The value 0 may be reserved to indicate that the last byte of the extent is not the last byte of a sub-image version contained in a progressive step's refinements. This progressive step number field may be named progressive_step_number_end. Possibly a flag may be used to indicate the presence of this field.
Item Embodiment In a third embodiment of the invention, hereinafter called the item embodiment, the progressive steps are indicated using indications of the sub-image versions they contain in their refinements.
Advantageously, this third embodiment provides a simple link between the sub-images composing a main image and the progressive steps for this main image. Advantageously some variants of this third embodiment require only a minimal signalizafion in the HEIF file.
In a first variant of this third embodiment, hereinafter called the version list variant, for each progressive step, the list of sub-image versions contained in this progressive step's refinements is specified in the progressive information stored in the HEIF file.
The sub-image versions contained in the progressive steps' refinements for an image item inside the HEIF file may be indicated with an item property associated with this image item. This progressive version list item property may be described by the following structure: aligned(8) class ProgressiveVersionListProperty extends ItemFullProperty('pvli', version = 0, flags = 0) { unsigned int(16) step_count; for (i=0; i < step_count; i++) { unsigned int(16) item_count; for (j=0; j < item_count; j++) { if (flags & 1) { unsigned int(32) item_ID; else { unsigned int(16) item_ID; In this structure, the step_count indicates the number of progressive steps described inside the progressive version list item property.
For each progressive step, the item_count value indicates the number of sub-image versions contained in this progressive step's refinements.
For each sub-image version, the item_ID value indicates the identifier of the image item corresponding to this sub-image version.
Inthisembodiment in step 510 ofFigure5, the progressive specifications include some information for defining the progressive steps. This information may include for example the list of sub-image versions contained in each progressive step's refinements.
In step 520 of Figure 5, a progressive version list item property is generated based on the definition of the progressive steps. This item property is written as a part of the meta' box of the FilEIF file at step 540 and is associated with the image item corresponding to the main image in the 1-11EIF file. For instance, the item property structure is mitten as partofthe item propertycontainerbox cipco' and is associated with the image item corresponding to the main image via an item property association box (ipma'.
In this embodiment, in step 620 of Figure 6, the progressive information is extracted by parsing the progressive version list item property. This progressive information is used to generate the listofprogressive steps for the main image. Each progressive step is specified by the list of sub-image versions it contains in its refinements.
Alstep640,ffis checkedwhetherallthedataforthe image items specifiedinthe list of sub-image versions contained in a progressive step's refinements have already been loaded.
Possibly, a sub-image may be encoded as a multklayerimage described by a single image item. The different versions of the sub-image are encoded as different layersofthisimage item. In this case, all the versionsofthesub-image may be described by a single image item. The progressive version list item property may be modified to indicate the encoding layer corresponding to a version of a sub-image. This modified progressive version list item property may be described bythefollowing structure: aligned(8) class ProgressiveVersionListProperty extends ItemFullProperty('pvlir, version = 0, flags = 0) { unsigned int(16) step_count; for (1=0; i < step_count; i++) { unsigned int(16) item_count; for (j=0; j < item_count; j++) { if (flags & 1) { unsigned int(32) item_ID; else { unsigned int(16) item_ID; unsigned int(1) has_layer; if (has_layer == 1) { unsigned int(15) layer_id; 40} else { unsigned int(7) reserved; In this structure, the has_layer value is a flag indicating whether the version of the sub-image corresponds to a specific encoding layer of the image item with the given item_ID. If the value of this has_layer flag is true, the layer_id value indicates the identifier of the encoding layer corresponding to the version of the sub-image.
Possibly, the different layers contained in an image item corresponding to a sub-image may be described by an item property associated to this image item. This progressive layer item property may be described by the following structure: aligned(8) class ProgressiveLayerProperty extends ItemFullProperty( 'play', version = 0, flags = 0) { unsigned int(8) layer_count; for (i=0; i < layer_count; i++) unsigned int(16) layer_id; In this structure, the layer_count value indicates the number of layers used for the progressive rendering. The total number of layers present in the image item may be greater than this indicated number of layers.
For each layer, the layer_id value indicates the identifier of this layer.
The progressive version list item property structure may be modified to take advantage of this progressive layer item property: aligned(8) class ProgressiveVersionListProperty extends ItemFullProperty('pvli', version = 0, flags = 0) { unsigned int(16) step_count; for (1=0; i < step_count; i++) { unsigned int(16) item_count; for (j=0; j < item_count; j++) { if (flags & 1) { unsigned int(32) item_ID; else { unsigned int(16) item_ID; 35} unsigned int(8) layer_index; In this structure, the layer_index value indicates the 1-based index of the layer inside the progressive layer item property associated to the image item corresponding to the sub-image version and specified by the value of item_ID. The value 0 is reserved for indicating that the sub-image version corresponds to the full content of the image item.
Possibly, the progressive layer item property may be replaced by several ls el' layer selection item properties, each lsel' layer selection item property indicating one of the layer contained in the image item. In this case, the layer_index value indicates the 1-based index of the rlsel' item property that is associated to the image item corresponding to the sub-image and specified by the value of item_ID and that indicates the layer corresponding to the sub-image version.
For these two variants, at step 640 of Figure 6, it is checked for each image item specified in the list of sub-image versions contained in a progressive step's refinements whether the corresponding sub-image version has been loaded. If there is no layer identifier specified for this image item and the whole data for this image item have been loaded, then the corresponding sub-image version has been loaded. If there is a layer identifier specified for this image item and the data for this layer and for all the previous layers have been loaded, then the corresponding sub-image version has been loaded.
Possibly, the identifiers for the sub-image version's items are allocated in the progressive refinement order. The progressive version list item property may be described by the following structure: aligned(8) class ProgressiveVersionListProperty extends ItemFullProperty('pvli', version = 0, flags = 0) { unsigned int(16) step_count; for (i=0; i < step_count; i++) { if (flags & 1) { unsigned int(32) item_ID; else { unsigned int(16) item_ID; In this structure, the item_ID value indicates the identifier of the last image item corresponding to a sub-image version contained in the refinements of the progressive step Item List Variant In a second variant of this third embodiment, the item list variant, for each progressive step, the list of sub-images for which a version is contained in this progressive step's refinements is specified in the progressive information stored in the HEIF file.
While the first variant indicates explicitly the sub-image versions contained in the refinements of a progressive step, this second variant indicates only the sub-images for which a version is contained in the refinements of a progressive step. This means that the renderer has to determine for each sub-image indicated in relation to a progressive step which of the sub-image versions is contained in the refinements of this progressive step The sub-images for which a version is contained in the progressive steps' refinements for an image item inside the HEIF file may be indicated with an item property associated with this image item. This progressive item list item property may be described by the following structure: aligned(8) class ProgressiveItemListProperty extends ItemFullProperty('pilir, version = 0, flags = 0) { unsigned int(16) step_count; for (1=0; i < step_count; i++) { unsigned int(16) item count; for (j=0; j < item_count; j++) { if (flags & 1) 1 unsigned int(32) item ID; 1 else { unsigned int(16) item_ID; 1 In this structure, the step_count indicates the number of progressive steps described inside the progressive item list item property.
For each progressive step, the item count value indicates the number of sub-images for which a version is contained in this progressive step's refinements.
For each sub-image, the item_ID value indicates the identifier of the image item for which a version is contained in this progressive step's refinements.
In this embodiment, in step 510 of Figure 5, the progressive specifications include some information for defining the progressive steps. This information may include for example the list of sub-image versions contained in each progressive step's refinements.
In step 520 of Figure 5, a progressive item list item property is generated based on the definition of the progressive steps. This item property is written as a part of the meta' box of the HEIF file at step 540 and is associated with the image item corresponding to the main image in the HEIF file. For instance, the item property structure is written as part of the item property container box ipco' and is associated with the image item corresponding to the main image via an item property association box (ipma'.
Preferably, for each sub-image that is a component of the main image, the list of its versions is explicitly included as a part of the meta' box of the HEIF file. If the versions are based on encoding layers of the sub-image, then a progressive layer item property or a set of layer selection item properties, lsel' , may be used to list the versions of this sub-image. Otherwise, an alt r' entity group may be used to list all the image items corresponding to versions of this sub-image. The highest-quality version of the sub-image should be listed first in this alt r' entity group.
In this embodiment, in step 620 of Figure 6, the progressive information is extracted by parsing the progressive item list item property. This progressive information is used to generate the list of progressive steps for the main image. Each progressive step is specified by the list of sub-images for which a version is contained in the progressive step's refinements.
At step 640, it is checked whether a new version is available for all the sub-images contained in the list associated to a progressive step.
As a variant, in step 620, the list of versions for each sub-image may be computed. Preferably, this list of versions is extracted from an explicit description in the HEIF file. If a sub-image version is an image item referenced by an 'aft r' entity group, then the list of all versions of the sub-image is the list of image items contained in this (altr' entity group. If a sub-image version is an image item with an associated progressive layer item property, then the list of all versions of the sub-image is the list of encoding layers indicated in this progressive layer item property. If a sub-image version is an image item referenced by an caltr' entity group and with an associated progressive layer item property, then the list of all versions of the sub-image is the combination of the lists for both cases.
Possibly, the list of versions for a sub-image is inferred from other structures from the 'meta' box of the HEIF file. For example, if a sub-image version is an image item that is associated with a thumbnail, then this thumbnail is a version of the sub-image.
Once this list of versions for each sub-image has been computed, these versions are ordered in a progressive refinement order, from the lowest quality to the highest quality and associated to the progressive step. For each sub-image, its versions are associated in order to the progressive steps that list this sub-image.
In this variant, at step 640, it is checked whether all the sub-image versions contained in a progressive step's refinements have already been loaded. If a sub-image version corresponds to an encoding layer of a multi-layer image item, then it is checked whether the data for this encoding layer and for all the previous encoding layers have been loaded. Otherwise, it is checked whether the data for the image item have been loaded.
Item reference variant A third variant of this third embodiment, hereinafter called the item reference variant, is similar to the second variant. Instead of storing the list of sub-images for which a version is contained in a progressive step's refinements in an item property, this list is stored in an entry of the Item Reference Box.
The Item Reference Box entry may have the prli' 4cc. The f rom_item_ID field of the entry is the identifier of the main image and the sub-images are specified using the to_item_ID fields of the entry. This reference type indicates for a main image the list of image items that are contained in the refinements of a progressive step. There is an Item Reference Box entry of type priii for each progressive step and these entries are ordered according to the progressive steps order.
Item Count Variant In a fourth variant of this third embodiment, hereinafter called the item count variant, for each progressive step, the number of sub-image versions contained in this progressive step's refinements is specified in the progressive information stored in the HEIF file.
This number of sub-image versions contained in the progressive step's refinements for an image item inside the HEIF file may be indicated with an item property associated with this image item. This progressive item count item property may be described by the following structure: aligned(8) class ProgressiveItemCountProperty extends ItemFullProperty('pcntr, version = 0, flags 0) { unsigned int(16) step_count; for (1=0; i < step_count; i++) { unsigned int(16) item_count; 25} In this structure, the step_count indicates the number of progressive steps described inside the progressive item count item property.
For each progressive step, the item_count value indicates the number of sub-images for which a version is contained in this progressive step's refinements.
In this embodiment, in step 510 of Figure 5, the progressive specifications include some information for defining the progressive steps. This information may include for example the number of sub-image versions contained in each progressive step's refinements.
In step 520 of Figure 5, a progressive item count item property is generated based on the definition of the progressive steps. This item property is written as a part of the 'meta' box of the HEIF file at step 540 and is associated with the image item corresponding to the main image in the HEIF file. For instance, the item property structure is written as part of the item property container box ipco' and is associated with the image item corresponding to the main image via an item property association box iprna' . Preferably, for each sub-image that composes the main image, the list of its versions is explicitly included as a part of the 'meta' box of the HEIF file. If the versions are based on encoding layers of the multi-layer sub-image, then a progressive layer item property may be used to list the versions of this multi-layer sub-image. Otherwise, an alt r' entity group may be used to list all the image items corresponding to versions of this sub-image. The highest-quality version of the sub-image should be listed first in this altr' entity group.
At step 640, the number of sub-images for which a new version has been loaded since the previous progressive step is counted. Then it is checked whether this number is greater than or equal to the number associated with the current progressive step.
As a variant, in step 620, the list of versions for each sub-image may be computed. Preferably, this list of versions is extracted from an explicit description in the HEIF file. If a sub-image version is an image item referenced by an 'aft r' entity group, then the list of all versions of the sub-image is the list of image items contained in this (altr' entity group. If a sub-image version is an image item with an associated progressive layer item property, then the list of all versions of the sub-image is the list of encoding layers indicated in this progressive layer item property. If a sub-image version is an image item referenced by an caltr' entity group and with an associated progressive layer item property, then the list of all versions of the sub-image is the combination of the lists for both cases.
Possibly, the list of versions for a sub-image is inferred from other structures of the meta' box of the HEIF file. For example, if a sub-image version is an image item that is associated with a thumbnail, then this thumbnail is a version of the sub-image. Once this list of versions for each sub-image has been computed, these versions are ordered in a progressive refinement order, from the lowest quality to the highest quality and associated to the progressive step. For each sub-image, its versions are associated in order to the progressive steps that list this sub-image.
In this variant, at step 640, it is checked whether all the sub-image versions contained in a progressive step's refinements have already been loaded. If a sub-image version corresponds to an encoding layer of a multi-layer image item, then it is checked whether the data for this encoding layer and for all the previous encoding layers have been loaded. Otherwise, it is checked whether the data for the image item have been loaded.
In a variant instead of storing the nurnberofsub-anage versions contained in a progressive step's refinements, the number of sub-image versions contained in the progressive step's contentisstored. Inthisvahant at step 640,thetotalnumberofsub-image versions loaded is counted and compared to the number associated with the progressive step.
In a variant, a single item_count value is used for all the progressive steps. In this variant, the progressive item count item property may be described bythefollowing 10 stricture: aligned(8) class ProgressiveItemCountProperty extends ItemFullProperty('pcnt', version = 0, flags = 0) { unsigned int(16) item count; In another variant, the numberofsub-anagesforvefich a version is contained in a progressive step's refinements is specified as a percentage ofthetc4alnumberofsubimageversions. Possibly, this total number doesn't take into account the low-quality sub-image versions.
Embodiment combination Possibly, two or more embodiments of the invention can be combined. In a combination of embodiments, the structures describing the progressive information related to the different embodiments may be combined. These structures may be kept separate.
In particular, it may be advantageous to combine the refinement pattern embodiment with one of the other embodiments. In this combination the progressive refinement pattern indicates the progressive refinement strategy used in the HEIF file, while the offset indication or the item indication specifies the refinements of each progressive step.
For example, an item property for specifying both a progressive refinement pattern and alocafionforeach progressive step may havethefollowing structure: aligned(8) class ProgressivePatternLocationProperty extends ItemFullProperty('pplor, version = 0, flags = 0) { unsigned int(8) pattern_type; unsigned int(1) reverse_flag; unsigned int(1) bidirectional_flag; unsigned int(1) parameter_flag; unsigned int(5) reserved; unsigned int(8) scale; if (parameter_flag == 1) { unsigned int(16) parameter; unsigned int(16) step_count; for (i=0; i < step_count; i++) { if (flags & 1) { unsigned int(32) item_index; 5} else { unsigned int(16) item_index; unsigned int(16) extent_index; Possibly some part of an embodiment of the invention can be combined with another embodiment of the invention.
Implementation variants In the variants that indicate sub-images through their identifiers, instead of specifying the identifier of each sub-image, the index of the sub-image in the Item Reference Entry linking the main image to the sub-images may be used.
Possibly instead of specifying several progressive steps in a single item property, each progressive step may be specified in its own item property. These item properties may be ordered according to the progressive step order.
Possibly a different item property for describing the progressive information is used for each type of main image. For example, there may be a specific item property for describing the progressive information linked to a grid item and a different item property for describing the progressive information linked to an overlay item, each with its own list of pattern_type values.
Possibly the progressive information may be contained in a new item, a progressive item, for example with the prog' type. This progressive item may have the same structure as one of the item properties described in this invention. This progressive item may be associated with the main image item it applies to with an item reference, for example with the prog' 4cc. This item reference may be from the progressive item to the image item or from the image item to the progressive item.
Possibly a progressive item may specify a single progressive step. In this case, an item reference may link the main image item to the progressive items describing progressive steps for this main image. The progressive items may be ordered in the item reference entry according to the order of their respective progressive steps.
Possibly the progressive information may be contained in a new box, a progressive box, for example with the prog' 4cc. This new box may be located inside the 'meta' box of the HEIF file. This progressive box may have the same structure as one of the item properties described in this invention. In addition, this progressive box may contain an indication of the main image it applies to, for example by specifying the identifier of the main image item. There may be as many instances of this progressive box as there are image items supporting a progressive rendering in the HEIF file. Possibly a progressive box may specify a single progressive step. In this case the progressive box may include the progressive step number.
Possibly, a progressive step description may include a progressive step number. Possibly different sizes may be used for some of the various fields described here. Possibly variable sizes may be used for some of the various fields described here. Possibly the size of a field may be based on the 'flags' parameter of the box containing the field, on the 'version' of this box, or on a flag field included in this box. Possibly the size of several fields may depend on the same indication.
Possibly, the progressive information may indicate the progressive step's content instead of the progressive step's refinements.
Possibly the progressive steps for a main image are not organized as a list but as a directed acyclic graph. This means that a progressive step may have two or more following progressive steps and that a progressive step may have two or more preceding progressive steps. This also means that a progressive step should not be followed or preceded by itself, directly or indirectly. Such an organization is advantageous when the HEIF file is not intended to be loaded sequentially or when the content of the main image span several files that may be loaded in an undetermined order.
Progressive step variants Preferably, the specification order for the progressive steps corresponds to the order of these progressive steps.
Possibly, some progressive steps may be empty or missing.
Possibly, the progressive information may define progressive steps for a progressive loading of the low-quality versions of the sub-images. In the example of Figure 4f, using a center to border progressive refinement pattern, the first progressive step may include only the low-quality version of the center sub-image: '15', then the second progressive step refinements may include the low-quality version of the four sub-images '12', '14', '16', and '18'.
Possibly, after the last progressive step, some high-quality version of some sub-images may be missing.
Possibly, one or more versions of a sub-image may be missing.
File with Multiple Content Possibly, an HEIF file may include several image items with associated progressive information. The data for these image items may be interleaved or ordered sequentially.
For example, an HEIF file may contain a first image item with a progressive refinement pattern as described by Figure 4f and a second image item with a progressive refinement pattern as described by Figure 4g. The different sub-image versions for these image items may be organized as follows. First, all the low-quality versions of the sub-image of the first image item are stored, followed by the low-quality versions of the sub-image of the second image item. Then the high-quality version of the first row of the first item are stored, followed by the high-quality version of the first and second column of the second item, and so on.
For the same example, the different sub-image versions of the two image items may be organized differently: first all the sub-image versions of the first image item are stored, followed by all the sub-image versions of the second image item.
Possibly the interleaving or sequential organization of an HEIF file regarding the image items supporting a progressive rendering may be indicated. This indication may be realized by a brand in the HEIF file. There may be a brand, for example 'spro', for indicating a sequential organization. There may be a brand, for example 'ipro', for indicating an interleaved organization.
Possibly an item property may indicate for each image item supporting a progressive rendering whether it is stored sequentially or in an interleaved way.
Possibly other image items not supporting a progressive rendering or other non-image items may be included in an HEIF file containing one or more image items supporting a progressive rendering. Possibly these other image items or non-image items are stored after all the progressive image items.
Possibly the progressive information associated with an image item indicates where the metadata associated with this item is stored in relation to the progressive content of this image item. For example, the progressive information may indicate that the metadata is stored after all the content of the image item. As another example, the progressive information may indicate that the metadata is stored before the high-level versions of the sub-images of the image item.
Possibly, a component of an image item may itself be composed of several sub-images. For example, an overlay item may use a grid item as one of its components. In such a case, the progressive information associated with the composed sub-image defines the different versions of this sub-image.
Figure 7 illustrates the main steps for a progressive rendering of an HEIF file generated according to embodiments of the invention. It is a more generic alternative to the main steps illustrated by Figure 6. Note that these steps are described in reference to image items, but could also be used for a progressive rendering of a track, in particular for a track for which the temporal information is not necessarily meaningful, like 'pia track or a sequence of independent still images encapsulated in a track.
In a first step 700, an image item to render is selected.
Then, in steps 710 to 735, the potential entities for progressive refinement are determined. First, the image item itself is regarded as a potential entity for progressive refinement.
At step 710, it is checked whether the selected image item is included in an entity group of type 'alb'. If this is the case, then at step 715, the image items contained in that entity group are regarded as potential entities for progressive refinement. The next step is step 730.
Otherwise, at step 720, it is checked whether the selected image item has an associated thumbnail image item. If this is the case, at step 725, the associated thumbnail is regarded as a potential entity for progressive refinement. The next step is step 730.
At step 730, it is checked whether the selected image item is encoded as a multi-layer image with several layers. If this is the case, at step 735, the different encoded layers are regarded as potential entities for progressive refinement. If a progressive layer item property is associated with the selected image item, only the layers indicated in this item property are regarded as potential entities for progressive refinement.
At step 740, the next potential entity is obtained. For example, the loading of data is monitored for checking when data corresponding to a new item or a new layer has been loaded. If the loaded data enables to obtain a new image item regarded as a potential entity, this image item is the next potential entity. If the loaded data enables to obtain a new layer regarded as a potential entity, then this layer is the next potential entity. If the loaded data enables to obtain a new image item that is a sub-image of a potential entity, then this potential entity is the next potential entity. Otherwise, the monitoring continues.
Then at step 750, it is checked whether any entity has been rendered. If no entity has been rendered, then at step 755, the next potential entity is rendered. The next step is step 790.
Otherwise, at step 760, it is checked if an 'alb' group was found at step 710, and if true, whether the next potential entity is earlier in this 'alb' group than the currently rendered entity. If true, then at step 765, the next potential entity is rendered. The next step is step 790.
Otherwise, at step 770, it is checked if the currently rendered entity is a thumbnail and the next potential entity its master image. If true, then at step 775, then next potential entity is rendered. The next step is step 790.
Otherwise, at step 780, it is checked if the next potential entity is the same as the currently rendered entity and has progressive steps associated with it. If this is true, it is further checked whether a new progressive step can be rendered. If this is true, then at step 785, this new progressive step is displayed. The next step is step 790.
At step 790, it is checked whether the end of the HEIF file has been reached. If this is the case, no more steps are executed. Otherwise, the next step is step 740.
The 'altr' grouping type may be used to associate with an image item other image items that may be used in a progressive rendering of this image item. Some variants of the embodiments of this invention rely on 'altr' entity groups for indicating all the versions of a sub-image. Other grouping types may be used for indicating which image items may be used in a progressive rendering of a given image item. A specific grouping type, for example prog', may be used for associating image items in view of progressive rendering.
Possibly, a progressive refinement pattern item property may be associated with any image item for indicating the progressive refinement strategy selected for this image item Possibly, a progressive refinement item property may be associated with an image item for indicating that this image item support progressive rendering. The structure of this progressive refinement item property may be the following: aligned(8) class ProgressiveProperty extends ItemFullProperty('proe, version = 0, flags = 0) { Possibly, some progressive information may be associated with an entity group for indicating that the entire group supports progressive rendering. For example, a left-right progressive rendering pattern may be associated with a panorama group for indicating that the panorama itself can be rendered in a progressive way.
Figure 8 is a schematic block diagram of a computing device 800 for implementation of one or more embodiments of the invention. The computing device 800 may be a device such as a micro-computer, a workstation or a light portable device. The computing device 800 comprises a communication bus connected to: -a central processing unit 801, such as a microprocessor, denoted CPU; -a random access memory 802, denoted RAM, for storing the executable code of the method of embodiments of the invention as well as the registers adapted to record variables and parameters necessary for implementing the method according to embodiments of the invention, the memory capacity thereof can be expanded by an optional RAM connected to an expansion port, for example; -a read-only memory 803, denoted ROM, for storing computer programs for implementing embodiments of the invention; -a network interface 804 is typically connected to a communication network over which digital data to be processed are transmitted or received. The network interface 804 can be a single network interface, or composed of a set of different network interfaces (for instance wired and wireless interfaces, or different kinds of wired or wireless interfaces). Data packets are written to the network interface for transmission or are read from the network interface for reception under the control of the software application running in the CPU 801; - a graphical user interface 805 may be used for receiving inputs from a user or to display information to a user; -a hard disk 806 denoted HD may be provided as a mass storage device; - an I/O module 807 may be used for receiving/sending data from/to external devices such as a video source or display.
The executable code may be stored either in read only memory 803, on the hard disk 806 or on a removable digital medium such as for example a disk. According to a variant, the executable code of the programs can be received by means of a communication network, via the network interface 804, in order to be stored in one of the storage means of the communication device 800, such as the hard disk 806, before being executed.
The central processing unit 801 is adapted to control and direct the execution of the instructions or portions of software code of the program or programs according to embodiments of the invention, which instructions are stored in one of the aforementioned storage means. After powering on, the CPU 801 is capable of executing instructions from main RAM memory 802 relating to a software application after those instructions have been loaded from the program ROM 803 or the hard-disc (HD) 806 for example. Such a software application, when executed by the CPU 801, causes the steps of the flowcharts of the invention to be performed.
Any step of the algorithms of the invention may be implemented in software by execution of a set of instructions or program by a programmable computing machine, such as a PC ("Personal Computer"), a DSP ("Digital Signal Processor") or a microcontroller; or else implemented in hardware by a machine or a dedicated component, such as an FPGA ("Field-Programmable Gate Array") or an ASIC ("Application-Specific Integrated Circuit").
Although the present invention has been described herein above with reference to specific embodiments, the present invention is not limited to the specific embodiments, and modifications will be apparent to a skilled person in the art which lie within the scope of the present invention.
Many further modifications and variations will suggest themselves to those versed in the art upon making reference to the foregoing illustrative embodiments, which are given by way of example only and which are not intended to limit the scope of the invention, that being determined solely by the appended claims. In particular the different features from different embodiments may be interchanged, where appropriate.
Each of the embodiments of the invention described above can be implemented solely or as a combination of a plurality of the embodiments. Also, features from different embodiments can be combined where necessary or where the combination of elements or features from individual embodiments in a single embodiment is beneficial.
In the claims, the word "comprising" does not exclude other elements or steps, and the indefinite article "a" or "an" does not exclude a plurality. The mere fact that different features are recited in mutually different dependent claims does not indicate that a combination of these features cannot be advantageously used.

Claims (23)

  1. CLAIMS1 A method of encapsulating image data in a media file, the image data being related to a main image to be generated based on a plurality of input images, the method comprising: obtaining the plurality of input images, at least one input image being associated with different versions of the input image; generating descriptive metadata for describing the main image; encapsulating the plurality of input images and the descriptive metadata in the media file; wherein the method further comprises: generating a progressive information defining a set of consecutive progressive steps for generating a version of the main image, each progressive step being associated with a set of input image versions required to generate the version of the main image; and embedding the progressive information in the descriptive metadata.
  2. 2. The method of claim 1, wherein the progressive information comprises an indication of a progressive refinement pattern, the progressive refinement pattern being indicative of a scanning order of the plurality of input images.
  3. 3. The method of claim 2, wherein the indication of a progressive refinement pattern is an index in a list of predefined progressive refinement patterns.
  4. 4. The method of claim 3, wherein the progressive information further comprises a parameter associated with the progressive refinement pattern.
  5. 5. The method of claim 2, wherein the indication of a progressive refinement pattern is based on an equation and associated parameters.
  6. 6. The method of claim 1, wherein the progressive information comprises, for each progressive step, a position in the file of the input image data associated with the progressive step.
  7. 7. The method of claim 6, wherein the position indicates the last byte of image data in the file associated with the progressive step.
  8. 8. The method of claim 6, wherein the position comprises an offset and a length to indicate the input image data associated with the progressive step.
  9. 9. The method of claim 6, wherein the position comprises a list of extents of the input image data associated with the progressive step.
  10. 10 The method of claim 1, wherein input image data being organized into one or more extent, each extent being composed of contiguous image data related to a version of one input image, the progressive information comprises, in descriptive metadata describing an extent, an information indicating that the last byte of the extent corresponds to a progressive step.
  11. 11. The method of claim 1, wherein input images being described as image items, the progressive information comprises for each progressive step a list of the image item identifiers identifying the image items associated with the progressive step.
  12. 12. The method of claim 1, wherein the progressive information comprises for each progressive step a number of input image version comprised in the progressive step.
  13. 13. The method of claim 11, wherein at least one input image being composed of a plurality of layers, each layer being associated with a version of the input image, the progressive information further comprises a layer identifier associated with the image item identifier of the input image.
  14. 14 The method of Claim 1, wherein generating a progressive information comprises generating a progressive refinement data structure comprising data for determining the location of reconstruction points in the media file, each reconstruction point indicating that a reconstruction of the main image is possible using image data associated to input images previously received.
  15. 15. The method of Claim 14, wherein the progressive refinement data structure further comprises a number of reconstruction points.
  16. 16. The method of Claim 2, wherein the progressive refinement pattern defines the location of reconstruction points in the media file, each reconstruction point indicating that a reconstruction of the main image is possible using image data associated to input images previously received.
  17. 17. The method of Claim 1, wherein the progressive information characterizes a construction of the main image so that its quality is gradually improved.
  18. 18 A method of generating a main image to be generated based on a plurality of input images from an image data file, wherein the method comprises: obtaining, from the image data file, descriptive metadata describing the main image; obtaining, from the descriptive metadata, a progressive information defining a set of consecutive progressive steps for generating a version of the main image, each progressive step being associated with a set of input image versions required to generate the version of the main image; obtaining image data corresponding to the input images from the image data file; and generating at least two versions of the main image corresponding to respective progressive steps, each version of the main image being generated when the set of input image versions associated with the respective progressive step is obtained from the image data file.
  19. 19. A computer program product for a programmable apparatus, the computer program product comprising a sequence of instructions for implementing a method according to any one of claims 1 to 18, when loaded into and executed by the programmable apparatus.
  20. 20. A computer-readable storage medium storing instructions of a computer program for implementing a method according to any one of claims 1 to 18.
  21. 21. A computer program which upon execution causes the method of any one of claims 1 to 18 to be performed.
  22. 22 A device for encapsulating image data in a media file, the image data being related to a main image to be generated based on a plurality of input images, the device comprising a processor configured for: obtaining the plurality of input images, at least one input image being associated with different versions of the input image; generating descriptive metadata for describing the main image; encapsulating the plurality of input images and the descriptive metadata in the media file; wherein the method further comprises: generating a progressive information defining a set of consecutive progressive steps for generating a version of the main image, each progressive step being associated with a set of input image versions required to generate the version of the main image; and - embedding the progressive information in the descriptive metadata.
  23. 23 A device for generating a main image to be generated based on a plurality of input images from an image data file, wherein the device comprises a processor configured for: - obtaining, from the image data file, descriptive metadata describing the main image; obtaining, from the descriptive metadata, a progressive information defining a set of consecutive progressive steps for generating a version of the main image, each progressive step being associated with a set of input image versions required to generate the version of the main image; obtaining image data corresponding to the input images from the image data file; and -generating at least two versions of the main image corresponding to respective progressive steps, each version of the main image being generated when the set of input image versions associated with the respective progressive step is obtained from the image data file.
GB2020066.3A 2020-12-17 2020-12-17 Method and apparatus for encapsulating image data in a file for progressive rendering Pending GB2602101A (en)

Priority Applications (8)

Application Number Priority Date Filing Date Title
GB2020066.3A GB2602101A (en) 2020-12-17 2020-12-17 Method and apparatus for encapsulating image data in a file for progressive rendering
GB2105269.1A GB2602169A (en) 2020-12-17 2021-04-13 Method and apparatus for encapsulating image data in a file for progressive rendering
GB2105852.4A GB2602170B (en) 2020-12-17 2021-04-23 Method and apparatus for encapsulating image data in a file for progressive rendering
GB2304517.2A GB2617913B (en) 2020-12-17 2021-04-23 Method and apparatus for encapsulating image data in a file for progressive rendering
EP21839925.1A EP4264949A1 (en) 2020-12-17 2021-12-15 Method and apparatus for encapsulating image data in a file for progressive rendering
US18/257,236 US20240107129A1 (en) 2020-12-17 2021-12-15 Method and apparatus for encapsulating image data in a file for progressive rendering
JP2023508604A JP2023552029A (en) 2020-12-17 2021-12-15 Method and apparatus for encapsulating image data within a file for progressive rendering
PCT/EP2021/086002 WO2022129235A1 (en) 2020-12-17 2021-12-15 Method and apparatus for encapsulating image data in a file for progressive rendering

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
GB2020066.3A GB2602101A (en) 2020-12-17 2020-12-17 Method and apparatus for encapsulating image data in a file for progressive rendering

Publications (2)

Publication Number Publication Date
GB202020066D0 GB202020066D0 (en) 2021-02-03
GB2602101A true GB2602101A (en) 2022-06-22

Family

ID=74221342

Family Applications (3)

Application Number Title Priority Date Filing Date
GB2020066.3A Pending GB2602101A (en) 2020-12-17 2020-12-17 Method and apparatus for encapsulating image data in a file for progressive rendering
GB2105269.1A Pending GB2602169A (en) 2020-12-17 2021-04-13 Method and apparatus for encapsulating image data in a file for progressive rendering
GB2105852.4A Active GB2602170B (en) 2020-12-17 2021-04-23 Method and apparatus for encapsulating image data in a file for progressive rendering

Family Applications After (2)

Application Number Title Priority Date Filing Date
GB2105269.1A Pending GB2602169A (en) 2020-12-17 2021-04-13 Method and apparatus for encapsulating image data in a file for progressive rendering
GB2105852.4A Active GB2602170B (en) 2020-12-17 2021-04-23 Method and apparatus for encapsulating image data in a file for progressive rendering

Country Status (1)

Country Link
GB (3) GB2602101A (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016128419A1 (en) * 2015-02-10 2016-08-18 Canon Kabushiki Kaisha Image data encapsulation
GB2538997A (en) * 2015-06-03 2016-12-07 Nokia Technologies Oy A method, an apparatus, a computer program for video coding
GB2538998A (en) * 2015-06-03 2016-12-07 Nokia Technologies Oy A method, an apparatus, a computer program for video coding
GB2582025A (en) * 2019-03-08 2020-09-09 Canon Kk Method and apparatus for encapsulating groups of images in a file

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016128419A1 (en) * 2015-02-10 2016-08-18 Canon Kabushiki Kaisha Image data encapsulation
GB2538997A (en) * 2015-06-03 2016-12-07 Nokia Technologies Oy A method, an apparatus, a computer program for video coding
GB2538998A (en) * 2015-06-03 2016-12-07 Nokia Technologies Oy A method, an apparatus, a computer program for video coding
GB2582025A (en) * 2019-03-08 2020-09-09 Canon Kk Method and apparatus for encapsulating groups of images in a file

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
"ISO/IEC 14496-12", vol. w18855, October 2019, article "Information technology — Coding of audiovisual objects — Part 12: ISO base media file format"

Also Published As

Publication number Publication date
GB202105852D0 (en) 2021-06-09
GB2602169A (en) 2022-06-22
GB2602170B (en) 2023-05-17
GB202105269D0 (en) 2021-05-26
GB2602170A (en) 2022-06-22
GB202020066D0 (en) 2021-02-03

Similar Documents

Publication Publication Date Title
US11962809B2 (en) Image data encapsulation with referenced description information
US11805304B2 (en) Method, device, and computer program for generating timed media data
US10567784B2 (en) Description of image composition with HEVC still image file format
CN112153386B (en) Method and apparatus for reading image file and storage medium
CN105900401B (en) Method, apparatus and computer program for encoding inter-layer dependencies
CN109257624B (en) Method and apparatus for generating and processing media file and storage medium
CN113949938B (en) Packaging method and device, processing method and device, and storage medium
CN110800311B (en) Method, apparatus and computer program for transmitting media content
US10595062B2 (en) Image data encapsulation
CN114747219A (en) Method and apparatus for storing and signaling sub-sample entry descriptions
KR102465188B1 (en) Method and apparatus for encapsulating images in files
US20160029091A1 (en) Method of displaying a region of interest in a video stream
CN113822966A (en) Method, apparatus and computer-readable storage medium for generating media file
US20240107129A1 (en) Method and apparatus for encapsulating image data in a file for progressive rendering
GB2602101A (en) Method and apparatus for encapsulating image data in a file for progressive rendering
WO2022129235A1 (en) Method and apparatus for encapsulating image data in a file for progressive rendering
WO2022148650A1 (en) Method, device, and computer program for encapsulating timed media content data in a single track of encapsulated media content data
GB2602642A (en) Method and apparatus for encapsulating uncompressed video data into a file
CN116648917A (en) Method and apparatus for encapsulating image data in a file for progressive rendering
GB2620584A (en) Method, Device, and computer program for extrapolation regions in region tracks
GB2620582A (en) Method, device, and computer program for improving indexing of portions of encapsulated media data
GB2560649A (en) Image data encapsulation with tile support