WO2019128342A1 - Video processing method and apparatus - Google Patents
Video processing method and apparatus Download PDFInfo
- Publication number
- WO2019128342A1 WO2019128342A1 PCT/CN2018/107236 CN2018107236W WO2019128342A1 WO 2019128342 A1 WO2019128342 A1 WO 2019128342A1 CN 2018107236 W CN2018107236 W CN 2018107236W WO 2019128342 A1 WO2019128342 A1 WO 2019128342A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- video
- dimensional
- frame
- dimensional image
- dimensional video
- Prior art date
Links
- 238000003672 processing method Methods 0.000 title claims abstract description 9
- 238000000034 method Methods 0.000 claims abstract description 31
- 238000013507 mapping Methods 0.000 claims description 17
- 230000011218 segmentation Effects 0.000 claims description 12
- 238000012545 processing Methods 0.000 claims description 8
- 238000009877 rendering Methods 0.000 claims description 6
- 238000010586 diagram Methods 0.000 description 21
- 230000015572 biosynthetic process Effects 0.000 description 13
- 238000003786 synthesis reaction Methods 0.000 description 13
- 230000008569 process Effects 0.000 description 6
- 230000005540 biological transmission Effects 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 3
- 230000002093 peripheral effect Effects 0.000 description 3
- 238000007654 immersion Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012856 packing Methods 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 239000002131 composite material Substances 0.000 description 1
- 239000003814 drug Substances 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/21—Server components or server architectures
- H04N21/218—Source of audio or video content, e.g. local disk arrays
- H04N21/21805—Source of audio or video content, e.g. local disk arrays enabling multiple viewpoints, e.g. using a plurality of cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
- H04N21/23418—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
- H04N21/2343—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/236—Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
- H04N21/2365—Multiplexing of several video streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/44008—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/4402—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/816—Monomedia components thereof involving special video data, e.g 3D video
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/845—Structuring of content, e.g. decomposing content into time segments
- H04N21/8455—Structuring of content, e.g. decomposing content into time segments involving pointers to the content, e.g. pointers to the I-frames of the video stream
Definitions
- the present invention relates to the field of computers, and in particular to a video processing method and apparatus.
- VR Virtual Reality
- VR video leads the continuous breakthrough and application of VR technology.
- the VR video can be viewed by using the helmet and the mobile terminal, the device is simple, the operation is convenient, and the market is recognized.
- VR video is spliced into a 360-degree panoramic video, pushed to the terminal, and equipped with a display peripheral such as a helmet to provide an immersive viewing experience.
- a display peripheral such as a helmet
- 360-degree viewing without dead ends can be achieved.
- the VR video in the related art also has limitations.
- the VR video can not reflect the immersion, and the VR video cannot be played normally, but it also takes up the bandwidth and traffic of the VR video.
- the embodiment of the invention provides a video processing method and device, so as to at least solve the problem that the VR video cannot be played when the video playing terminal does not support the VR in the related art.
- a video processing method comprising: restoring a picture of a virtual reality VR video to a two-dimensional image; and generating a two-dimensional video of the plurality of views using the two-dimensional image.
- restoring the picture of the virtual reality VR video to the two-dimensional image comprises: parsing each frame of the VR video, and restoring each frame picture from the tile mode to the panorama mode by the spherical model rendering; Each frame of the panoramic mode performs view segmentation to obtain a two-dimensional image of a plurality of views.
- dividing each frame of the panoramic mode into a view angle to obtain a two-dimensional image of multiple views including one of: splitting a spherical surface of each frame in the panoramic mode into a hexahedron, Using the texture mapping method with the center of the circle as the center point, using a 120-degree angle of view, projecting in different directions, and obtaining images of six angles of up, down, left, and right; and customizing the spherical surface of each frame in the panoramic mode
- generating the two-dimensional video of the plurality of viewing angles by using the two-dimensional image comprises: merging the two-dimensional image into a multiplexed code stream according to the location information of the two-dimensional image; and performing the multiple multiplexed code streams Packing slices produces a two-dimensional video of multiple perspectives.
- the two-dimensional video of the multiple views is described by different index numbers.
- outputting the two-dimensional video includes: receiving perspective selection information; selecting a specified two-dimensional video corresponding to the perspective in all the two-dimensional videos according to the perspective selection information; and outputting the code stream of the specified two-dimensional video.
- the method before the outputting the two-dimensional video, the method further includes: determining that the video playing terminal meets at least one of the following conditions: VR play is not supported, and the VR wearable device is not connected.
- the method further includes: outputting the two-dimensional video.
- outputting the two-dimensional video comprises: outputting a two-dimensional video of one or more viewing angles.
- the two-dimensional video of the multiple viewing angles is output, the two-dimensional video of the multiple viewing angles is played on the screen in the video playing terminal.
- a video playback apparatus comprising: a restoration module configured to restore a picture of a virtual reality VR video to a two-dimensional image; and a generation module configured to generate the plurality of images using the two-dimensional image Two-dimensional video of a perspective.
- the restoring module includes: a parsing unit configured to parse each frame of the VR video, and restore each frame from a tile mode to a pan mode by a spherical model rendering; dividing the unit, setting In order to divide the viewing angle of each frame of the panoramic mode, a two-dimensional image of a plurality of viewing angles is obtained.
- the dividing unit is further configured to: divide a spherical surface of each frame in the panoramic mode into a hexahedron, and adopt a texture mapping manner with a center coordinate as a center point, and adopt a 120 degree angle of view.
- the generating module includes: a convergence unit configured to aggregate the two-dimensional image into multiple streams according to position information of the two-dimensional image; and a processing unit configured to perform the multiple streams Packing slices produces a two-dimensional video of multiple perspectives.
- the two-dimensional video of the multiple views is described by different index numbers.
- the device further includes: an output module configured to output the two-dimensional video.
- the output module includes: a receiving unit configured to receive view selection information; and a selecting unit configured to select a specified two-dimensional video corresponding to the view angle among all the two-dimensional videos according to the view selection information; an output unit, setting To output the code stream of the specified two-dimensional video.
- the device further includes: a determining module, configured to determine, before the output module outputs the two-dimensional video, that the video playing terminal meets at least one of the following conditions: the VR playing is not supported, and the VR wearing device is not connected.
- a determining module configured to determine, before the output module outputs the two-dimensional video, that the video playing terminal meets at least one of the following conditions: the VR playing is not supported, and the VR wearing device is not connected.
- a storage medium is also provided.
- the storage medium is set to store program code set to perform the following steps:
- a two-dimensional video of a plurality of viewing angles is generated using the two-dimensional image.
- the video of the two-dimensional (2D) view is regenerated by the reverse reduction of the video, which solves the problem that the VR video cannot be played when the video play terminal does not support the VR in the related art, and reduces the transmission bandwidth of the video play terminal. Requirements to reduce traffic consumption and storage capacity.
- FIG. 1 is a flow chart of a video processing method according to an embodiment of the present invention.
- FIG. 2 is a block diagram showing the structure of a video processing apparatus according to an embodiment of the present invention.
- FIG. 3 is a schematic view of a spherical projection in an embodiment of the present invention.
- FIG. 4 is a schematic diagram of a hexahedron division in an embodiment of the present invention.
- FIG. 5 is a schematic diagram of texture mapping in an embodiment of the present invention.
- FIG. 6 is a schematic diagram of a free view splitting in an embodiment of the present invention.
- FIG. 7 is a schematic diagram of code stream re-generation in an embodiment of the present invention.
- FIG. 8 is a flowchart of a client selection perspective in an embodiment of the present invention.
- FIG. 9 is a schematic diagram of a VR live broadcast in an embodiment of the present invention.
- FIG. 10 is a schematic diagram of VR on-demand in an embodiment of the present invention.
- FIG. 11 is a schematic diagram of comparison of two players in an embodiment of the present invention.
- FIG. 13 is a flowchart of automatically switching a viewing angle according to a connection of an external device according to an embodiment of the present invention
- FIG. 14 is a schematic diagram of clicking a video selection view video in an embodiment of the present invention.
- 15 is a schematic diagram of a video frame selection angle of view in an embodiment of the present invention.
- FIG. 1 is a flowchart of a video processing method according to an embodiment of the present invention. As shown in FIG. 1, the process includes the following steps:
- Step S102 restoring the picture of the virtual reality VR video to a two-dimensional image
- Step S104 generating a two-dimensional video of a plurality of viewing angles using the two-dimensional image.
- the video of the two-dimensional (2D) view is regenerated by the reverse reduction of the video, which solves the problem that the VR video cannot be played when the video play terminal does not support the VR in the related art, and reduces the transmission bandwidth of the video play terminal. Requirements to reduce traffic consumption and storage capacity.
- the execution body of the foregoing steps may be a video distribution platform, such as a content delivery network (CDN), or a video presentation platform, such as a playback terminal, but is not limited thereto.
- a video distribution platform such as a content delivery network (CDN)
- CDN content delivery network
- a video presentation platform such as a playback terminal
- restoring the picture of the virtual reality VR video to the two-dimensional image comprises:
- S12 Performing a view angle division on each frame of the panoramic mode to obtain a two-dimensional image of a plurality of viewing angles.
- the angle of view can be up, down, left, and right, front and rear, or angles of view according to different shooting positions.
- each frame of the panoramic mode is divided into a view angle to obtain a two-dimensional image of multiple viewing angles, which may be, but is not limited to, the following:
- the spherical surface of each frame in the panoramic mode is divided into hexahedrons, and the center of the center coordinates is used as a center point by using texture mapping, and the image is projected in different directions by using a 120-degree angle of view, and images of six angles of up, down, left, and right are obtained.
- the spherical surface of each frame in the panoramic mode is automatically divided.
- the texture mapping is used as the center point of the center of the circle, the Y axis is the positive direction, and the positive X direction is the starting point of 0 degrees, and is sequentially divided around the Z axis.
- M maps are generated and rotated around the X axis to generate M maps, and images of different viewing angles of 2M frames are obtained, and M is an integer greater than 1.
- Taking a custom angle of view as a 90-degree angle of view taking the origin as the center, the Y-axis is the positive direction, and the X-axis positive direction is the 0-degree starting point, which is sequentially divided around the Z-axis.
- the principle of segmentation also uses texture mapping. Generate 4 graphs. At the same time, four pictures are generated by rotating around the X axis, for a total of 8 frames.
- generating a two-dimensional video of the plurality of views using the two-dimensional image includes:
- S21 converge the two-dimensional image into a multi-stream according to the position information of the two-dimensional image; each frame is marked according to the position after the segmentation is completed, and the position information of the image frame of the same view has relevance;
- the two-dimensional video of multiple viewing angles is described by different index numbers, and the index numbers that can be generated in the order of cutting are sequentially.
- the method further includes: outputting the two-dimensional video.
- Output 2D video including:
- S31 Receive view selection information; the view selection information is used to indicate a viewing angle selected by the user or a preset policy selection (a presentation angle of the video playing terminal), and may also be a switch when the video is currently being played;
- the outputting the two-dimensional video comprises: outputting two-dimensional video of one or more viewing angles, and when simultaneously outputting the two-dimensional video of the multiple viewing angles, the video playing terminal may play and display by using a split screen or a multi-frequency manner.
- the method before the outputting the two-dimensional video, the method further includes: determining that the video playing terminal meets at least one of the following conditions: the VR playing is not supported, and the VR wearing device is not connected. This saves network bandwidth and storage space and prevents the terminal from downloading video types that cannot be played.
- the method according to the above embodiment can be implemented by means of software plus a necessary general hardware platform, and of course, by hardware, but in many cases, the former is A better implementation.
- the technical solution of the present invention which is essential or contributes to the prior art, may be embodied in the form of a software product stored in a storage medium (such as ROM/RAM, disk,
- the optical disc includes a number of instructions for causing a terminal device (which may be a cell phone, a computer, a server, or a network device, etc.) to perform the methods described in various embodiments of the present invention.
- a video processing device is also provided, which is used to implement the above-mentioned embodiments and preferred embodiments, and has not been described again.
- the term "module” may implement a combination of software and/or hardware of a predetermined function.
- the apparatus described in the following embodiments is preferably implemented in software, hardware, or a combination of software and hardware, is also possible and contemplated.
- FIG. 2 is a structural block diagram of a video processing apparatus according to an embodiment of the present invention. As shown in FIG. 2, the apparatus includes:
- the restoration module 20 is configured to restore the picture of the virtual reality VR video to a two-dimensional image
- the generating module 22 is configured to generate a two-dimensional video of a plurality of viewing angles using the two-dimensional image.
- the restoration module includes: a parsing unit configured to parse each frame of the VR video, and restore each frame from the tile mode to the panorama mode through the spherical model rendering; the segmentation unit is set to the pan mode
- a parsing unit configured to parse each frame of the VR video, and restore each frame from the tile mode to the panorama mode through the spherical model rendering; the segmentation unit is set to the pan mode
- Each frame of the picture is divided into a view angle to obtain a two-dimensional image of a plurality of angles of view.
- the dividing unit is further configured to: divide the spherical surface of each frame in the panoramic mode into a hexahedron, adopt a texture mapping manner with the center of the circle as a center point, adopt a 120 degree angle of view, and perform in different directions.
- Projection obtain images of six perspectives from top to bottom, left and right, and left and right; separate the spherical surface of each frame in the panorama mode, using texture mapping as the center point of the center of the circle, the positive direction of the Y axis, and the positive direction of the X axis It is a starting point of 0 degrees, and is sequentially divided around the Z axis to generate M maps, and rotates around the X axis to generate M maps, and obtain images of different viewing angles of 2M frames.
- the generating module includes: a convergence unit configured to aggregate the two-dimensional image into multiple streams according to the position information of the two-dimensional image; and the processing unit is configured to package the multiple streams to generate multiple perspectives. Dimensional video.
- the two-dimensional video of multiple viewing angles is described by different index numbers.
- the device further includes: an output module configured to output the two-dimensional video.
- the output module may include: a receiving unit configured to receive the viewing angle selection information; and a selecting unit configured to select the specified two-dimensional video of the corresponding viewing angle among all the two-dimensional videos according to the viewing angle selection information; and the output unit is configured to output the specified two-dimensional video Code stream.
- the device further includes: a determining module, configured to determine, before the output module outputs the two-dimensional video, that the video playing terminal meets at least one of the following conditions: the VR playing is not supported, and the VR wearing device is not connected.
- a determining module configured to determine, before the output module outputs the two-dimensional video, that the video playing terminal meets at least one of the following conditions: the VR playing is not supported, and the VR wearing device is not connected.
- each of the above modules may be implemented by software or hardware.
- the foregoing may be implemented by, but not limited to, the foregoing modules are all located in the same processor; or, the above modules are in any combination.
- the forms are located in different processors.
- the invention relates to a video processing scheme for VR video playing, and particularly relates to a method for selecting a viewing angle.
- the invention provides a method for multi-view selection of VR video based on reverse synthesis.
- the VR video is generally combined with a multi-channel video signal to generate a 360-degree panoramic video.
- the splicing seam is sharpened and the like to eliminate the traces to the greatest extent.
- the plane projection is performed to form a tiled image for transmission.
- the video stream is distributed to the video inverse synthesizer, the video is first restored to a spherical video by a reverse algorithm, and the video is divided into multiple 2D video streams according to a defined perspective and pushed to the CDN side.
- the terminal presents the video of the VR, the related view video is selected and pulled down to play, so that different viewing angle videos can be viewed without the related viewing device such as a helmet.
- This embodiment can be divided into three parts from the flow.
- the first part is the reverse synthesis of the video inverse synthesizer.
- the VR video is restored by the inverse algorithm and segmented into 2D images.
- the second part is to regenerate the code stream, slice the 2D image, package it into a new code stream, and describe it in the main m3u8 file.
- the third part is to select the perspective of the client and present the 2D video. The following instructions are presented:
- FIG. 3 is a schematic diagram of a spherical projection in the embodiment of the present invention, wherein the planar video forms a spherical texture VR video through spherical projection.
- FIG. 4 is a schematic diagram of the hexahedron segmentation in the embodiment of the present invention, and finally forms six figures of up, down, left, and right.
- FIG. 5 is a schematic diagram of texture mapping in the embodiment of the present invention.
- a 120-degree angle of view is used to shoot in different directions, that is, to project in different directions, and finally an image of each face is formed.
- Customized viewing angle segmentation you can customize the segmentation angle of view, taking the 90-degree angle of view as an example, taking the origin as the center, the Y-axis is the positive direction, and the X-axis positive direction is the 0-degree starting point, which is sequentially divided around the Z-axis.
- the principle of segmentation is the same.
- Four maps are generated by texture mapping.
- four pictures are generated by rotating around the X axis, for a total of 8 frames.
- FIG. 6 is a schematic diagram of free viewing angle division in the embodiment of the present invention.
- FIG. 7 is a schematic diagram of re-generation of a code stream in an embodiment of the present invention. After each frame is divided, the frame is marked according to the position, and the frame of the same position is aggregated and coded to form a single block.
- the description file of the code stream such as index1.m3u8, index2.m3u8....
- the index numbers generated in the order in which they are cut are described in the main index.m3u8:
- FIG. 8 is a flowchart of a client selection perspective in the embodiment of the present invention.
- the client first downloads the index.m3u8 file of the VR video for parsing, and obtains the angle between the angle and the sub-m3u8 file. Correspondence.
- the sub-m3u8 description file of the corresponding perspective is downloaded, and the stream is pulled down for playback.
- FIG. 9 is a schematic diagram of a VR live broadcast in an embodiment of the present invention.
- the head end collects images through a multi-channel camera, transmits the image to a synthesizer of the panoramic video, completes the stitching and composite projection of the panoramic video, and pushes the real-time message transmission protocol (
- the code of the Real Time Message Protocol (RTMP) is forwarded to the transcoder.
- the transcoder converts the rtmp protocol stream into a hls (HTTP Live Streaming) format stream and pushes it to the inverse synthesis server.
- RTMP Real Time Message Protocol
- the reverse synthesis of the VR panoramic video is completed in real time on the server of the video reverse synthesis, and the segmented 2D video is buffered into the cache and pushed to the CDN node.
- FIG. 10 is a schematic diagram of VR on-demand in the embodiment of the present invention.
- the VR panoramic video is injected into the encoder in the format of mp4, and the encoder is transcoded into the HLS format and pushed to the video inverse synthesis server.
- the reverse synthesis of the VR panoramic video is completed on the video inverse synthesis node, and the segmented 2D video is buffered into the cache and pushed to the CDN node. Compared with VR live reverse synthesis, no real-time considerations are needed.
- Example 3 the client perspective selection of VR video playback is not supported.
- FIG. 11 is a schematic diagram of comparison of two players in the embodiment of the present invention.
- FIG. 12 is a flow chart of client selection that does not support VR video in the embodiment of the present invention, including:
- the user selects one of the videos to play in full screen.
- the viewing angle is automatically switched according to the connection of the external device.
- FIG. 13 is a flowchart of automatically switching a viewing angle according to a connection of an external device according to an embodiment of the present invention, including:
- the user selects one of the videos to play in full screen.
- Example 5 When playing VR video, switch to the corresponding perspective 2D video playback.
- the VR video can be adjusted by gestures and gyroscopes.
- FIG. 14 is a schematic diagram of clicking a video selection view video in an embodiment of the present invention.
- FIG. 15 is a schematic diagram of the selection of the viewing angle video in the embodiment of the present invention. .
- the code stream of the 2D view of the code stream is regenerated by the reverse synthesis of the video, and the client downloads and plays the game through different modes, which effectively solves the problem that the terminal does not support VR video playback, and has no peripherals.
- the 2D video playback corresponding to the selection of the viewing angle reduces the bandwidth requirement and reduces the consumption of the traffic.
- Embodiments of the present invention also provide a storage medium.
- the above storage medium may be configured to store program code for performing the following steps:
- the foregoing storage medium may include, but not limited to, a USB flash drive, a Read-Only Memory (ROM), a Random Access Memory (RAM), a mobile hard disk, and a magnetic memory.
- ROM Read-Only Memory
- RAM Random Access Memory
- a mobile hard disk e.g., a hard disk
- magnetic memory e.g., a hard disk
- the processor executes according to the stored program code in the storage medium:
- a two-dimensional image of multiple perspectives is generated using a two-dimensional image.
- modules or steps of the present invention described above can be implemented by a general-purpose computing device that can be centralized on a single computing device or distributed across a network of multiple computing devices. Alternatively, they may be implemented by program code executable by the computing device such that they may be stored in the storage device by the computing device and, in some cases, may be different from the order herein.
- the steps shown or described are performed, or they are separately fabricated into individual integrated circuit modules, or a plurality of modules or steps thereof are fabricated as a single integrated circuit module.
- the invention is not limited to any specific combination of hardware and software.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Databases & Information Systems (AREA)
- Television Signal Processing For Recording (AREA)
- Image Generation (AREA)
- Controls And Circuits For Display Device (AREA)
Abstract
Description
Claims (18)
- 一种视频处理方法,包括:A video processing method comprising:将虚拟现实VR视频的画面还原成二维图像;Restore the picture of the virtual reality VR video to a two-dimensional image;使用所述二维图像生成多个视角的二维视频。A two-dimensional video of a plurality of viewing angles is generated using the two-dimensional image.
- 根据权利要求1所述的方法,其中,将虚拟现实VR视频的画面还原成二维图像包括:The method of claim 1 wherein restoring the picture of the virtual reality VR video to the two-dimensional image comprises:解析所述VR视频中的每一帧画面,并通过球面模型渲染将每一帧画面从平铺模式还原成全景模式;Parsing each frame of the VR video and restoring each frame from the tile mode to the panorama mode through spherical model rendering;将所述全景模式中的每一帧画面进行视角分割,得到多个视角的二维图像。Each frame of the panoramic mode is divided into a view angle to obtain a two-dimensional image of a plurality of views.
- 根据权利要求2所述的方法,其中,将所述全景模式中的每一帧画面进行视角分割,得到多个视角的二维图像,包括以下之一:The method according to claim 2, wherein each frame of the panoramic mode is subjected to view segmentation to obtain a two-dimensional image of a plurality of views, including one of the following:将所述全景模式中的每一帧画面的球面分割成六面体,采用纹理映射的方式以圆心坐标为中心点,采用120度视角,在不同的方向进行投影,得到上下左右前后六个视角的图像;The spherical surface of each frame in the panoramic mode is divided into hexahedrons, and the center of the center coordinates is used as a center point by using texture mapping, and the image is projected in different directions by using a 120-degree angle of view, and images of six angles of up, down, left, and right are obtained. ;将所述全景模式中的每一帧画面的球面进行自定义分割,采用纹理映射的方式以圆心坐标为中心点,Y轴为正方向,X轴正方向为0度起始点,绕Z轴依次分割,生成M个图,并绕X轴旋转生成M个图,得到2M帧不同视角的图像。Customizing the spherical surface of each frame in the panoramic mode, using texture mapping as the center point of the center of the circle, the Y axis is the positive direction, the positive X direction is the 0 degree starting point, and the Z axis is sequentially Splitting, generating M maps, and rotating around the X-axis to generate M maps, and obtaining images of different viewing angles of 2M frames.
- 根据权利要求1所述的方法,其中,使用所述二维图像生成多个视角的二维视频包括:The method of claim 1, wherein generating the two-dimensional video of the plurality of views using the two-dimensional image comprises:按照所述二维图像的位置信息将所述二维图像汇聚成多路码流;Converging the two-dimensional images into multiple streams according to position information of the two-dimensional image;对所述多路码流进行打包切片生成多个视角的二维视频。The multiplexed code stream is packetized and sliced to generate a two-dimensional video of multiple viewing angles.
- 根据权利要求1所述的方法,其中,所述多个视角的二维视频通过不同的索引号进行描述。The method of claim 1, wherein the two-dimensional video of the plurality of views is described by a different index number.
- 根据权利要求1所述的方法,其中,在使用所述二维图像生成多个视角的二维视频之后,所述方法还包括:The method of claim 1, wherein after generating the two-dimensional video of the plurality of views using the two-dimensional image, the method further comprises:输出所述二维视频。The two-dimensional video is output.
- 根据权利要求6所述的方法,其中,输出所述二维视频包括:The method of claim 6 wherein outputting the two-dimensional video comprises:接收视角选择信息;Receiving perspective selection information;根据所述视角选择信息在所有二维视频中选择对应视角的指定二维视频;Selecting a specified two-dimensional video corresponding to the view angle among all the two-dimensional videos according to the view selection information;输出所述指定二维视频的码流。Outputting the code stream of the specified two-dimensional video.
- 根据权利要求1所述的方法,其中,输出所述二维视频包括:The method of claim 1 wherein outputting the two-dimensional video comprises:输出一个或多个视角的二维视频。Output two-dimensional video of one or more views.
- 根据权利要求8所述的方法,其中,在输出多个视角的二维视频时,所述多个视角的二维视频在视频播放终端分屏播放。The method according to claim 8, wherein the two-dimensional video of the plurality of viewing angles is played on a split screen of the video playing terminal when the two-dimensional video of the plurality of viewing angles is output.
- 一种视频处理装置,包括:A video processing device comprising:还原模块,设置为将虚拟现实VR视频的画面还原成二维图像;a restoration module configured to restore a virtual reality VR video to a two-dimensional image;生成模块,设置为使用所述二维图像生成多个视角的二维视频。A generating module is configured to generate a two-dimensional video of a plurality of viewing angles using the two-dimensional image.
- 根据权利要求10所述的装置,其中,所述还原模块包括:The apparatus of claim 10 wherein said restoration module comprises:解析单元,设置为解析所述VR视频中的每一帧画面,并通过球面模型渲染将每一帧画面从平铺模式还原成全景模式;a parsing unit configured to parse each frame of the VR video and restore each frame from a tile mode to a pan mode by spherical model rendering;分割单元,设置为将所述全景模式中的每一帧画面进行视角分割,得到多个视角的二维图像。The dividing unit is configured to divide the viewing angle of each frame of the panoramic mode to obtain a two-dimensional image of a plurality of viewing angles.
- 根据权利要求11所述的装置,其中,所述分割单元还设置为以下之一:The apparatus according to claim 11, wherein said dividing unit is further set to one of:将所述全景模式中的每一帧画面的球面分割成六面体,采用纹理映射的方式以圆心坐标为中心点,采用120度视角,在不同的方向进行投影,得到上下左右前后六个视角的图像;The spherical surface of each frame in the panoramic mode is divided into hexahedrons, and the center of the center coordinates is used as a center point by using texture mapping, and the image is projected in different directions by using a 120-degree angle of view, and images of six angles of up, down, left, and right are obtained. ;将所述全景模式中的每一帧画面的球面进行自定义分割,采用纹理映射的方式以圆心坐标为中心点,Y轴为正方向,X轴正方向为0度起始点,绕Z轴依次分割,生成M个图,并绕X轴旋转生成M个图,得到2M帧不同视角的图像。Customizing the spherical surface of each frame in the panoramic mode, using texture mapping as the center point of the center of the circle, the Y axis is the positive direction, the positive X direction is the 0 degree starting point, and the Z axis is sequentially Splitting, generating M maps, and rotating around the X-axis to generate M maps, and obtaining images of different viewing angles of 2M frames.
- 根据权利要求10所述的装置,其中,所述生成模块包括:The apparatus of claim 10, wherein the generating module comprises:汇聚单元,设置为按照所述二维图像的位置信息将所述二维图像汇聚成多路码流;a convergence unit, configured to aggregate the two-dimensional images into multiple streams according to position information of the two-dimensional image;处理单元,设置为对所述多路码流进行打包切片生成多个视角的二维视频。The processing unit is configured to perform packet slicing on the multiple code streams to generate a two-dimensional video of multiple viewing angles.
- 根据权利要求10所述的装置,其中,所述多个视角的二维视频通过不同的索引号进行描述。The apparatus of claim 10, wherein the two-dimensional video of the plurality of views is described by a different index number.
- 根据权利要求10所述的装置,其中,所述装置还包括:The device of claim 10, wherein the device further comprises:输出模块,设置为输出所述二维视频。An output module configured to output the two-dimensional video.
- 根据权利要求15所述的装置,其中,所述输出模块包括:The apparatus of claim 15 wherein said output module comprises:接收单元,设置为接收视角选择信息;a receiving unit configured to receive viewing angle selection information;选择单元,设置为根据所述视角选择信息在所有二维视频中选择对应视角的指定二维视频;a selecting unit, configured to select a specified two-dimensional video corresponding to the viewing angle among all the two-dimensional videos according to the viewing angle selection information;输出单元,设置为输出所述指定二维视频的码流。An output unit configured to output a code stream of the specified two-dimensional video.
- 一种存储介质,所述存储介质包括存储的程序,其中,所述程序运行时执行权利要求1至9中任一项所述的方法。A storage medium, the storage medium comprising a stored program, wherein the program is executed to perform the method of any one of claims 1 to 9.
- 一种处理器,所述处理器设置为运行程序,其中,所述程序运行时执行权利要求1至9中任一项所述的方法。A processor, the processor being arranged to run a program, wherein the program is executed to perform the method of any one of claims 1 to 9.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711423373.XA CN108124193A (en) | 2017-12-25 | 2017-12-25 | Method for processing video frequency and device |
CN201711423373.X | 2017-12-25 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2019128342A1 true WO2019128342A1 (en) | 2019-07-04 |
Family
ID=62231481
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2018/107236 WO2019128342A1 (en) | 2017-12-25 | 2018-09-25 | Video processing method and apparatus |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN108124193A (en) |
WO (1) | WO2019128342A1 (en) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108124193A (en) * | 2017-12-25 | 2018-06-05 | 中兴通讯股份有限公司 | Method for processing video frequency and device |
CN109729338B (en) * | 2018-11-28 | 2021-10-01 | 北京虚拟动点科技有限公司 | Display data processing method, device and system |
CN112188303A (en) * | 2020-09-03 | 2021-01-05 | 北京火眼目测科技有限公司 | VR (virtual reality) streaming media playing method and device based on visual angle |
CN113489935A (en) * | 2020-09-11 | 2021-10-08 | 青岛海信电子产业控股股份有限公司 | Video interaction method and device |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030063816A1 (en) * | 1998-05-27 | 2003-04-03 | Industrial Technology Research Institute, A Taiwanese Corporation | Image-based method and system for building spherical panoramas |
CN106162207A (en) * | 2016-08-25 | 2016-11-23 | 北京字节跳动科技有限公司 | A kind of panoramic video parallel encoding method and apparatus |
CN107169924A (en) * | 2017-06-14 | 2017-09-15 | 歌尔科技有限公司 | The method for building up and system of three-dimensional panoramic image |
CN107197135A (en) * | 2016-03-21 | 2017-09-22 | 成都理想境界科技有限公司 | A kind of video generation method, player method and video-generating device, playing device |
CN107205140A (en) * | 2017-07-12 | 2017-09-26 | 赵政宇 | A kind of panoramic video segmentation projecting method and apply its system |
CN108124193A (en) * | 2017-12-25 | 2018-06-05 | 中兴通讯股份有限公司 | Method for processing video frequency and device |
-
2017
- 2017-12-25 CN CN201711423373.XA patent/CN108124193A/en active Pending
-
2018
- 2018-09-25 WO PCT/CN2018/107236 patent/WO2019128342A1/en active Application Filing
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030063816A1 (en) * | 1998-05-27 | 2003-04-03 | Industrial Technology Research Institute, A Taiwanese Corporation | Image-based method and system for building spherical panoramas |
CN107197135A (en) * | 2016-03-21 | 2017-09-22 | 成都理想境界科技有限公司 | A kind of video generation method, player method and video-generating device, playing device |
CN106162207A (en) * | 2016-08-25 | 2016-11-23 | 北京字节跳动科技有限公司 | A kind of panoramic video parallel encoding method and apparatus |
CN107169924A (en) * | 2017-06-14 | 2017-09-15 | 歌尔科技有限公司 | The method for building up and system of three-dimensional panoramic image |
CN107205140A (en) * | 2017-07-12 | 2017-09-26 | 赵政宇 | A kind of panoramic video segmentation projecting method and apply its system |
CN108124193A (en) * | 2017-12-25 | 2018-06-05 | 中兴通讯股份有限公司 | Method for processing video frequency and device |
Also Published As
Publication number | Publication date |
---|---|
CN108124193A (en) | 2018-06-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11870967B2 (en) | Methods and apparatus for supporting content generation, transmission and/or playback | |
US11303881B2 (en) | Method and client for playing back panoramic video | |
JP7223106B2 (en) | Method, device and computer program for adaptive streaming of virtual reality media content | |
WO2019128342A1 (en) | Video processing method and apparatus | |
JP6979035B2 (en) | How to Improve Streaming of Virtual Reality Media Content, Devices and Computer Programs | |
CN109076255B (en) | Method and equipment for sending and receiving 360-degree video | |
US10897646B2 (en) | Video stream transmission method and related device and system | |
KR20220081390A (en) | Methods and apparatus for delivering content and/or playing back content | |
JP2020503792A (en) | Information processing method and apparatus | |
RU2733218C2 (en) | Method, apparatus and a stream for formatting an immersive video image for traditional and immersive playback devices | |
TWI786572B (en) | Immersive media providing method and acquiring method, device, equipment and storage medium | |
Romero-Rondón et al. | Foveated streaming of virtual reality videos | |
JP2023529748A (en) | Support for multi-view video motion with disocclusion atlas | |
Podborski et al. | Virtual reality and DASH | |
CN107438203B (en) | Method for establishing and receiving list, network equipment and terminal | |
KR102243666B1 (en) | How to transmit 360 degree video, how to receive 360 degree video, 360 degree video transmission device, 360 degree video receiving device | |
JP7054351B2 (en) | System to play replay video of free viewpoint video | |
CN114513702B (en) | Web-based block panoramic video processing method, system and storage medium | |
US10593122B1 (en) | Mirroring edge pixels | |
KR20220054283A (en) | Methods for transmitting and rendering a 3D scene, a method for generating patches, and corresponding devices and computer programs | |
US20220329886A1 (en) | Methods and devices for handling media data streams | |
JP2022036123A (en) | System and method for reproducing replay video of free-viewpoint video | |
JP2023544049A (en) | Bidirectional presentation data stream | |
CN116235429A (en) | Bi-directional presentation of data streams using control plane channels and data plane channels |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 18896429 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18896429 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 17.11.2020) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18896429 Country of ref document: EP Kind code of ref document: A1 |