CN113766229B - Encoding method, decoding method, device, equipment and readable storage medium - Google Patents

Encoding method, decoding method, device, equipment and readable storage medium Download PDF

Info

Publication number
CN113766229B
CN113766229B CN202111160289.XA CN202111160289A CN113766229B CN 113766229 B CN113766229 B CN 113766229B CN 202111160289 A CN202111160289 A CN 202111160289A CN 113766229 B CN113766229 B CN 113766229B
Authority
CN
China
Prior art keywords
matrix
point
point cloud
target
sub
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111160289.XA
Other languages
Chinese (zh)
Other versions
CN113766229A (en
Inventor
冯亚楠
李琳
周冰
徐嵩
邢刚
马思伟
王苫社
徐逸群
胡玮
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Peking University
China Mobile Communications Group Co Ltd
MIGU Culture Technology Co Ltd
Original Assignee
Peking University
China Mobile Communications Group Co Ltd
MIGU Culture Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Peking University, China Mobile Communications Group Co Ltd, MIGU Culture Technology Co Ltd filed Critical Peking University
Priority to CN202111160289.XA priority Critical patent/CN113766229B/en
Publication of CN113766229A publication Critical patent/CN113766229A/en
Priority to PCT/CN2022/123245 priority patent/WO2023051783A1/en
Application granted granted Critical
Publication of CN113766229B publication Critical patent/CN113766229B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • G06F18/232Non-hierarchical techniques
    • G06F18/2321Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
    • G06F18/23213Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions with fixed number of clusters, e.g. K-means clustering
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/42Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/44Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • H04N19/517Processing of motion vectors by encoding
    • H04N19/52Processing of motion vectors by encoding by predictive encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Probability & Statistics with Applications (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Image Processing (AREA)

Abstract

The application discloses an encoding method, a decoding method, a device, equipment and a readable storage medium, which relate to the technical field of image processing and are used for improving processing performance. The method comprises the following steps: clustering point cloud data to be processed of the current frame to obtain a plurality of sub point clouds; generating a generalized Laplace matrix for any target sub-point cloud in the plurality of sub-point clouds according to Euclidean distances between a plurality of point pairs in the target sub-point cloud and Euclidean distances between a target point in the target sub-point cloud and a corresponding point of the target point; performing inter-frame prediction and graph Fourier residual transformation on the target sub-point cloud by using the generalized Laplacian matrix; respectively quantizing and encoding the transformed multiple sub-point clouds to obtain an encoded code stream; the corresponding point is located in a reference point cloud of the target sub-point cloud, and the reference point cloud is located in a reference frame of the current frame. The embodiment of the application can improve the processing performance.

Description

Encoding method, decoding method, device, equipment and readable storage medium
Technical Field
The present disclosure relates to the field of image processing technologies, and in particular, to an encoding method, a decoding method, an apparatus, a device, and a readable storage medium.
Background
With the development of computer hardware and algorithms, the acquisition of three-dimensional point cloud data is more convenient, and the data volume of the point cloud data is also larger. The point cloud data is composed of a large number of three-dimensional unordered points, each of which includes position information (X, Y, Z) and several attribute information (colors, normal vectors, etc.).
In order to facilitate storage and transmission of point cloud data, point cloud compression technology is becoming a focus of attention. The prior art provides a scheme for selectively encoding one or more 3D point clouds using inter-coding (e.g., motion compensation) techniques of previously encoded/decoded frames. However, such a scheme has poor processing performance such as encoding.
Disclosure of Invention
The embodiment of the application provides an encoding method, a decoding method, a device, equipment and a readable storage medium, so as to improve processing performance.
In a first aspect, an embodiment of the present application provides an encoding method, which is applied to an encoding device, including:
clustering point cloud data to be processed of the current frame to obtain a plurality of sub point clouds;
generating a generalized Laplace matrix for any target sub-point cloud in the plurality of sub-point clouds according to Euclidean distances between a plurality of point pairs in the target sub-point cloud and Euclidean distances between a target point in the target sub-point cloud and a corresponding point of the target point;
performing inter-frame prediction and graph Fourier residual transformation on the target sub-point cloud by using the generalized Laplacian matrix;
respectively quantizing and encoding the transformed multiple sub-point clouds to obtain an encoded code stream;
the corresponding point is located in a reference point cloud of the target sub-point cloud, and the reference point cloud is located in a reference frame of the current frame.
In a second aspect, embodiments of the present application further provide a decoding method, applied to a decoding device, where the method includes:
acquiring a coded code stream;
performing graph Fourier inverse transformation based on Euclidean distance weight on the coded code stream to obtain a transformation result;
based on the transformation result, obtaining a decoded code stream;
the coding code stream is obtained by coding the results of inter-frame prediction and graph Fourier residual transformation of the sub-point cloud by coding equipment.
In a third aspect, embodiments of the present application further provide an encoding apparatus, including:
the first acquisition module is used for clustering the point cloud data to be processed of the current frame to obtain a plurality of sub point clouds;
the first generation module is used for generating a generalized Laplace matrix for any target sub-point cloud in the multiple sub-point clouds according to Euclidean distances between multiple point pairs in the target sub-point cloud and Euclidean distances between a target point in the target sub-point cloud and a corresponding point of the target point;
the first transformation module is used for carrying out inter-frame prediction and graph Fourier residual transformation on the target sub-point cloud by utilizing the generalized Laplace matrix;
the first coding module is used for respectively quantizing and coding the transformed multiple sub-point clouds to obtain a coded code stream;
the corresponding point is located in a reference point cloud of the target sub-point cloud, and the reference point cloud is located in a reference frame of the current frame.
In a fourth aspect, embodiments of the present application further provide a decoding apparatus, including:
the first acquisition module is used for acquiring the coded code stream;
the first transformation module is used for carrying out graph Fourier inverse transformation based on Euclidean distance weight on the coded code stream to obtain a transformation result;
the first decoding module is used for obtaining a decoding code stream based on the transformation result;
the coding code stream is obtained by coding the results of inter-frame prediction and graph Fourier residual transformation of the sub-point cloud by coding equipment.
In a fifth aspect, embodiments of the present application further provide an electronic device, including: a memory, a processor and a program stored on the memory and executable on the processor, which processor implements the steps in the encoding method or decoding method as described above when executing the program.
In a sixth aspect, embodiments of the present application also provide a readable storage medium having a program stored thereon, which when executed by a processor, implements steps in an encoding method or a decoding method as described above.
In the embodiment of the application, the point cloud data to be processed of the current frame are clustered to obtain a plurality of sub-point clouds, and for any target sub-point cloud, a generalized Laplace matrix is generated according to Euclidean distances between a plurality of point pairs in the target sub-point clouds and Euclidean distances between target points in the target sub-point clouds and corresponding points of the target points, and the generalized Laplace matrix is utilized to respectively perform inter-frame prediction and graph Fourier residual transformation on the plurality of sub-point clouds, so that a coding code stream is obtained based on a transformation result. Because the generalized Laplace matrix is generated by utilizing Euclidean distance between points, global correlation characteristics can be utilized in the embodiment of the application, so that the correlation between the points is more fully expressed, the similarity between the point cloud data can be removed as much as possible, and the coding performance is improved.
The performance of the encoding end is improved, and correspondingly, for the decoding end, the data to be decoded is optimized, so that the decoding efficiency and performance can be correspondingly improved.
Drawings
FIG. 1 is a flow chart of an encoding method provided by an embodiment of the present application;
FIGS. 2 and 3 are schematic illustrations of the effects of the method of the embodiments of the present application and the prior art method;
FIG. 4 is a flow chart of a decoding method provided by an embodiment of the present application;
FIG. 5 is a block diagram of an encoding apparatus provided in an embodiment of the present application;
fig. 6 is a block diagram of a decoding apparatus according to an embodiment of the present application.
Detailed Description
In the embodiment of the application, the term "and/or" describes the association relationship of the association objects, which means that three relationships may exist, for example, a and/or B may be represented: a exists alone, A and B exist together, and B exists alone. The character "/" generally indicates that the context-dependent object is an "or" relationship.
The term "plurality" in the embodiments of the present application means two or more, and other adjectives are similar thereto.
The following description of the technical solutions in the embodiments of the present application will be made clearly and completely with reference to the accompanying drawings in the embodiments of the present application, and it is apparent that the described embodiments are only some embodiments of the present application, but not all embodiments. All other embodiments, which can be made by one of ordinary skill in the art without undue burden from the present disclosure, are within the scope of the present disclosure.
Referring to fig. 1, fig. 1 is a flowchart of an encoding method provided in an embodiment of the present application, which is applied to an encoding apparatus. As shown in fig. 1, the method comprises the following steps:
step 101, clustering point cloud data to be processed of a current frame to obtain a plurality of sub point clouds.
In the step, the point cloud data to be processed is subjected to voxel conversion to obtain point cloud voxels, and then the point cloud voxels are clustered to obtain a plurality of sub point clouds.
Specifically, a three-dimensional grid with a preset size is constructed, point cloud data to be processed are placed in the constructed three-dimensional grid to obtain coordinates of each point, and the three-dimensional grid containing the points is used as point cloud voxels to obtain a plurality of point cloud voxels. In addition, the coordinate and attribute information of each point cloud voxel can be obtained. Wherein the attribute information includes intensity, color, and the like. In the embodiment of the application, the coordinates of the point cloud voxels are specifically the coordinates of the center point of each point in the point cloud voxels; the color information of the point cloud voxel is specifically an average value of the color information of each point in the point cloud voxel. In practical application, the point cloud data to be processed can be subjected to voxel conversion in an octree mode and the like, so that a plurality of point cloud voxels are obtained.
And clustering the point cloud data by using a space uniform partitioning method. The clustering method may employ, for example, a K-means clustering method.
In the embodiment of the application, the point cloud data to be processed is divided into a plurality of sub-point clouds based on the position information, and the space is uniformly divided. Each sub-point cloud may be encoded independently.
Step 102, generating a generalized laplace matrix according to Euclidean distances between a plurality of point pairs in the target sub-point cloud and Euclidean distances between a target point in the target sub-point cloud and a corresponding point of the target point for any one of the plurality of sub-point clouds. The corresponding point is located in a reference point cloud of the target sub-point cloud, and the reference point cloud is located in a reference frame of the current frame.
Any one of the multiple sub-point clouds can be used as the target sub-point cloud. In practical application, the processing mode of each target sub-point cloud is the same.
Specifically, in this step, the following may be included:
s1021, obtaining a weight matrix according to Euclidean distances among a plurality of point pairs in the target sub-point cloud.
Wherein the target sub-point cloud can comprise a plurality of points, and each two points form the implementationOne point pair in the example. In the embodiment of the application, the Euclidean distance between two points in each pair of points is calculated. For example, for the ith point and the jth point in the target sub-point cloud, the euclidean distance between the ith point and the jth point is calculated. Specifically, for point i (x 1 ,x 2 ……x n ) And point j (y 1 ,y 2 ……y n ) The Euclidean distance d (i, j) between the two can be calculated according to the following formula in practical application:
Figure GDA0004014290900000051
wherein i is more than or equal to 1 and less than or equal to M, j is more than or equal to 1 and less than or equal to M, i, j and M are integers, and M is the total number of points included in the target sub-point cloud.
Thereafter, weights are calculated according to the following formula, and the weight matrix W is formed using the weights:
Figure GDA0004014290900000052
wherein ,Wij Representing the weight corresponding to the edge from the ith point to the jth point in the target sub-point cloud; distance represents the Euclidean distance between the ith point and the jth point; σ is a constant that is not equal to 0, representing the adjustment parameter.
And S1022, obtaining the Laplace matrix according to the degree matrix and the weight matrix.
In this step, the difference between the availability matrix and the weight matrix is used as a laplace matrix.
Specific: l=d-W, L represents a laplace matrix, D represents a degree matrix, and W represents a weight matrix.
Wherein the diagonal element d of the degree matrix i =∑ j W ij The other element is 0. Wherein d i The ith diagonal element of the degree matrix, W ij Representing the weight corresponding to the edge from the ith point to the jth point in the target sub-point cloud
S1023, generating a diagonal matrix.
The diagonal matrix is generated from Euclidean distances between target points in the target sub-point cloud and corresponding points of the target points.
Specifically, a reference point cloud of the target sub-point cloud may be first determined in the reference frame. For example, motion estimation is performed in a reference frame, finding a matching reference point cloud. The target sub-point clouds and the reference points are in one-to-one correspondence. For example, using an iterative closest point algorithm, in the reference frame, a reference point cloud for the target sub-point cloud may be determined based on Euclidean distance. Then, the diagonal matrix D is generated based on Euclidean distance between each point in the target sub-point cloud and the corresponding point of each point in the reference point cloud w . Wherein the value on the ith diagonal of the diagonal matrix is the reciprocal of the Euclidean distance between the ith point and the point p, and the other elements are 0. Wherein the point p is the corresponding point of the ith point in the reference point cloud.
S1024, obtaining the generalized Laplace matrix according to the diagonal matrix and the Laplace matrix.
In this step, the sum of the diagonal matrix and the laplacian matrix is used as the generalized laplacian matrix.
Specific: lg=l+d w Wherein Lg represents a generalized Laplace matrix, L represents a Laplace matrix, and D w Representing a diagonal matrix.
And 103, performing inter-frame prediction and graph Fourier residual transformation on the target sub-point cloud by using the generalized Laplacian matrix.
In this step, the inter-prediction and graph fourier residual transform may be understood as inter-prediction and graph fourier residual transform based on euclidean distance weights, and may include the following contents:
s1031, obtaining an attribute predicted value of the reference frame on the target attribute of the current frame.
In the embodiment of the application, an inter-frame prediction method is adopted, and a reference frame is utilized to predict the attribute value of the current frame. Wherein the attributes may include color, intensity, normal vector, etc. Then the target attribute may be any attribute.
Specifically, in this step, the attribute prediction value of the target attribute of the reference frame to the current frame is obtained according to the following formula:
Figure GDA0004014290900000061
wherein ,
Figure GDA0004014290900000062
attribute predictors, x, representing target attributes of a reference frame to a current frame t-1 Attribute values representing target attributes of the reference frame, lg represents a generalized Laplacian matrix.
S1032, generating a residual error of the target attribute of the current frame according to the attribute value of the target attribute of the current frame and the attribute predicted value of the target attribute of the reference frame to the current frame.
Specifically, here, a difference between an attribute value of the target attribute of the current frame and an attribute prediction value of the target attribute of the reference frame to the current frame may be used as the residual:
Figure GDA0004014290900000063
wherein delta represents the residual of the target property of the current frame,
Figure GDA0004014290900000064
attribute predictors, x, representing target attributes of a reference frame to a current frame t An attribute value representing a target attribute of the current frame.
The residual is obtained by the inter prediction method, so that the difference between two frames can be obtained as much as possible. Since the same part between two frames does not need additional processing, the code rate can be saved by calculating the residual error.
S1033, transforming the residual error of the target attribute of the current frame based on the generalized Laplace matrix.
In this step, a transformation matrix is obtained by using the generalized laplace matrix, and then, a residual error of the target attribute of the current frame is transformed by using the transformation matrix.
Specifically, the following formula is solved to obtain a transformation matrix:
Figure GDA0004014290900000071
wherein Lg represents a generalized Laplace matrix,
Figure GDA0004014290900000072
representing the transformation matrix.
The transformation result is obtained using the following formula:
Figure GDA0004014290900000073
wherein ,
Figure GDA0004014290900000075
representing the transformation result +.>
Figure GDA0004014290900000074
Representing a transformation matrix, delta representing the residual of the target attribute of the current frame.
In the embodiment of the application, the concept of the generalized graph Fourier transform is introduced on the basis of the traditional graph Fourier transform, and prediction and residual transformation are performed on the inter-frame attribute of the point cloud data, so that the redundancy among the data can be further removed, and the coding efficiency is improved.
The processing mode of other sub-point clouds is the same as the processing mode of the target sub-point clouds.
And 104, respectively quantizing and encoding the transformed multiple sub-point clouds to obtain an encoded code stream.
In the step, the transformed multiple sub-point clouds are uniformly quantized and arithmetically encoded to obtain an encoded code stream.
Taking the target attribute as a color as an example, here, the color can be decomposed into three 3×1 vectors (YUV or RGB). Taking the Y component as an example, the attribute value of the current frame is predicted according to the procedure in S1031, and a residual is generated according to S1032. After that, the residual is transformed by S1033. And uniformly quantizing and arithmetic coding the transformed Y component to obtain a code stream. For each component, the processing may be performed in the same manner as for the Y component.
In the embodiment of the application, since the generalized Laplace matrix is generated by using Euclidean distance between points, global correlation characteristics can be utilized in the embodiment of the application, and the correlation between the points is more fully expressed, so that the similarity between the point cloud data can be removed as much as possible, and the coding performance is improved.
In practical applications, tests are performed on a sequence of actual point clouds. In the test, first, a test is performed on a 16-frame dynamic point cloud, and a comparison of the performance of the method of the embodiment of the present application and the performance of the RAHT (Region-Adaptive Hierarchical Transform, region adaptive level transform) NWGFT (main direction weight map fourier transform) method is shown in fig. 2. In order to quantify the gain, a comparison of the data with the RAHT method was made in the experiment, as shown in fig. 3.
As can be seen from fig. 2 and fig. 3, in the embodiment of the present application, based on the conventional graph fourier transform, the concept of the generalized graph fourier transform is introduced, and prediction and residual transformation are performed on the inter-frame attribute of the point cloud, so that redundancy between data can be further removed, and coding efficiency is improved. Experimental results show that the method can improve subjective and objective performance and can be applied to compression, transmission and storage systems of actual point clouds.
Referring to fig. 4, fig. 4 is a flowchart of a decoding method provided in an embodiment of the present application, which is applied to an encoding apparatus. As shown in fig. 4, the method comprises the following steps:
step 401, obtaining a coded code stream.
The coding code stream is obtained by coding a result of inter-frame prediction and graph Fourier residual transformation of the sub-point cloud by using a generalized Laplace matrix by coding equipment.
And step 402, performing graph Fourier transform based on Euclidean distance weight on the coded code stream to obtain a transform result.
At the decoding end, after entropy decoding is carried out on the coded code stream, inverse quantization is carried out on the coded code stream. And then, performing graph Fourier inverse transformation based on Euclidean distance weight on the code stream after inverse quantization to obtain a transformation result.
Specifically, the following formula may be used to perform the graph fourier transform based on the euclidean distance weight on the dequantized encoded code stream:
Figure GDA0004014290900000081
wherein ,
Figure GDA0004014290900000082
representing the inverse transform residual value,/->
Figure GDA0004014290900000083
Representing a transformation matrix->
Figure GDA0004014290900000084
The quantized residual value representing the target property of the current frame, epsilon represents the dequantization coefficient.
And step 403, obtaining a decoding code stream based on the transformation result.
In the embodiment of the application, since the generalized Laplace matrix is generated by using Euclidean distance between points, global correlation characteristics can be utilized in the embodiment of the application, and the correlation between the points is more fully expressed, so that the similarity between the point cloud data can be removed as much as possible, and the coding performance is improved. The performance of the encoding end is improved, and correspondingly, for the decoding end, the data to be decoded is optimized, so that the decoding efficiency and performance can be correspondingly improved.
The embodiment of the application also provides a coding device. Referring to fig. 5, fig. 5 is a block diagram of an encoding apparatus provided in an embodiment of the present application. Since the principle of the coding device for solving the problem is similar to that of the coding method in the embodiment of the present application, the implementation of the coding device can refer to the implementation of the method, and the repetition is omitted.
As shown in fig. 5, the encoding apparatus 500 includes:
the first obtaining module 501 is configured to cluster point cloud data to be processed of a current frame to obtain multiple sub-point clouds; a first generating module 502, configured to generate, for any one of the multiple sub-point clouds, a generalized laplace matrix according to euclidean distances between multiple point pairs in the target sub-point cloud and euclidean distances between a target point in the target sub-point cloud and a corresponding point of the target point; a first transform module 503, configured to perform inter-frame prediction and graph fourier residual transform on a target sub-point cloud by using the generalized laplace matrix; a first encoding module 504, configured to quantize and encode the transformed multiple sub-point clouds respectively, to obtain an encoded code stream; the corresponding point is located in a reference point cloud of the target sub-point cloud, and the reference point cloud is located in a reference frame of the current frame.
Optionally, the first obtaining module includes: the first processing sub-module is used for carrying out voxel formation on the point cloud data to be processed to obtain point cloud voxels; and the first acquisition sub-module is used for clustering the point cloud voxels to obtain a plurality of sub-point clouds.
Optionally, the first generating module includes:
the first acquisition sub-module is used for obtaining a weight matrix according to Euclidean distances among a plurality of point pairs in the target sub-point cloud; the second acquisition sub-module is used for obtaining a Laplace matrix according to the degree matrix and the weight matrix; the first generation submodule is used for generating a diagonal matrix; and the second generation submodule is used for obtaining the generalized Laplace matrix according to the diagonal matrix and the Laplace matrix.
Specifically, the second obtaining submodule is configured to use a difference between the degree matrix and the weight matrix as a laplace matrix; the second generating submodule is used for utilizing the sum of the diagonal matrix and the Laplace matrix as the generalized Laplace matrix.
Wherein the diagonal element d of the degree matrix i =∑ j W ij, wherein ,di The ith diagonal element of the degree matrix, W ij Representing the weight corresponding to the edge from the ith point to the jth point in the target sub-point cloud; 1.ltoreq.i.ltoreq.M, 1.ltoreq.j.ltoreq.M, i, j, M being an integer, M being the total number of points included in the target sub-point cloud;
the diagonal matrix is generated from Euclidean distances between target points in the target sub-point cloud and corresponding points of the target points.
Optionally, the first obtaining submodule includes:
the first calculation unit is used for calculating the Euclidean distance between the ith point and the jth point in the target sub-point cloud;
a first acquisition unit for calculating weights according to the following formula, and forming the weight matrix by using the weights:
Figure GDA0004014290900000101
wherein ,Wij Representing the weight corresponding to the edge from the ith point to the jth point in the target sub-point cloud; distance represents the Euclidean distance between the ith point and the jth point; sigma is a constant not equal to 0, representing the adjustment parameter; and i is more than or equal to 1 and less than or equal to M, j is more than or equal to 1 and less than or equal to M, i, j, M is an integer, and M is the total number of points included in the target sub-point cloud.
Optionally, the first generating sub-module includes:
a first determining unit, configured to determine a reference point cloud of the target sub-point cloud in the reference frame; a first generation unit, configured to generate the diagonal matrix based on a euclidean distance between each point in a target sub-point cloud and a corresponding point of each point in the reference point cloud; the value on the ith diagonal of the diagonal matrix is the inverse of the Euclidean distance between the ith point and a point p, wherein the point p is the corresponding point of the ith point in the reference point cloud.
Optionally, the first determining unit is configured to determine, in the reference frame, a reference point cloud of the target sub-point cloud by using an iterative closest point algorithm.
Optionally, the first transformation module includes:
a first obtaining sub-module, configured to obtain an attribute prediction value of the target attribute of the current frame by using the reference frame; a first generation sub-module, configured to generate a residual error of the target attribute of the current frame according to the attribute value of the target attribute of the current frame and the attribute prediction value of the target attribute of the reference frame to the target attribute of the current frame; and the first transformation submodule is used for transforming the residual error of the target attribute of the current frame based on the generalized Laplace matrix.
Optionally, the first obtaining submodule is configured to obtain an attribute prediction value of the target attribute of the reference frame to the current frame according to the following formula:
Figure GDA0004014290900000102
wherein ,
Figure GDA0004014290900000103
attribute predictors, x, representing target attributes of a reference frame to a current frame t-1 Attribute values representing target attributes of the reference frame, lg represents a generalized Laplacian matrix.
Optionally, the first generating sub-module is configured to use, as the residual, a difference between an attribute value of the target attribute of the current frame and an attribute prediction value of the target attribute of the reference frame to the current frame.
Optionally, the first transformation submodule includes:
the first acquisition unit is used for obtaining a transformation matrix by using the generalized Laplace matrix; and the first transformation unit is used for transforming the residual error of the target attribute of the current frame by utilizing the transformation matrix.
Optionally, the first obtaining unit is configured to solve the following formula to obtain a transformation matrix:
Figure GDA0004014290900000111
wherein Lg represents a generalized Laplace matrix,
Figure GDA0004014290900000112
representing the transformation matrix.
Optionally, the first transforming unit is configured to obtain the transforming result by using the following formula:
Figure GDA0004014290900000113
wherein ,
Figure GDA0004014290900000119
representing the transformation result +.>
Figure GDA0004014290900000114
Representing a transformation matrix, delta representing the residual of the target attribute of the current frame.
The device provided in the embodiment of the present application may execute the above method embodiment, and its implementation principle and technical effects are similar, and this embodiment will not be described herein again.
The embodiment of the application also provides a decoding device. Referring to fig. 6, fig. 6 is a block diagram of a decoding apparatus provided in an embodiment of the present application. Since the principle of the decoding device for solving the problem is similar to that of the decoding method in the embodiment of the present application, the implementation of the decoding device can refer to the implementation of the method, and the repetition is omitted.
As shown in fig. 6, the decoding apparatus 600 includes:
a first obtaining module 601, configured to obtain a coded code stream; the first transformation module 602 is configured to perform inverse fourier transform on the encoded code stream based on the euclidean distance weight, to obtain a transformation result; a first decoding module 603, configured to obtain a decoded code stream based on the transformation result; the coding code stream is obtained by coding the results of inter-frame prediction and graph Fourier residual transformation of the sub-point cloud by coding equipment.
Optionally, the first transformation module includes: the first processing submodule is used for carrying out inverse quantization on the coded code stream; and the first transformation submodule is used for carrying out graph Fourier inverse transformation based on Euclidean distance weight on the code stream after inverse quantization to obtain a transformation result.
Optionally, the first transform submodule is configured to perform graph fourier transform based on the euclidean distance weight on the dequantized coded code stream by using the following formula:
Figure GDA0004014290900000115
wherein ,
Figure GDA0004014290900000116
representing the inverse transform residual value,/->
Figure GDA0004014290900000117
Representing a transformation matrix->
Figure GDA0004014290900000118
The quantized residual value representing the target property of the current frame, epsilon represents the dequantization coefficient.
The device provided in the embodiment of the present application may execute the above method embodiment, and its implementation principle and technical effects are similar, and this embodiment will not be described herein again.
It should be noted that, in the embodiment of the present application, the division of the units is schematic, which is merely a logic function division, and other division manners may be implemented in actual practice. In addition, each functional unit in each embodiment of the present application may be integrated in one processing unit, or each unit may exist alone physically, or two or more units may be integrated in one unit. The integrated units may be implemented in hardware or in software functional units.
The integrated units, if implemented in the form of software functional units and sold or used as stand-alone products, may be stored in a processor-readable storage medium. Based on such understanding, the technical solution of the present application may be embodied in essence or a part contributing to the prior art or all or part of the technical solution, in the form of a software product stored in a storage medium, including several instructions to cause a computer device (which may be a personal computer, a server, or a network device, etc.) or a processor (processor) to perform all or part of the steps of the methods described in the embodiments of the present application. And the aforementioned storage medium includes: a U-disk, a removable hard disk, a Read-Only Memory (ROM), a random access Memory (Random Access Memory, RAM), a magnetic disk, or an optical disk, or other various media capable of storing program codes.
The embodiment of the application also provides electronic equipment, which comprises: a memory, a processor and a program stored on the memory and executable on the processor, which processor implements the steps in the encoding method or decoding method as described above when executing the program.
The embodiment of the present application further provides a readable storage medium, on which a program is stored, where the program, when executed by a processor, implements each process of the foregoing encoding or decoding method embodiment, and the same technical effects can be achieved, so that repetition is avoided, and no further description is given here. The readable storage medium may be any available medium or data storage device that can be accessed by a processor, including, but not limited to, magnetic memories (e.g., floppy disks, hard disks, magnetic tapes, magneto-optical disks (MO), etc.), optical memories (e.g., CD, DVD, BD, HVD, etc.), semiconductor memories (e.g., ROM, EPROM, EEPROM, nonvolatile memories (NAND FLASH), solid State Disks (SSD)), etc.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
From the above description of the embodiments, it will be clear to those skilled in the art that the above-described embodiment method may be implemented by means of software plus a necessary general hardware platform, but of course may also be implemented by means of hardware, but in many cases the former is a preferred embodiment. In light of such understanding, the technical solutions of the present application may be embodied essentially or in part in the form of a software product stored in a storage medium (e.g., ROM/RAM, magnetic disk, optical disk) and including instructions for causing a terminal (which may be a cell phone, computer, server, air conditioner, or network device, etc.) to perform the methods described in the embodiments of the present application.
The embodiments of the present application have been described above with reference to the accompanying drawings, but the present application is not limited to the above-described embodiments, which are merely illustrative and not restrictive, and many forms may be made by those of ordinary skill in the art without departing from the spirit of the present application and the scope of the claims, which are also within the protection of the present application.

Claims (18)

1. An encoding method applied to an encoding apparatus, comprising:
clustering point cloud data to be processed of the current frame to obtain a plurality of sub point clouds;
generating a generalized Laplace matrix for any target sub-point cloud in the plurality of sub-point clouds according to Euclidean distances between a plurality of point pairs in the target sub-point cloud and Euclidean distances between a target point in the target sub-point cloud and a corresponding point of the target point;
performing inter-frame prediction and graph Fourier residual transformation on the target sub-point cloud by using the generalized Laplace matrix;
respectively quantizing and encoding the transformed multiple sub-point clouds to obtain an encoded code stream;
the corresponding point is located in a reference point cloud of the target sub-point cloud, and the reference point cloud is located in a reference frame of the current frame;
wherein the generating a generalized laplace matrix according to the euclidean distance between the plurality of point pairs in the target sub-point cloud and the euclidean distance between the target point in the target sub-point cloud and the corresponding point of the target point comprises:
obtaining a weight matrix according to Euclidean distances among a plurality of point pairs in the target sub-point cloud;
obtaining a Laplace matrix according to the degree matrix and the weight matrix;
generating a diagonal matrix;
obtaining the generalized Laplace matrix according to the diagonal matrix and the Laplace matrix;
wherein, the difference between the degree matrix and the weight matrix is used as a Laplacian matrix;
using the sum of the diagonal matrix and the Laplace matrix as the generalized Laplace matrix;
wherein the diagonal element d of the degree matrix i =∑ j W ij, wherein ,di The ith diagonal element of the degree matrix, W ij Representing the weight corresponding to the edge from the ith point to the jth point in the target sub-point cloud; 1.ltoreq.i.ltoreq.M, 1.ltoreq.j.ltoreq.M, i, j, M being an integer, M being the total number of points included in the target sub-point cloud;
the diagonal matrix is generated from Euclidean distances between target points in the target sub-point cloud and corresponding points of the target points.
2. The method of claim 1, wherein clustering the point cloud data to be processed of the current frame to obtain a plurality of sub point clouds comprises:
performing voxel conversion on the point cloud data to be processed to obtain point cloud voxels;
and clustering the point cloud voxels to obtain a plurality of sub-point clouds.
3. The method of claim 1, wherein the obtaining a weight matrix according to euclidean distances between a plurality of point pairs in the target sub-point cloud comprises:
calculating the Euclidean distance between the ith point and the jth point in the target sub-point cloud;
calculating weights according to the following formula, and forming the weight matrix by using the weights:
Figure FDA0004087268170000021
wherein ,Wij Representing the weight corresponding to the edge from the ith point to the jth point in the target sub-point cloud; distance represents the Euclidean distance between the ith point and the jth point; sigma is a constant not equal to 0, representing the adjustment parameter; and i is more than or equal to 1 and less than or equal to M, j is more than or equal to 1 and less than or equal to M, i, j, M is an integer, and M is the total number of points included in the target sub-point cloud.
4. The method of claim 1, wherein the generating a diagonal matrix comprises:
determining a reference point cloud of the target sub-point cloud in the reference frame;
generating the diagonal matrix based on Euclidean distance between each point in the target sub-point cloud and the corresponding point of each point in the reference point cloud;
the value on the ith diagonal of the diagonal matrix is the inverse of the Euclidean distance between the ith point and a point p, wherein the point p is the corresponding point of the ith point in the reference point cloud.
5. The method of claim 4, wherein the determining the reference point cloud of the target sub-point cloud in the reference frame comprises:
and determining a reference point cloud of the target sub-point cloud in the reference frame by using an iterative nearest point algorithm.
6. The method of claim 1, wherein performing inter-prediction and graph fourier residual transforms on the plurality of sub-point clouds using the generalized laplacian matrix, respectively, comprises:
acquiring an attribute predicted value of the reference frame on the target attribute of the current frame;
generating a residual error of the target attribute of the current frame according to the attribute value of the target attribute of the current frame and the attribute predicted value of the target attribute of the reference frame to the current frame;
and transforming the residual error of the target attribute of the current frame based on the generalized Laplace matrix.
7. The method of claim 6, wherein the obtaining the attribute prediction value of the target attribute of the reference frame to the current frame comprises:
obtaining an attribute predicted value of the target attribute of the reference frame to the current frame according to the following formula:
Figure FDA0004087268170000031
wherein ,
Figure FDA0004087268170000032
an attribute predictor representing a target attribute of the reference frame to the current frame,x t-1 attribute values representing target attributes of the reference frame, lg represents a generalized Laplacian matrix.
8. The method of claim 6, wherein generating the residual of the target attribute of the current frame based on the attribute value of the target attribute of the current frame and the attribute prediction value of the target attribute of the reference frame to the target attribute of the current frame, comprises:
and using the difference between the attribute value of the target attribute of the current frame and the attribute predicted value of the target attribute of the reference frame to the current frame as the residual error.
9. The method of claim 6, wherein transforming the residual of the target property of the current frame based on the generalized laplacian matrix comprises:
obtaining a transformation matrix by using the generalized Laplace matrix;
and transforming the residual error of the target attribute of the current frame by using the transformation matrix.
10. The method of claim 6, wherein the deriving a transformation matrix using the generalized laplacian matrix comprises:
solving the following formula to obtain a transformation matrix:
Figure FDA0004087268170000033
wherein Lg represents a generalized Laplace matrix,
Figure FDA0004087268170000034
representing the transformation matrix.
11. The method of claim 9, wherein transforming the residual of the target property of the current frame using the transformation matrix comprises:
the transformation result is obtained using the following formula:
Figure FDA0004087268170000035
wherein, theta represents the transformation result,
Figure FDA0004087268170000036
representing a transformation matrix, delta representing the residual of the target attribute of the current frame.
12. A decoding method applied to a decoding device, the method comprising:
acquiring a coded code stream;
performing graph Fourier inverse transformation based on Euclidean distance weight on the coded code stream to obtain a transformation result;
based on the transformation result, obtaining a decoded code stream;
the coding code stream is obtained by coding a result of inter-frame prediction and graph Fourier residual transformation of the sub-point cloud by using a generalized Laplace matrix by coding equipment;
the generalized Laplace matrix is generated by the following steps:
obtaining a weight matrix according to Euclidean distances among a plurality of point pairs in the target sub-point cloud;
obtaining a Laplace matrix according to the degree matrix and the weight matrix;
generating a diagonal matrix;
obtaining the generalized Laplace matrix according to the diagonal matrix and the Laplace matrix;
wherein, the difference between the degree matrix and the weight matrix is used as a Laplacian matrix;
using the sum of the diagonal matrix and the Laplace matrix as the generalized Laplace matrix;
wherein the diagonal element d of the degree matrix i =∑ j W ij, wherein ,di The ith diagonal element of the degree matrix, W ij Representing the weight corresponding to the edge from the ith point to the jth point in the target sub-point cloud; 1.ltoreq.i.ltoreq.M, 1.ltoreq.j.ltoreq.M, i, j, M being an integer, M being the total number of points included in the target sub-point cloud;
the diagonal matrix is generated from Euclidean distances between target points in the target sub-point cloud and corresponding points of the target points.
13. The method of claim 12, wherein performing an inverse fourier transform on the encoded code stream based on euclidean distance weights to obtain a transform result comprises:
performing inverse quantization on the coded code stream;
and carrying out graph Fourier inverse transformation based on Euclidean distance weight on the code stream after inverse quantization to obtain a transformation result.
14. The method of claim 13, wherein performing an inverse fourier transform on the encoded code stream based on euclidean distance weights to obtain a transform result comprises:
performing graph Fourier transform based on Euclidean distance weight on the code stream after inverse quantization by using the following formula:
Figure FDA0004087268170000041
wherein ,
Figure FDA0004087268170000042
representing the inverse transform residual value,/->
Figure FDA0004087268170000043
Representing a transformation matrix->
Figure FDA0004087268170000044
The quantized residual value representing the target property of the current frame, epsilon represents the dequantization coefficient.
15. An encoding device, comprising:
the first acquisition module is used for clustering the point cloud data to be processed of the current frame to obtain a plurality of sub point clouds;
the first generation module is used for generating a generalized Laplace matrix for any target sub-point cloud in the multiple sub-point clouds according to Euclidean distances between multiple point pairs in the target sub-point cloud and Euclidean distances between a target point in the target sub-point cloud and a corresponding point of the target point;
the first transformation module is used for carrying out inter-frame prediction and graph Fourier residual transformation on the target sub-point cloud by utilizing the generalized Laplace matrix;
the first coding module is used for respectively quantizing and coding the transformed multiple sub-point clouds to obtain a coded code stream;
the corresponding point is located in a reference point cloud of the target sub-point cloud, and the reference point cloud is located in a reference frame of the current frame;
wherein the first generation module comprises:
the first acquisition sub-module is used for obtaining a weight matrix according to Euclidean distances among a plurality of point pairs in the target sub-point cloud; the second acquisition sub-module is used for obtaining a Laplace matrix according to the degree matrix and the weight matrix; the first generation submodule is used for generating a diagonal matrix; the second generation submodule is used for obtaining the generalized Laplace matrix according to the diagonal matrix and the Laplace matrix;
the second obtaining submodule is used for using the difference between the degree matrix and the weight matrix as a Laplacian matrix; the second generating submodule is used for utilizing the sum of the diagonal matrix and the Laplace matrix as the generalized Laplace matrix;
wherein the diagonal element d of the degree matrix i =∑ j W ij, wherein ,di The ith diagonal element of the degree matrix, W ij Representing the weight corresponding to the edge from the ith point to the jth point in the target sub-point cloud; 1.ltoreq.i.ltoreq.M, 1.ltoreq.j.ltoreq.M, i, j, M being an integer, M being the total number of points included in the target sub-point cloud;
the diagonal matrix is generated from Euclidean distances between target points in the target sub-point cloud and corresponding points of the target points.
16. A decoding apparatus, comprising:
the first acquisition module is used for acquiring the coded code stream;
the first transformation module is used for carrying out graph Fourier inverse transformation based on Euclidean distance weight on the coded code stream to obtain a transformation result;
the first decoding module is used for obtaining a decoding code stream based on the transformation result;
the coding code stream is obtained by coding a result of inter-frame prediction and graph Fourier residual transformation of the sub-point cloud by using a generalized Laplace matrix by coding equipment;
the generalized Laplace matrix is generated by the following steps:
obtaining a weight matrix according to Euclidean distances among a plurality of point pairs in the target sub-point cloud;
obtaining a Laplace matrix according to the degree matrix and the weight matrix;
generating a diagonal matrix;
obtaining the generalized Laplace matrix according to the diagonal matrix and the Laplace matrix;
wherein, the difference between the degree matrix and the weight matrix is used as a Laplacian matrix;
using the sum of the diagonal matrix and the Laplace matrix as the generalized Laplace matrix;
wherein the diagonal element d of the degree matrix i =∑ j W ij, wherein ,di Ith pair of representation degree matrixAngle line element, W ij Representing the weight corresponding to the edge from the ith point to the jth point in the target sub-point cloud; 1.ltoreq.i.ltoreq.M, 1.ltoreq.j.ltoreq.M, i, j, M being an integer, M being the total number of points included in the target sub-point cloud;
the diagonal matrix is generated from Euclidean distances between target points in the target sub-point cloud and corresponding points of the target points.
17. An electronic device, comprising: a memory, a processor, and a program stored on the memory and executable on the processor; it is characterized in that the method comprises the steps of,
the processor for reading a program in a memory to implement the steps in the encoding method according to any one of claims 1 to 11; or to implement the steps in the decoding method according to any one of claims 12 to 14.
18. A readable storage medium storing a program, wherein the program when executed by a processor implements the steps of the encoding method according to any one of claims 1 to 11; or to implement the steps in the decoding method according to any one of claims 12 to 14.
CN202111160289.XA 2021-09-30 2021-09-30 Encoding method, decoding method, device, equipment and readable storage medium Active CN113766229B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202111160289.XA CN113766229B (en) 2021-09-30 2021-09-30 Encoding method, decoding method, device, equipment and readable storage medium
PCT/CN2022/123245 WO2023051783A1 (en) 2021-09-30 2022-09-30 Encoding method, decoding method, apparatus, device, and readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111160289.XA CN113766229B (en) 2021-09-30 2021-09-30 Encoding method, decoding method, device, equipment and readable storage medium

Publications (2)

Publication Number Publication Date
CN113766229A CN113766229A (en) 2021-12-07
CN113766229B true CN113766229B (en) 2023-04-28

Family

ID=78798550

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111160289.XA Active CN113766229B (en) 2021-09-30 2021-09-30 Encoding method, decoding method, device, equipment and readable storage medium

Country Status (2)

Country Link
CN (1) CN113766229B (en)
WO (1) WO2023051783A1 (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113766229B (en) * 2021-09-30 2023-04-28 咪咕文化科技有限公司 Encoding method, decoding method, device, equipment and readable storage medium
WO2023173238A1 (en) * 2022-03-12 2023-09-21 Oppo广东移动通信有限公司 Encoding method, decoding method, code stream, encoder, decoder, and storage medium
CN114785998A (en) * 2022-06-20 2022-07-22 北京大学深圳研究生院 Point cloud compression method and device, electronic equipment and storage medium
WO2023245982A1 (en) * 2022-06-20 2023-12-28 北京大学深圳研究生院 Point cloud compression method and apparatus, electronic device, and storage medium
WO2024077911A1 (en) * 2022-10-13 2024-04-18 Beijing Bytedance Network Technology Co., Ltd. Method, apparatus, and medium for point cloud coding
CN116797625B (en) * 2023-07-20 2024-04-19 无锡埃姆维工业控制设备有限公司 Monocular three-dimensional workpiece pose estimation method

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108171761B (en) * 2017-12-13 2020-10-16 北京大学 Point cloud intra-frame coding method and device based on Fourier image transformation
WO2020197086A1 (en) * 2019-03-25 2020-10-01 엘지전자 주식회사 Point cloud data transmission device, point cloud data transmission method, point cloud data reception device, and point cloud data reception method
CN112385238B (en) * 2019-07-10 2023-03-24 深圳市大疆创新科技有限公司 Data encoding method, data decoding method, equipment and storage medium
CN110572655B (en) * 2019-09-30 2023-01-10 北京大学深圳研究生院 Method and equipment for encoding and decoding point cloud attribute based on neighbor weight parameter selection and transmission
CN110418135B (en) * 2019-08-05 2022-05-27 北京大学深圳研究生院 Point cloud intra-frame prediction method and device based on neighbor weight optimization
CN113766229B (en) * 2021-09-30 2023-04-28 咪咕文化科技有限公司 Encoding method, decoding method, device, equipment and readable storage medium

Also Published As

Publication number Publication date
WO2023051783A1 (en) 2023-04-06
CN113766229A (en) 2021-12-07

Similar Documents

Publication Publication Date Title
CN113766229B (en) Encoding method, decoding method, device, equipment and readable storage medium
CN113678466A (en) Method and apparatus for predicting point cloud attribute encoding
US8805097B2 (en) Apparatus and method for coding a three dimensional mesh
CN111641832A (en) Encoding method, decoding method, device, electronic device and storage medium
US20220303537A1 (en) Method and device for encoding/decoding video signal by using optimized conversion based on multiple graph-based model
CN114245896A (en) Vector query method and device, electronic equipment and storage medium
US20240202982A1 (en) 3d point cloud encoding and decoding method, compression method and device based on graph dictionary learning
CN107113426B (en) Method and apparatus for performing graph-based transformations using generalized graph parameters
CN115130571A (en) Feature encoding method, feature decoding method, feature encoding device, feature decoding device, electronic device, and storage medium
CN114332500A (en) Image processing model training method and device, computer equipment and storage medium
CN104392207A (en) Characteristic encoding method for recognizing digital image content
JP6748022B2 (en) Division shape determining apparatus, learning apparatus, division shape determining method, and division shape determining program
US10715802B2 (en) Method for encoding/decoding video signal by using single optimized graph
Hajizadeh et al. Predictive compression of animated 3D models by optimized weighted blending of key‐frames
CN114565693A (en) Image generation and neural network training method, device, equipment and medium
EP4216553A1 (en) Point cloud decoding and encoding method, and decoder, encoder and encoding and decoding system
US20230274467A1 (en) Image encoding apparatus, image encoding method and program
Akutsu et al. End-to-End Deep ROI Image Compression
Valarmathi et al. Iteration-free fractal image compression using Pearson’s correlation coefficient-based classification
CN105706447A (en) Moving image coding device, moving image decoding device, moving image coding method, moving image decoding method, and program
Thakur et al. Image‐independent optimal non‐negative integer bit allocation technique for the DCT‐based image transform coders
CN114189692B (en) Point cloud attribute coding method and decoding method based on continuous subspace diagram transformation
CN113902844A (en) Method, apparatus, electronic device, medium, and program product for point cloud encoding
CN117459727B (en) Image processing method, device and system, electronic equipment and storage medium
JP5561611B2 (en) Image coding method, image coding apparatus, and image coding program

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant