CN104915978A - Realistic animation generation method based on Kinect - Google Patents

Realistic animation generation method based on Kinect Download PDF

Info

Publication number
CN104915978A
CN104915978A CN201510341839.6A CN201510341839A CN104915978A CN 104915978 A CN104915978 A CN 104915978A CN 201510341839 A CN201510341839 A CN 201510341839A CN 104915978 A CN104915978 A CN 104915978A
Authority
CN
China
Prior art keywords
frame
matrix
skeleton
kinect
sequence
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510341839.6A
Other languages
Chinese (zh)
Other versions
CN104915978B (en
Inventor
李坤
杨敬钰
刘雷杰
宋晓林
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Lingyun Shixun Technology Co.,Ltd.
Original Assignee
Tianjin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin University filed Critical Tianjin University
Priority to CN201510341839.6A priority Critical patent/CN104915978B/en
Publication of CN104915978A publication Critical patent/CN104915978A/en
Application granted granted Critical
Publication of CN104915978B publication Critical patent/CN104915978B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Abstract

The invention belongs to the field of computer vision and computer graphics, and provides a method which can ensure a newly-generated video sequence to be vivid and reliable and solve the problem of high requirements on an experimental device. According to the technical scheme adopted by the invention, the realistic animation generation method based on Kinect uses a single Kinect v2.0 to carry out static three dimensional model and color depth skeleton information acquisition on a person. A marker-free motion capture method is used, skeleton motion information of each frame and a three dimensional model after deformation are obtained. After a user designates a new action, on the basis of the built data collection, a video for the new action by the person is generated through the proposed sparse representation-based low rank matrix filling texture synthesis method. The method of the invention is mainly applied to video processing.

Description

Based on the Realistic animation generation method of body sense camera Kinect
Technical field
The invention belongs to computer vision and field of Computer Graphics, particularly a kind of new element video method adopting a Kinect somatosensory camera to generate strong sense of reality.
Background technology
For a long time, the research for new animation producing method did not all stop always, and its solution method also emerges in an endless stream.These methods are divided three classes substantially, and first method utilizes skeleton and model rendering the earliest, carry out Model Reconstruction use Effective Pictures By A Computer to obtain new animated video sequence to simple scenario.But this method is better for the object effect of rigidity, but cannot form the video sequence of strong sense of reality for nonrigid object.Second method utilizes the method for physical modeling to synthesize the three-dimensional model of a small amount of high precision, high realism off-line, afterwards with these three-dimensional models for sample, synthesize under various human motion, there is the three-dimensional model of geometric detail true to nature.But the control of these class methods to the setting of apparel construction and material parameters is all very limited, its apparel construction generating motion corresponding all cannot conform to livery a certain in reality completely with material.The third method is the method based on Video Textures, extract by some frames concentrated existing video sequence and recombinate, finally generate new video sequence, this method can form the comparatively real new animation result of effect for simple scene, but can not process comparatively complicated scene.Marx's Planck Computer Sciences Laboratory (Feng Xu, Yebin Liu, Carsten Stoll, James Tompkin, Gaurav Bharaj, Qionghai Dai, Hans-Peter Seidel, Jan Kautz, Christian Theobalt.Video-based characters:creating new human performances from a multi-view video database.ACM Transactions on Graphics (TOG), 2011, 30 (4): 32.) by adopting various visual angles camera array to carry out data acquisition, model and video texture is utilized to generate the very strong new animated video sequence of the sense of reality.But this scheme cost is high, high to requirement for experiment condition, is unfavorable for Industry Promotion.
Body sense camera Kinect (Zhengyou Zhang.Microsoft kinect sensor and its effect.MultiMedia, IEEE, 2012,19 (2): 4-10.) appearance, making to realize Realistic animation based on cheap apparatus becomes possibility.In November, 2010, Kinect was gone on the market by Microsoft, became the focus of whole world E-consumer soon.It can catch chromatic information, depth information, skeleton motion information, voice messaging etc. in real time.And the research application based on it is rapidly broken out.The people such as Tong (Tong J, Zhou J, Liu L, et al.Scanning 3D full human bodies using kinects.IEEE Transactions onVisualization and Computer Graphics, 2012,18 (4): 643-650.) Kinect is utilized to achieve three-dimensional reconstruction for human body. deng people ( m, Martinek M, Greiner G, et al..Automatic reconstruction of personalized avatars from 3D face scans.Computer Animation and Virtual Worlds, 2011,22 (2 ?3): 195-202.) the real character's mask achieved based on Kinect is rebuild.2014, in Kinect bis-generation, was issued, and was greatly improved than precision before.Can more accurate Model Reconstruction being carried out based on Kinect v2.0, the cromogram of more real 1080p can be preserved in real time, and precision be the depth map of 1mm simultaneously, providing good data basis for generating based on the Realistic animation of Kinect v2.0.
Summary of the invention
For overcoming the deficiencies in the prior art, realizing can ensureing to generate the true to nature credible of new video sequences, the method that experimental facilities requires high difficult point can be solved again.For this reason, the technical scheme that the present invention takes is, based on the Realistic animation generation method of body sense camera Kinect, utilizes single Kinect v2.0 personage to be carried out to the collection of static three-dimensional model and color depth framework information.Utilize unmarked motion capture method, obtain the skeleton motion information of every frame and the three-dimensional model after being out of shape.After user specifies a new element, based on set up data set, pass through the texture synthesis method that the proposed low-rank matrix based on rarefaction representation fills and generate the video that this personage does new element; Specifically comprise the following steps:
1) based on the method for Kinect Fusion, personage is carried out to the reconstruction of static three-dimensional model, and at this model namely with reference on the basis of attitude mode, manually embed skeleton;
2) Database: the frame sequence of color video, deep video and correspondence when adopting single this personage of Kinect v2.0 cameras capture to carry out various elemental motion; IK (inverse kinematics) algorithm is utilized to calculate the kinematic parameter of each frame skeleton node; Adopt linear hybrid skinning method (linear blend skinning) to calculate skin weight and be out of shape and obtain three-dimensional model corresponding to every frame with reference to attitude mode; Model after adopting depth information to correct distortion makes it more accurate; Data set only needs to set up once, can generate the video sequence of the new element of this personage based on this database;
3) user specifies the frame sequence of new action, and the topological structure of this skeleton is identical with the topological structure of data centralization skeleton;
4) for every frame of fresh target action finds similar candidate frame in data centralization;
5) utilize Moving Least Squares method that the human body parts of the frame retrieved is carried out deformation to be estimated at first, then utilize weighting low-rank matrix to fill up interpolation method and be optimized;
6) by step 5) final image that obtains puts together the new video sequences obtaining requiring.
The design of database building plan, specifically comprises the following steps:
2-1) for the seizure of color video, deep video and framework information, Kinect v2.0 is utilized to carry out Real-time Collection preservation, arranging Kinect v2.0 frame per second is 30fps, and color image resolution is 1920 × 1080, and depth image resolution is set to 512 × 424; The collection of frame sequence comprises the preservation of the preservation of node space three-dimensional information and the mapping relations with color depth information; The personage's elemental motion gathered comprises as walked, running, stride, wave, stretch, frontly to play, side is played, chest expanding, lateral bending and time;
2-2) utilize the three-dimensional information of each frame frame sequence collected, input each node space three dimensional local information and carry out IK Algorithm for Solving, obtain the rotation translation matrix of each node, that comprise its overall situation with kinematic matrix that is local;
2-3) calculate skin weight with linear hybrid skinning method (linear blend skinning) and distortion obtains three-dimensional model corresponding to every frame with reference to attitude mode;
2-4) adopting based on non-rigid surface's alignment schemes of rarefaction representation, utilize depth information, by step 2-3) three-dimensional model that obtains is out of shape further, makes it to coincide with depth information and more accurate.
The design of similar frame search method, specifically comprises the following steps:
4-1) first the skeleton in target skeleton and database being snapped in same world coordinates, rotates skeleton and make it towards same direction, then obtaining candidate frame by minimizing the retrieval of following energy function:
E ( F ) = Σ i ∈ I ( F ) D s ( S d i - S q i ) + α ( Σ i ∈ I ( F ) D s ( S d i - S d i - 1 ) - Σ i ∈ M D s ( S q i - S q i - 1 ) ) - β N ( F ) - - - ( 1 )
Wherein F is unknown candidate frame, and N (F) is candidate sequence frame number, and I (F) is the frame number of database original series, and M is target sequence frame number, the skeleton knot vector of database sequence i-th frame, the skeleton knot vector of database sequence i-th-1 frame, the skeleton knot vector of target sequence i-th frame, be the skeleton knot vector of target sequence i-th-1 frame, α, β represent the weight allocation between two constraints, and skeleton similarity is defined as follows:
D s ( S m - S n ) = Σ j = 1 J ( S m j - S n j ) 2 σ j - - - ( 2 )
Wherein m and n can represent d or q in formula (1) respectively, S jbe the position of a jth skeleton node, and J is interstitial content, σ jit is the variance of node j position in a database; In formula (1), Section 1 is space constraint, ensures the similarity of candidate frame and target frame; Two item constraints are below to ensure temporal continuity, avoid shake.
Synthesis target frame sequence is the comparatively popular sparse representation method based on occurring recently, removes noise step as follows:
5-1) 3 d surface model of every frame is all obtained by same three-dimensional model deformation in assumption database and in target sequence, and they have identical summit and Topology connection; Dividing processing is carried out to three-dimensional model, is always divided into 16 parts and by mapping, the character image of target frame and retrieval frame is also divided into 16 parts; Calculate target frame and the respective pixel of retrieval frame, be guide with corresponding point, utilize Moving Least Squares method will retrieve frame distortion;
5-2) step 5-1) early results that obtains may comprise some missing information, because retrieval frame can not comprise all information of target frame, the method adopting the matrix based on rarefaction representation to fill up carries out interpolation to these regions, finally obtain every two field picture, synthesize new video sequence; For the method that the matrix based on rarefaction representation is filled up, first will ask for priority in the edge on current imperfect image, the sort criteria of priority considers texture and the depth information of block; Then, fill up according to the priority orders of each piece, for some blocks to be filled up, find the K similar to it similar piece; If the pixel P repaired iresiding w × w block B 0, and K similar piece is designated B respectively 1, B 2, B 3... B k, obtain a matrix D with these matrixes as each row.The method filled up based on the matrix of rarefaction representation is expressed as following optimization problem:
min A,Erank(A)+λ||W°E|| 0s.t.P Ω(D)=P Ω(A+E) (3)
Wherein, A is original matrix to be asked, and E is nonzero element number, and D is observing matrix, and W is the weight matrix considering similar piece of similarity, and λ is the weight of noise size; The order of rank (A) representing matrix A, || || 00 norm of representing matrix, " ° " represents that two matrix corresponding elements are multiplied, and Ω is the index set of known elements, P Ωfor projecting to the operator in Ω territory.Above optimization problem is NP hard (Non-deterministic Polynomial-time hard) problem, therefore uses the nuclear norm of matrix || || *rank of matrix is replaced to minimize, by matrix 1 norm || || 1replace 0 norm, above-mentioned problem is converted into:
min A,E||A|| *+λ||W°E|| 1,s.t.P Ω(D)=P Ω(A+E) (4)
After solving out and obtaining matrix A, take out first row, be restructured as the block of pixels size of w × w, be the current block that method that the matrix based on rarefaction representation fills up recovers, the block of pixels of other positions by that analogy.
Compared with the prior art, technical characterstic of the present invention and effect:
The problems such as the experiment condition needs that inventive process avoids polyphaser collecting work room and the sense of reality having new animation producing method is poor, action limitation, by adopting a Kinect v2.0 collected by camera, formation base database, by retrieval coupling, distortion, repairing etc., realize the new video sequences of the strong sense of reality generating any action.There is following characteristics:
1, equipment is cheap and experiment condition simple, is easy to realize.
2, the sense of reality is very strong, can obtain effect comparatively true to nature for tiny texture variations.
3, database movement collection only need be carried out once, based on this data set, can generate the new video sequences that validity that action edits arbitrarily is very high.
The present invention can adopt single Kinect v2.0 to carry out three-dimensional reconstruction and data acquisition, simple in realization, and has the effect allowing people comparatively be satisfied with.The method proposed has good extensibility: can by utilizing the flexible change on the increase of the information of multiple Kinect v2.0 fulfillment database and visual angle.
Accompanying drawing explanation
The present invention above-mentioned and/or additional aspect and advantage will become obvious and easy understand from the following description of the accompanying drawings of embodiments, wherein:
Fig. 1 is the Realistic animation generation method flow diagram of the embodiment of the present invention based on body sense camera Kinect v2.0;
Fig. 2 is that the embodiment of the present invention adopts proposed invention method to generate the result of new element video sequence;
Fig. 3 is the result result of three kinds of different new elements being placed on a certain frame of the video synthesized in captured background video.
Embodiment
The present invention is intended to the deficiency overcoming existing new animation producing scheme, realizes can ensureing to generate the true to nature credible of new video sequences, can solve again the method that experimental facilities requires high difficult point.The present invention is by using a Kinect v2.0 as the equipment of acquisition database, and by carrying out the process such as motion tracking to data, set up elemental motion database, based on this, by generating the realistic video of new element based on the texture synthesis method of rarefaction representation.
The present invention proposes a kind of Realistic animation generation method based on body sense camera Kinect v2.0, by reference to the accompanying drawings and embodiment be described in detail as follows:
The present invention utilizes single Kinect v2.0 camera personage to be carried out to the collection of static three-dimensional model and color depth framework information.Utilize unmarked motion capture method, obtain the skeleton motion information of every frame and the three-dimensional model after being out of shape.After user specifies a new element, based on set up data set, pass through the texture synthesis method that the proposed low-rank matrix based on rarefaction representation fills and generate the video that this personage does new element.As shown in Figure 1, for the Realistic animation based on body sense camera Kinect v2.0 of the embodiment of the present invention generates method flow diagram, comprise the following steps:
1) based on the method for Kinect Fusion (project that the real-time static three-dimensional model that Kinect official issues is rebuild), personage is carried out to the reconstruction of static three-dimensional model, and on the basis of this model (with reference to attitude mode), manually embed skeleton;
2) Database: the frame sequence of color video, deep video and correspondence when adopting single this personage of Kinect v2.0 cameras capture to carry out various elemental motion; IK (inverse kinematics) algorithm is utilized to calculate the kinematic parameter of each frame skeleton node; Adopt linear hybrid skinning method (linear blend skinning) to calculate skin weight and be out of shape and obtain three-dimensional model corresponding to every frame with reference to attitude mode; Model after adopting depth information to correct distortion makes it more accurate.Data set only needs to set up once, can generate the video sequence of the new element of this personage based on this database;
3) user specifies the frame sequence of new action.The topological structure of this skeleton is identical with the topological structure of data centralization skeleton;
4) for every frame of fresh target action finds similar candidate frame in data centralization;
5) utilize Moving Least Squares method that the human body parts of the frame retrieved is carried out deformation to be estimated at first, then utilize weighting low-rank matrix to fill up interpolation method and be optimized;
6) by step 5) final image that obtains puts together the new video sequences obtaining requiring.
The design of database building plan, specifically comprises the following steps:
2-1) for the seizure of color video, deep video and framework information, Kinect v2.0 is utilized to carry out Real-time Collection preservation, arranging Kinect v2.0 frame per second is 30fps, and color image resolution is 1920 × 1080, and depth image resolution is set to 512 × 424; The collection of frame sequence comprises the preservation of the preservation of node space three-dimensional information and the mapping relations with color depth information; The personage's elemental motion gathered comprises as walked, running, stride, wave, stretch, frontly to play, side is played, chest expanding, lateral bending and time etc.;
2-2) utilize the three-dimensional information of each frame frame sequence collected, input each node space three dimensional local information and carry out IK Algorithm for Solving, obtain the rotation translation matrix of each node, that comprise its overall situation with kinematic matrix that is local;
2-3) calculate skin weight with linear hybrid skinning method (linear blend skinning) and distortion obtains three-dimensional model corresponding to every frame with reference to attitude mode;
2-4) adopting based on non-rigid surface's alignment schemes of rarefaction representation, utilize depth information, by step 2-3) three-dimensional model that obtains is out of shape further, makes it to coincide with depth information and more accurate.
The design of similar frame search method, specifically comprises the following steps:
4-1) first the skeleton in target skeleton and database being snapped in same world coordinates, rotates skeleton and make it towards same direction, then obtaining candidate frame by minimizing the retrieval of following energy function:
E ( F ) = Σ i ∈ I ( F ) D s ( S d i - S q i ) + α ( Σ i ∈ I ( F ) D s ( S d i - S d i - 1 ) - Σ i ∈ M D s ( S q i - S q i - 1 ) ) - β N ( F ) - - - ( 1 )
Wherein F is unknown candidate frame, and N (F) is candidate sequence frame number, and I (F) is the frame number of database original series, and M is target sequence frame number, the skeleton knot vector of database sequence i-th frame, the skeleton knot vector of database sequence i-th-1 frame, the skeleton knot vector of target sequence i-th frame, be the skeleton knot vector of target sequence i-th-1 frame, α, β represent the weight allocation between two constraints, and skeleton similarity is defined as follows:
D s ( S m - S n ) = Σ j = 1 J ( S m j - S n j ) 2 σ j - - - ( 2 )
Wherein m and n can represent d or q in formula (1) respectively, S jbe the position of a jth skeleton node, and J is interstitial content, σ jit is the variance of node j position in a database; In formula (1), Section 1 is space constraint, ensures the similarity of candidate frame and target frame; Two item constraints are below to ensure temporal continuity, avoid shake.
Synthesis target frame sequence is the comparatively popular sparse representation method based on occurring recently: low-rank matrix is filled up.The method can not only repair image lack part, can also remove noise; Concrete grammar comprises the following steps:
5-1) 3 d surface model of every frame is all obtained by same three-dimensional model deformation in assumption database and in target sequence, and they have identical summit and Topology connection; Dividing processing is carried out to three-dimensional model, is always divided into 16 parts and by mapping, the character image of target frame and retrieval frame is also divided into 16 parts; Calculate target frame and the respective pixel of retrieval frame, be guide with corresponding point, utilize Moving Least Squares method will retrieve frame distortion;
5-2) step 5-1) early results that obtains may comprise some missing information, because retrieval frame can not comprise all information of target frame.The method that we adopt the matrix based on rarefaction representation to fill up carries out interpolation to these regions, finally obtains every two field picture, synthesizes new video sequence; For the method that the matrix based on rarefaction representation is filled up, first will ask for priority in the edge on current imperfect image, the sort criteria of priority considers texture and the depth information of block.Then, fill up according to the priority orders of each piece.For some blocks to be filled up, find the K similar to it similar piece.If the pixel P repaired iresiding w × w block B 0, and K similar piece is designated B respectively 1, B 2, B 3... B k, obtain a matrix D with these matrixes as each row.The method filled up based on the matrix of rarefaction representation is expressed as following optimization problem:
min A,Erank(A)+λ||W°E|| 0s.t.P Ω(D)=P Ω(A+E) (3)
Wherein, A is original matrix to be asked, and E is nonzero element number, and D is observing matrix, and W is the weight matrix considering similar piece of similarity, and λ is the weight of noise size; The order of rank (A) representing matrix A, || || 00 norm of representing matrix, " ° " represents that two matrix corresponding elements are multiplied, and Ω is the index set of known elements, P Ωfor projecting to the operator in Ω territory.Above optimization problem is NP hard (Non-deterministic Polynomial-time hard) problem, therefore uses the nuclear norm of matrix || || *rank of matrix is replaced to minimize, by matrix 1 norm || || 1replace 0 norm, above-mentioned problem can be converted into:
min A,E||A|| *+λ||W°E|| 1,s.t.P Ω(D)=P Ω(A+E) (4)
After solving out and obtaining matrix A, take out first row, be restructured as the block of pixels size of w × w, be the current block that method that the matrix based on rarefaction representation fills up recovers.The block of pixels of other positions by that analogy.

Claims (4)

1. one kind based on the Realistic animation generation method of body sense camera Kinect, it is characterized in that, utilizes single Kinect v2.0 personage to be carried out to the collection of static three-dimensional model and color depth framework information.Utilize unmarked motion capture method, obtain the skeleton motion information of every frame and the three-dimensional model after being out of shape, after user specifies a new element, based on set up data set, pass through the texture synthesis method that the proposed low-rank matrix based on rarefaction representation fills and generate the video that this personage does new element; Specifically comprise the following steps:
1) based on the method for Kinect Fusion, personage is carried out to the reconstruction of static three-dimensional model, and at this model namely with reference on the basis of attitude mode, manually embed skeleton;
2) Database: the frame sequence of color video, deep video and correspondence when adopting single this personage of Kinect v2.0 cameras capture to carry out various elemental motion; IK (inverse kinematics) algorithm is utilized to calculate the kinematic parameter of each frame skeleton node; Adopt linear hybrid skinning method (linear blend skinning) to calculate skin weight and be out of shape and obtain three-dimensional model corresponding to every frame with reference to attitude mode; Model after adopting depth information to correct distortion makes it more accurate; Data set only needs to set up once, can generate the video sequence of the new element of this personage based on this database;
3) user specifies the frame sequence of new action, and the topological structure of this skeleton is identical with the topological structure of data centralization skeleton;
4) for every frame of fresh target action finds similar candidate frame in data centralization;
5) utilize Moving Least Squares method that the human body parts of the frame retrieved is carried out deformation to be estimated at first, then utilize weighting low-rank matrix to fill up interpolation method and be optimized;
6) by step 5) final image that obtains puts together the new video sequences obtaining requiring.
2., as claimed in claim 1 based on the Realistic animation generation method of body sense camera Kinect, it is characterized in that, the design of database building plan, specifically comprises the following steps:
2-1) for the seizure of color video, deep video and framework information, Kinect v2.0 is utilized to carry out Real-time Collection preservation, arranging Kinect v2.0 frame per second is 30fps, and color image resolution is 19201080, and depth image resolution is set to 512424; The collection of frame sequence comprises the preservation of the preservation of node space three-dimensional information and the mapping relations with color depth information; The personage's elemental motion gathered comprises as walked, running, stride, wave, stretch, frontly to play, side is played, chest expanding, lateral bending and time;
2-2) utilize the three-dimensional information of each frame frame sequence collected, input each node space three dimensional local information and carry out IK Algorithm for Solving, obtain the rotation translation matrix of each node, that comprise its overall situation with kinematic matrix that is local;
2-3) calculate skin weight with linear hybrid skinning method (linear blend skinning) and distortion obtains three-dimensional model corresponding to every frame with reference to attitude mode;
2-4) adopting based on non-rigid surface's alignment schemes of rarefaction representation, utilize depth information, by step 2-3) three-dimensional model that obtains is out of shape further, makes it to coincide with depth information and more accurate.
3., as claimed in claim 1 based on the Realistic animation generation method of body sense camera Kinect, it is characterized in that, the design of similar frame search method, specifically comprises the following steps:
4-1) first the skeleton in target skeleton and database being snapped in same world coordinates, rotates skeleton and make it towards same direction, then obtaining candidate frame by minimizing the retrieval of following energy function:
E ( F ) = Σ i ∈ I ( F ) D s ( S d i - S q i ) + α ( Σ i ∈ I ( F ) D s ( S d i - S d i - 1 ) - Σ i ∈ M D s ( S q i - S q i - 1 ) ) - β N ( F ) - - - ( 1 )
Wherein F is unknown candidate frame, and N (F) is candidate sequence frame number, and I (F) is the frame number of database original series, and M is target sequence frame number, the skeleton knot vector of database sequence i-th frame, the skeleton knot vector of database sequence i-th-1 frame, the skeleton knot vector of target sequence i-th frame, be the skeleton knot vector of target sequence i-th-1 frame, α, β represent the weight allocation between two constraints, and skeleton similarity is defined as follows:
D s ( S m - S n ) = Σ j = 1 J ( S m j - S n j ) 2 σ j - - - ( 2 )
Wherein m and n can represent d or q in formula (1) respectively, S jthe position of the skeleton node, but interstitial content, σ jit is the variance of node location in a database; In formula (1), Section 1 is space constraint, ensures the similarity of candidate frame and target frame; Two item constraints are below to ensure temporal continuity, avoid shake.
4. as claimed in claim 1 based on the Realistic animation generation method of body sense camera Kinect, it is characterized in that, synthesis target frame sequence is the comparatively popular sparse representation method based on occurring recently, removes noise step as follows:
5-1) 3 d surface model of every frame is all obtained by same three-dimensional model deformation in assumption database and in target sequence, and they have identical summit and Topology connection; Dividing processing is carried out to three-dimensional model, is always divided into 16 parts and by mapping, the character image of target frame and retrieval frame is also divided into 16 parts; Calculate target frame and the respective pixel of retrieval frame, be guide with corresponding point, utilize Moving Least Squares method will retrieve frame distortion;
5-2) step 5-1) early results that obtains may comprise some missing information, because retrieval frame can not comprise all information of target frame, the method adopting the matrix based on rarefaction representation to fill up carries out interpolation to these regions, finally obtain every two field picture, synthesize new video sequence; For the method that the matrix based on rarefaction representation is filled up, first will ask for priority in the edge on current imperfect image, the sort criteria of priority considers texture and the depth information of block; Then, fill up according to the priority orders of each piece, for some blocks to be filled up, find the K similar to it similar piece; If the pixel P repaired iresiding w × w block B 0, and K similar piece is designated B respectively 1, B 2, B 3... B k, obtain a matrix D with these matrixes as each row.The method filled up based on the matrix of rarefaction representation is expressed as following optimization problem:
Wherein, A is original matrix to be asked, and E is nonzero element number, and D is observing matrix, and W is the weight matrix considering similar piece of similarity, and λ is the weight of noise size; The order of rank (A) representing matrix A, || || 00 norm of representing matrix, " ° " represents that two matrix corresponding elements are multiplied, and Ω is the index set of known elements, P Ωfor projecting to the operator in Ω territory.Above optimization problem is NP hard (Non-deterministic Polynomial-time hard) problem, therefore uses the nuclear norm of matrix || || *rank of matrix is replaced to minimize, by matrix 1 norm || || 1replace 0 norm, above-mentioned problem is converted into:
After solving out and obtaining matrix A, take out first row, be restructured as the block of pixels size of w × w, be the current block that method that the matrix based on rarefaction representation fills up recovers, the block of pixels of other positions by that analogy.
CN201510341839.6A 2015-06-18 2015-06-18 Realistic animation generation method based on body-sensing camera Kinect Active CN104915978B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510341839.6A CN104915978B (en) 2015-06-18 2015-06-18 Realistic animation generation method based on body-sensing camera Kinect

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510341839.6A CN104915978B (en) 2015-06-18 2015-06-18 Realistic animation generation method based on body-sensing camera Kinect

Publications (2)

Publication Number Publication Date
CN104915978A true CN104915978A (en) 2015-09-16
CN104915978B CN104915978B (en) 2018-04-03

Family

ID=54085014

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510341839.6A Active CN104915978B (en) 2015-06-18 2015-06-18 Realistic animation generation method based on body-sensing camera Kinect

Country Status (1)

Country Link
CN (1) CN104915978B (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105338369A (en) * 2015-10-28 2016-02-17 北京七维视觉科技有限公司 Method and apparatus for synthetizing animations in videos in real time
CN105427386A (en) * 2015-12-24 2016-03-23 南京邮电大学 Garment deformation method based on input human body posture real-time generation
CN105787995A (en) * 2016-01-17 2016-07-20 郑州财经学院 Method for processing plane figure image
CN106683178A (en) * 2016-12-30 2017-05-17 天津大学 Method for recovering three-dimensional framework by low-rank matrix on basis of graph theory
CN106780673A (en) * 2017-02-13 2017-05-31 杨金强 A kind of animation method and system
CN106920263A (en) * 2017-03-10 2017-07-04 大连理工大学 Undistorted integration imaging 3 D displaying method based on Kinect
CN107204000A (en) * 2017-04-24 2017-09-26 天津大学 Human body segmentation's method based on Kinect depth cameras
CN107403463A (en) * 2016-05-18 2017-11-28 西门子保健有限责任公司 The human body with nonrigid portions represents in imaging systems
CN108022278A (en) * 2017-12-29 2018-05-11 清华大学 Character animation method for drafting and system based on motion tracking in video
CN108055479A (en) * 2017-12-28 2018-05-18 暨南大学 A kind of production method of animal behavior video
CN108648261A (en) * 2018-03-25 2018-10-12 成都明镜视觉科技有限公司 A kind of threedimensional model defines method
CN109325442A (en) * 2018-09-19 2019-02-12 福州大学 A kind of face identification method of image pixel missing
CN111930231A (en) * 2020-07-27 2020-11-13 歌尔光学科技有限公司 Interaction control method, terminal device and storage medium
CN111968169A (en) * 2020-08-19 2020-11-20 北京拙河科技有限公司 Dynamic human body three-dimensional reconstruction method, device, equipment and medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1920886A (en) * 2006-09-14 2007-02-28 浙江大学 Video flow based three-dimensional dynamic human face expression model construction method
US20090284529A1 (en) * 2008-05-13 2009-11-19 Edilson De Aguiar Systems, methods and devices for motion capture using video imaging
CN102222361A (en) * 2010-04-06 2011-10-19 清华大学 Method and system for capturing and reconstructing 3D model

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1920886A (en) * 2006-09-14 2007-02-28 浙江大学 Video flow based three-dimensional dynamic human face expression model construction method
US20090284529A1 (en) * 2008-05-13 2009-11-19 Edilson De Aguiar Systems, methods and devices for motion capture using video imaging
CN102222361A (en) * 2010-04-06 2011-10-19 清华大学 Method and system for capturing and reconstructing 3D model

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
孙鑫、杨敬钰等: ""基于超像素与自回归模型的深度恢复"", 《光电子技术》 *
李红波等: ""一种基于Kinect的角色骨骼动画方法"", 《重庆邮电大学学报》 *
陈姝等: ""基于Kinect深度相机的实时三维人体动画"", 《计算机工程与科学》 *
黄浩、李坤: ""基于Kinect的虚拟视点生成技术"", 《信息技术》 *

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105338369A (en) * 2015-10-28 2016-02-17 北京七维视觉科技有限公司 Method and apparatus for synthetizing animations in videos in real time
CN105427386A (en) * 2015-12-24 2016-03-23 南京邮电大学 Garment deformation method based on input human body posture real-time generation
CN105787995B (en) * 2016-01-17 2018-09-28 郑州财经学院 A kind of graphical image plane processing method
CN105787995A (en) * 2016-01-17 2016-07-20 郑州财经学院 Method for processing plane figure image
CN107403463A (en) * 2016-05-18 2017-11-28 西门子保健有限责任公司 The human body with nonrigid portions represents in imaging systems
CN106683178A (en) * 2016-12-30 2017-05-17 天津大学 Method for recovering three-dimensional framework by low-rank matrix on basis of graph theory
CN106683178B (en) * 2016-12-30 2020-04-28 天津大学 Graph theory-based low-rank matrix three-dimensional framework recovery method
CN106780673A (en) * 2017-02-13 2017-05-31 杨金强 A kind of animation method and system
CN106920263A (en) * 2017-03-10 2017-07-04 大连理工大学 Undistorted integration imaging 3 D displaying method based on Kinect
CN106920263B (en) * 2017-03-10 2019-07-16 大连理工大学 Undistorted integration imaging 3 D displaying method based on Kinect
CN107204000A (en) * 2017-04-24 2017-09-26 天津大学 Human body segmentation's method based on Kinect depth cameras
CN108055479A (en) * 2017-12-28 2018-05-18 暨南大学 A kind of production method of animal behavior video
CN108022278A (en) * 2017-12-29 2018-05-11 清华大学 Character animation method for drafting and system based on motion tracking in video
CN108022278B (en) * 2017-12-29 2020-12-22 清华大学 Character animation drawing method and system based on motion tracking in video
CN108648261A (en) * 2018-03-25 2018-10-12 成都明镜视觉科技有限公司 A kind of threedimensional model defines method
CN109325442A (en) * 2018-09-19 2019-02-12 福州大学 A kind of face identification method of image pixel missing
CN111930231A (en) * 2020-07-27 2020-11-13 歌尔光学科技有限公司 Interaction control method, terminal device and storage medium
CN111930231B (en) * 2020-07-27 2022-02-25 歌尔光学科技有限公司 Interaction control method, terminal device and storage medium
CN111968169A (en) * 2020-08-19 2020-11-20 北京拙河科技有限公司 Dynamic human body three-dimensional reconstruction method, device, equipment and medium
CN111968169B (en) * 2020-08-19 2024-01-19 北京拙河科技有限公司 Dynamic human body three-dimensional reconstruction method, device, equipment and medium

Also Published As

Publication number Publication date
CN104915978B (en) 2018-04-03

Similar Documents

Publication Publication Date Title
CN104915978B (en) Realistic animation generation method based on body-sensing camera Kinect
Rematas et al. Soccer on your tabletop
Achenbach et al. Fast generation of realistic virtual humans
CN109636831B (en) Method for estimating three-dimensional human body posture and hand information
Yang et al. S3: Neural shape, skeleton, and skinning fields for 3d human modeling
CN113012282B (en) Three-dimensional human body reconstruction method, device, equipment and storage medium
US11195318B2 (en) Rapid avatar capture and simulation using commodity depth sensors
CN101303772A (en) Method for modeling non-linear three-dimensional human face based on single sheet image
US11928778B2 (en) Method for human body model reconstruction and reconstruction system
JP7446566B2 (en) Volumetric capture and mesh tracking based machine learning
JP2013120556A (en) Object attribute estimation device and video plotting device
Cheung et al. Markerless human motion transfer
CN114863035A (en) Implicit representation-based three-dimensional human motion capturing and generating method
Yang et al. Image translation based synthetic data generation for industrial object detection and pose estimation
CN108230431B (en) Human body action animation generation method and system of two-dimensional virtual image
Zhu et al. Mvp-human dataset for 3d human avatar reconstruction from unconstrained frames
Gomes et al. Do as I do: Transferring human motion and appearance between monocular videos with spatial and temporal constraints
CN109903360A (en) 3 D human face animation control system and its control method
Remondino et al. Human motion reconstruction and animation from video sequences
Wang et al. Hierarchical facial expression animation by motion capture data
Hetang et al. Novel View Synthesis from a Single RGBD Image for Indoor Scenes
Remondino et al. Markerless motion capture from single or multi-camera video sequence
CN116612256B (en) NeRF-based real-time remote three-dimensional live-action model browsing method
Yang et al. Disentangled human action video generation via decoupled learning
CN115049764B (en) Training method, device, equipment and medium of SMPL parameter prediction model

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20200703

Address after: 411, block a, Zhizao street, Zhongguancun, No. 45, Chengfu Road, Haidian District, Beijing 100080

Patentee after: Beijing Youke Nuclear Power Technology Development Co.,Ltd.

Address before: 300072 Tianjin City, Nankai District Wei Jin Road No. 92

Patentee before: Tianjin University

TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20201010

Address after: 100084 automation teaching staff, No.1 Qinghua garden, Haidian District, Beijing

Patentee after: Liu Yebin

Address before: 411, block a, Zhizao street, Zhongguancun, No. 45, Chengfu Road, Haidian District, Beijing 100080

Patentee before: Beijing Youke Nuclear Power Technology Development Co.,Ltd.

TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20210301

Address after: 518005 1001, building C2, Nanshan wisdom garden, 1001 Xueyuan Avenue, Changyuan community, Taoyuan Street, Nanshan District, Shenzhen City, Guangdong Province

Patentee after: Shenzhen Lingyun Shixun Technology Co.,Ltd.

Address before: 100084 automation staff, No.1 Qinghua garden, Haidian District, Beijing

Patentee before: Liu Yebin