CN111915693A - Sketch-based face image generation method and system - Google Patents

Sketch-based face image generation method and system Download PDF

Info

Publication number
CN111915693A
CN111915693A CN202010439641.2A CN202010439641A CN111915693A CN 111915693 A CN111915693 A CN 111915693A CN 202010439641 A CN202010439641 A CN 202010439641A CN 111915693 A CN111915693 A CN 111915693A
Authority
CN
China
Prior art keywords
feature
sketch
image
module
vector
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010439641.2A
Other languages
Chinese (zh)
Other versions
CN111915693B (en
Inventor
高林
傅红波
苏万超
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Computing Technology of CAS
City University of Hong Kong CityU
Original Assignee
Institute of Computing Technology of CAS
City University of Hong Kong CityU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Computing Technology of CAS, City University of Hong Kong CityU filed Critical Institute of Computing Technology of CAS
Priority to CN202010439641.2A priority Critical patent/CN111915693B/en
Publication of CN111915693A publication Critical patent/CN111915693A/en
Application granted granted Critical
Publication of CN111915693B publication Critical patent/CN111915693B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Software Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Processing (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The invention provides a face image generation method and a face image generation system based on a sketch, which comprise the following steps: extracting a plurality of parts of the human face in the hand-drawn sketch through a feature extraction network to obtain a plurality of feature vectors, and reasonably expressing each feature vector in a manifold space as an optimization vector; and decoding and mapping the optimized vector into a feature tensor by using a feature mapping network, splicing all feature tensors to obtain a complete human face feature tensor, and synthesizing the human face feature tensor into a human face image with high reality by using an image synthesis network. The invention aims to solve the problems of high quality requirement of input images and poor reality of generated results in the prior art, and provides a high-quality face image generation method and system based on sketch interaction.

Description

Sketch-based face image generation method and system
Technical Field
The present invention relates to computer graphics and computer vision, and more particularly to a method for synthesizing and editing a sketch of a face image.
Background
With the development of deep learning, image editing and translating work in the visual field is more and more, and the work of transferring from hand-drawn sketch images to real faces gradually appears. However, most of the existing solutions are based on professional draft images or contour maps extracted from real images as input, and the existing solutions do not perform well on draft images drawn by common users, and at present, no interactive system for combining facial images appears. The technology for generating the real face image from the sketch image has wide application in real life, and can be used for various aspects such as crime investigation, character design and the like. In the above applications, the existing research often adopts an image-to-image migration method, however, the method restricts the details of depicting the image of a person as much as possible, and when the depicted details are not fine enough, it is often difficult to generate a real face image, thereby limiting the range of users.
In view of the simplicity and the ease of use of sketches, which are often used to depict objects or face images, the prior art can implement the operation of editing the face images by using the sketches, but it is difficult to generate high-quality face images when only the sketches are input. Classical network models can enable the translation of sketch images into real images, however when the input is a rough or incomplete hand-drawn sketch, it is often difficult to generate real results. The prior art can complement a sketch and generate a real image, has good performance on simple objects such as pineapples and strawberries, but has poor performance on the generation of images migrated to human faces.
Disclosure of Invention
In order to solve the generation and interactive work from the hand-drawn sketch image to the real face image, the invention designs an interactive system based on the sketch. And (3) dividing the face image by utilizing a deep neural network, data prior and an optimization idea of local-global processing, and integrating the sub-regions to generate a high-quality face image after optimization processing. The invention aims to solve the problems of high quality requirement of input images and poor reality of generated results in the prior art, and provides a high-quality face image generation method and system based on sketch interaction. And designing a use software for the method, guiding/helping a user to carry out drawing creation in a shadow map form, and further generating a corresponding real image.
Aiming at the defects of the prior art, the invention provides a face image generation method based on a sketch, which comprises the following steps:
step 1, extracting a plurality of parts of a human face in a hand-drawn sketch through a feature extraction network to obtain a plurality of feature vectors, and reasonably expressing each feature vector in a manifold space as an optimization vector;
and 2, decoding and mapping the optimized vector into a feature tensor by using a feature mapping network, splicing all the feature tensors to obtain a complete face feature tensor, and synthesizing the face feature tensor into a face feature image by using an image synthesis network.
The face image generation method based on the sketch, wherein the step 1 comprises the following steps:
step 11, setting segmentation areas of a left eye area, a right eye area, a nose area and a mouth area of a face image in the hand-drawn sketch by adopting overlapping windows, obtaining other areas by reversely selecting the segmentation areas of the left eye area, the right eye area, the nose area and the mouth area, and independently extracting characteristics of each area c
Figure BDA0002503651040000021
Step 12, training the automatic coding network by taking the characteristics of each region as training data to obtain a coder model EcThe automatic coding network comprises a multi-layer decoder and a multi-layer encoder, a full connection layer is arranged between the encoder and the decoder, and a residual block is added after each convolution or deconvolution operation of the encoder and the decoder to construct a hidden layer descriptor.
The face image generation method based on the sketch, wherein the step 1 comprises the following steps:
step 13, constructing a draft image data set S-S for trainingiTo the training set of the hidden layer, extracting the characteristics of each picture in the training set, and for a certain area c, utilizing the encoder model EcConstructing a set of hidden layer features for the training composition
Figure BDA0002503651040000022
The hidden layer feature set FcAll the characteristic points are distributed in a low-dimensional manifold space McInternal;
step 14, inputting the hand-drawn sketch
Figure BDA0002503651040000029
Then, the encoder model E is usedcExtracting the feature vector of the corresponding region c
Figure BDA0002503651040000023
Searching and interpolating the characteristic vector of the c-th part
Figure BDA0002503651040000024
Projection onto manifold space McAbove, is represented as
Figure BDA0002503651040000025
To Mc
The retrieval and interpolation method specifically comprises the following steps:
step 141, feature vector for region c
Figure BDA0002503651040000026
Searching the training set to the feature set F by calculating Euclidean distance between feature vectorscThe most similar K samples
Figure BDA0002503651040000027
Represents a set of K most recent samples to represent
Figure BDA0002503651040000028
In manifold space McAdjacent features of (a);
step 142, solving the interpolation weight by the minimization problem as follows:
Figure BDA0002503651040000031
wherein
Figure BDA0002503651040000032
Is a sample
Figure BDA0002503651040000033
The weight of each component can be solved by solving the constraint least square of a single component;
step 143, weight of given solution
Figure BDA0002503651040000034
Figure BDA0002503651040000035
In manifold space McThe projected points on may be represented as:
Figure BDA0002503651040000036
wherein
Figure BDA0002503651040000037
Is that
Figure BDA0002503651040000038
For image synthesis.
The face image generation method based on the sketch, wherein the step 2 comprises the following steps:
step 21, mapping the optimized vector in the hidden layer space to multi-channel features to generate a three-dimensional feature tensor, and mapping the three-dimensional feature tensor of each region according to a freehand sketch
Figure BDA00025036510400000310
And splicing the exact positions of the middle human face components to obtain a complete human face characteristic tensor.
The face image generation method based on the sketch further comprises the following steps:
and 3, when a current drawing sketch is given, weighting and superposing the sketch image data set and K most similar sketch images of the drawing sketch together, displaying the sketch images on a drawing board in a form of a shadow image with the transparency of 30%, and updating the shadow image in real time when a user draws.
The invention also provides a face image generation system based on the sketch, which comprises the following steps:
the method comprises the following steps that a module 1 extracts a plurality of parts of a human face in a hand-drawn sketch through a feature extraction network to obtain a plurality of feature vectors, and reasonable expression of each feature vector in a manifold space is used as an optimization vector;
and the module 2 decodes and maps the optimized vector into a feature tensor by using the feature mapping network, combines all the feature tensors to obtain a complete face feature tensor, and combines the face feature tensor into a face feature image by using the image combining network.
The face image generation system based on sketch, wherein the module 1 includes:
the module 11 sets segmentation areas of a left eye area, a right eye area, a nose area and a mouth area of a face image in the hand-drawn sketch by adopting overlapping windows, obtains other areas by reversely selecting the segmentation areas of the left eye area, the right eye area, the nose area and the mouth area, and independently extracts features of each area c
Figure BDA0002503651040000039
The module 12 trains the automatic coding network by taking the characteristics of each region as training data to obtain a coder model EcThe automatic coding network comprises a multi-layer decoder and a multi-layer encoder, and a full connection layer is arranged between the encoder and the decoder, and each convolution or reverse convolution of the encoder and the decoderThe convolution operation is followed by the addition of a block of residuals to build the hidden layer descriptor.
The face image generation system based on sketch, wherein the module 1 includes:
module 13, construct a draft image dataset S ═ S for trainingiTo the training set of the hidden layer, extracting the characteristics of each picture in the training set, and for a certain area c, utilizing the encoder model EcConstructing a set of hidden layer features for the training composition
Figure BDA0002503651040000041
The hidden layer feature set FcAll the characteristic points are distributed in a low-dimensional manifold space McInternal;
module 14, input hand-drawing sketch
Figure BDA00025036510400000416
Then, the encoder model E is usedcExtracting the feature vector of the corresponding region c
Figure BDA0002503651040000042
Searching and interpolating the characteristic vector of the c-th part
Figure BDA0002503651040000043
Projection onto manifold space McAbove, is represented as
Figure BDA0002503651040000044
To Mc
The system for retrieving and interpolating specifically comprises:
module 141, feature vector for region c
Figure BDA0002503651040000045
Searching the training set to the feature set F by calculating Euclidean distance between feature vectorscThe most similar K samples
Figure BDA0002503651040000046
Represents a set of K most recent samples to represent
Figure BDA0002503651040000047
In manifold space McAdjacent features of (a);
module 142 solves the interpolation weights by the minimization problem by:
Figure BDA0002503651040000048
wherein
Figure BDA0002503651040000049
Is a sample
Figure BDA00025036510400000410
The weight of each component can be solved by solving the constraint least square of a single component;
module 143, weight given solution
Figure BDA00025036510400000411
Figure BDA00025036510400000412
In manifold space McThe projected points on may be represented as:
Figure BDA00025036510400000413
wherein
Figure BDA00025036510400000414
Is that
Figure BDA00025036510400000415
For image synthesis.
The face image generation system based on sketch, wherein the module 2 includes:
module 21, converting the hidden layer spaceThe optimized vector is mapped into the multi-channel features to generate a three-dimensional feature tensor, and the three-dimensional feature tensor of each region is mapped according to the hand-drawn sketch
Figure BDA00025036510400000417
And splicing the exact positions of the middle human face components to obtain a complete human face characteristic tensor.
The face image generation system based on sketch still includes:
and a module 3, when a current drawing sketch is given, displaying the sketch image data set and K most similar sketch images of the drawing sketch on a drawing board in a shadow map form with 30% transparency after weighted superposition, and updating the shadow map in real time when a user draws.
According to the scheme, the invention has the advantages that:
the system designed by the invention can realize the input of any hand-drawn sketch and automatically realize the shadow map guide and high-reality human face image generation operation.
The system flow chart is shown in fig. 1, and the system relates to a plurality of technologies such as image feature extraction, manifold projection of feature space, feature interpolation of feature space, high-quality image generation and interactive interface design.
As shown in fig. 2, an image generation effect after feature interpolation is performed in the feature space by looking up the sketch of the adjacent component in the feature space according to the present invention is shown. The first and fifth columns of the graph show two adjacent sketch images in a feature space, and the middle three columns of the graph are feature vectors which are uniformly interpolated and input into a decoder to decode the generated sketch images. As can be seen from the illustration results, the feature interpolation in the feature space can obtain a smooth image transition effect, thereby proving the continuity of the feature space distribution.
As shown in fig. 3, the generated effect of the present invention after adjusting the mixing weight of other regions is demonstrated. The upper left is a sketch image drawn by the user, and the rest five images are generated effects obtained by setting different mixing weight parameters in other areas. As can be seen from the figure, the generated image has higher correspondence to the input as the mixing weight coefficient is increased.
As shown in fig. 4, a comparison of the generated effects exhibited by the present invention after adjusting the mixing weights of multiple regions is shown. The first image shows the input sketch image, and the second image shows the effect generated directly by the input sketch. The third graph shows the effect of the generation after feature optimization, and the fourth graph shows the result of the generation after different pairs of mixing parameters are adjusted for different regions. It can be seen from the figure that the user can make a smooth adjustment between the optimized result and the user-generated result by adjusting the blending parameters.
As shown in FIG. 5, the effect of adding lines and editing details when the user draws is demonstrated. The image on the left part shows that the line constraint is increased along with the increase of the strokes of the user, the generated result is more similar, and the overall shadow of the generated image is changed. The right image shows the effect of the user on the detail adjustment, so that the method is free in detail control, and the shape expression of other areas cannot be influenced by the change of local detail lines.
As shown in fig. 6, the interactive interface design of the present invention is shown, in which there are two image display areas, a drawing board "SketchingCanvas" and a presentation board "SythesizedFace", which respectively display the sketch content drawn by the user and the generated image effect. Meanwhile, an adjusting bar is arranged on the upper right of the interface, and the user can adjust the weight of the corresponding area to realize the strong and weak influence of the user hand-drawn sketch on the generated result. Meanwhile, different controls are arranged above the interface, and a user can select a painting brush and an eraser and adjust the corresponding size. Whether to turn on the real-time generation function and the Shadow guidance function may also be decided by selecting "Realtime" and "Shadow".
As shown in fig. 7, it shows different image displays on the drawing board and the display board during the actual use process of the user. Wherein, on the drawing board, there is shadow guide, and the shadow map can be updated in real time according to the painting brush of the user. Meanwhile, the generated result can be displayed on the display board in real time, and the user can see the current generation effect.
As shown in fig. 8, the result generated when the present invention is actually used is shown, the first three lines are sketch images drawn by the user, and the second four lines are the corresponding face image generation results.
As shown in fig. 9, the effect generated when the interpolation is applied to the face image according to the present invention is shown. The left and right images are used, and the image in the middle is a face image obtained by interpolation, so that the smooth deformation sequence can be obtained when the face image interpolation is processed.
As shown in fig. 10, the effect of the present invention generated on face blending is shown, different parts of different faces are pieced together, sketch images of different parts are input, and a new complete face image is generated.
Drawings
FIG. 1 is a schematic flow chart of the system of the present invention;
FIG. 2 is a diagram of interpolation generation effects of different parts on a hidden layer;
FIG. 3 is a diagram showing the effect of setting different tuning mixing parameters in other areas after feature optimization;
FIG. 4 is a graph showing the effect of adjusting mixing parameters for different regions after feature optimization;
FIG. 5 is an add detail generate effect (left) and a partial detail edit generate effect (right) diagram;
FIG. 6 is an interactive interface effect display diagram;
FIG. 7 is a drawing board display diagram and a generated effect display diagram on the operation interface;
FIG. 8 is a production result display diagram;
FIG. 9 is a diagram of image interpolation generation effect;
fig. 10 is a diagram showing the mosaic composition effect of the face image.
Detailed Description
In the prior art, due to the fact that the quality of images input by users is not considered to be uneven, an untrained ordinary user is difficult to efficiently draw a face sketch image, and the input sketch information is not sufficiently expressed, so that the generated face image is poor in reality sense or cannot be generated; meanwhile, the common user often has difficulty in mastering the position proportion of facial features during drawing, and certain difficulty exists in drawing creation.
The inventor analyzes the structure of the facial features, and takes the distribution similarity of the facial features into consideration, so that the facial image is divided into five parts (left eye, right eye, nose, mouth and other regions), and each region is processed independently to better depict local details. The method aims to edit each region independently and prevent local editing from influencing the global effect. The local details can be better expressed by a separate operation after the face region is blocked. The effect of the partial editing can be referred to fig. 5.
When a user draws a sketch, the method can independently extract the features of each part, optimize the features by adopting a local linear embedding method (namely LLE (locally linear embedding), and generate a real face image by utilizing the optimized features. Meanwhile, in order to solve the problem that a common user cannot draw a reasonable face sketch, the invention designs an interface, when the user draws, the user can predict and recommend the sketch according to the drawing of the user, and the sketch is placed and drawn under a drawing board in a shadow mode so as to guide the user to draw. In consideration of different professional levels of users, the control strip is arranged, and the users can control the strength of the drawing sketch to the generation effect by adjusting the weight of each part.
The invention provides a high-quality face image generation method and system for sketch interaction, which comprises the following key points:
key point 1, face image feature extraction module
By adopting a method of processing the face image in a subarea manner, the face image is divided into 5 parts (a left eye, a right eye, a nose, a mouth and other parts) in view of the clear structure of the face, and the characteristics of each area image are independently extracted; to better control the details of each portion, the local features of each face portion are learned. And (3) independently training each region data by adopting an automatic encoding network (namely AE), and acquiring a corresponding feature vector of the sketch image in the hidden layer through an encoder.
Key point 2, feature optimization Module
In consideration of the diversity of sketch expression, in order to optimize the extracted features to be reasonable in manifold space, the feature projection module is utilized to project each part to the corresponding manifold distribution space, so that the optimized feature vectors conform to the reasonable expression of data, and the optimization of hidden layer features extracted from any sketch image in the reasonable expression space is realized.
Key point 3, feature mapping Module
Since each sub-region is independent of the feature extraction operation, after the optimized feature vector is obtained, the hidden layer feature needs to be decoded separately for each sub-region to obtain the respective feature tensor. And a certain replacing operation is adopted to realize the merging operation from the sub-region feature tensor to the complete human face feature map.
Key point 4, image composition Module
After the fused feature tensor is obtained, the image synthesis module converts the feature tensor into a real face image. The module adopts a conditional generation confrontation network structure, takes a face feature image as input, and realizes the generation of a high-quality face image through confrontation learning and image constraint.
Key point 5, design and display of Interactive interface
The interactive interface design supports users with different drawing levels to create, and can generate drawing results in real time. The shadow map is used for guiding the user in the drawing process, so that more layout and detail guidance is provided for the user. The user can select the corresponding function through the control in the toolbar, and the generation effect is adjusted by adjusting the strength of the control area constraint through the adjusting bar.
In order to make the aforementioned features and effects of the present invention more comprehensible, embodiments accompanied with figures are described in detail below.
As shown in fig. 1, the method of the present invention comprises:
s1: inputting user hand-drawn sketches
Figure BDA0002503651040000081
S2: drawing a sketch of a user's hand using a feature extraction network
Figure BDA0002503651040000082
Extracting to obtain c pairs of feature vectors of each part
Figure BDA0002503651040000083
S3: aligning feature vectors in a feature manifold space
Figure BDA0002503651040000084
Carrying out projection optimization to obtain a characteristic vector
Figure BDA0002503651040000085
In manifold space McReasonable expression of
Figure BDA0002503651040000086
Reasonable means that an expression closest to the real data distribution is found in the manifold space of the real data distribution;
s4: using the feature mapping network to optimize the feature vector
Figure BDA0002503651040000087
Decoding and mapping the feature tensor into a feature tensor, and splicing the feature tensor of each part to obtain a complete face feature tensor;
s5: and inputting a human face feature tensor by using an image synthesis network to generate a human face image with high reality as a human face image generation result of the sketch.
S6: interactive interface design and implementation
Wherein, the method in S2 includes:
s21: after the method and the device acquire the user hand-drawn sketch, the characteristics of the image need to be extracted. Processing the whole image to hardly express the diversity of face irrelevant pairs, and adopting a method of processing the face image in regions for this purpose; in view of the structure of the human faceClearly, the invention divides the face image into 5 parts (left eye, right eye, nose, mouth and other parts), and extracts the characteristic of each area image separately, which is marked as ScAnd c is {1,2,3,4,5 }; in order to better process details among the parts, the invention adopts an overlapping window to set the segmentation areas of the first four parts (the selection of the area size is determined according to the distribution range of the corresponding parts of the human beings in the database), and for the selection of other parts, we select the face image with the first four parts removed.
S22: to better control the details of each part, learning the local features of each face part requires designing a feature extraction network and training for real-time operation optimization. The method comprises two parts of feature network and training data processing.
S221: the invention adopts automatic coding network (AE) to train each region data independently, and after finishing AE network training from draft image to draft image, we can train through encoder EcTo obtain the corresponding feature vector of the sketch image in the hidden layer, and then the corresponding feature vector is processed by a decoder DcTo realize the generation of the feature vector to the sketch image; for the design of the automatic coding network, the invention adopts the form of a five-layer decoder and a five-layer encoder, and adopts the form of a full connection layer between the encoder and the decoder to ensure that the hidden layer characteristics generated by each part are 512 dimensions. For the dimension of the hidden layer features, the hidden layer features (128,256,512) with different dimensions are tested, and experiments show that the expression effect of 512 dimensions is better for the detail reconstruction and expression of the sketch, and the reconstruction result is fuzzy due to the low-dimensional features. Also after trial and error we add a residual block to build the hidden layer descriptor after each convolution/deconvolution operation of the encoding/decoding layer. The residual block is added in the network layer, so that the information loss of the network can be reduced, and the expression effect of the network model is improved.
S222: when training the feature extraction network, firstly, training data needs to be constructed. Since there is no real face image dataset containing a sketch image in the future, the present invention performs boundary contour extraction on color images to simulate sketch images. For the construction of the sketch image data set, after line extraction is carried out by adopting Photoshop, line reduction is carried out by adopting a sketch simplification method. When the network parameters are trained, the training is carried out in an automatic supervision mode, the Mean Square Error (MSE) loss between the input sketch image and the reconstructed image is caused, and an Adam optimizer is used for carrying out optimization solution on the network parameters.
Wherein the method in S3 comprises:
s31: considering that the sketch image input by the user may be different from the real face image, we need to map it into the real distribution space (manifold space). For an input sketch image s, a feature extraction module is used for acquiring a feature vector corresponding to each part c
Figure BDA0002503651040000091
Given the randomness and diversity of sketch representations, the features extracted by an image may not fit into the feature distribution of a dataset. In order to optimize the extracted features to be reasonable in a manifold space and achieve the purpose that a reasonable feature expression can be obtained even if the drawing sketch image is incomplete, the invention designs a feature optimization module which can extract hidden layer features from any sketch image and optimize the hidden layer features in a reasonable expression space.
S32: first, a draft image dataset S-S needs to be constructed for trainingiTo the hidden layer feature set. Extracting the characteristics of each picture in the training set, and for a certain part c, using an encoder model E trained in a characteristic extraction modulecTo construct a set of hidden layer features composed of a training set
Figure BDA0002503651040000092
The results shown in fig. 2 indicate that sketch images similar in feature space are similar in feature space pair distribution, so we consider all feature points F to be similarcDistributed in a low-dimensional manifold space, defined as Mc
S33: when inputting the sketch
Figure BDA0002503651040000101
Then, using the pre-trained encoder EcExtracting the feature vector of the corresponding part c, defined as
Figure BDA0002503651040000102
In order to increase the reality of expressing a human face, under the assumption of local linearity, the feature vector of the c part is searched and interpolated by following the idea of LLE
Figure BDA0002503651040000103
Projection to popular space McAbove, is represented as
Figure BDA0002503651040000104
To Mc. The method comprises two parts of nearest point search and characteristic interpolation:
s331: feature vector for part c
Figure BDA0002503651040000105
The invention firstly finds the characteristic set F from the training set by calculating the Euclidean distance between the characteristic vectorscThe K most similar samples. Experiments and comparisons show that when K is 10, the resultant can ensure both surface authenticity and proper variation. Definition of
Figure BDA0002503651040000106
To represent a set of K most recent samples,
Figure BDA0002503651040000107
to represent
Figure BDA0002503651040000108
In manifold space McOf the adjacent feature.
S332: using adjacent features
Figure BDA0002503651040000109
To the feature vector
Figure BDA00025036510400001010
A reconstruction representation is performed and a weight parameter solution is performed by minimizing the reconstruction error. This is equivalent to solving the interpolation weights by the following minimization problem:
Figure BDA00025036510400001011
wherein
Figure BDA00025036510400001012
Is a sample
Figure BDA00025036510400001013
Is determined. The weights of the components can be found by solving a constrained least squares problem for the individual components. Given the weight of the solution
Figure BDA00025036510400001014
Figure BDA00025036510400001019
In manifold space McThe projected points on may be represented as:
Figure BDA00025036510400001015
Figure BDA00025036510400001016
is that
Figure BDA00025036510400001017
The optimized feature vector can be transmitted to a feature mapping module and an image synthesis module for image synthesis.
Wherein, the method in S4 includes:
s41: given an input sketch image
Figure BDA00025036510400001018
Then, the feature extraction module can be used for extracting features and the feature projection module is used for projecting all parts to the corresponding spatial distribution manifold, so that the optimized feature vector is in a reasonable expression space. After the feature vectors of each part are acquired, the features are required to be used for synthesizing a complete face image, and if the features are directly decoded into sketch images for splicing, images appear at the joint of the regions, so that the generation effect is poor. Given that draft images have only one pass, the inconsistencies of adjacent components in the overlap region are difficult to resolve automatically by the sketch generation network. This forces us to map the eigenvectors in the hidden layer space into the multi-channel features (i.e., generate the three-dimensional feature tensor). It was found in experiments that this greatly improved the information flow, and effectively fused the feature map instead of simple stitching on the draft image, helping to resolve inconsistencies between parts. Because the features of different parts have different semantic meanings, a decoder model is designed for each part, the problem that the spatial distribution of high-dimensional features extracted from different parts is different is solved, and the conversion from a 512-dimensional feature vector to a three-dimensional feature tensor is realized. Each decoder consists of a full link layer and five decoding layers. Each feature tensor obtained by decoding has 32 channels and the size of the feature tensor is the same as the size of the area of the corresponding part in the sketch image.
S42: after the tensor representation of each part is obtained, the resultant feature tensors of the left eye, the right eye, the nose and the mouth need to be put back into the feature tensors of other parts according to the exact position of the face component in the input face sketch image, so as to retain the original spatial relationship between the face components. At the time of merging, we use a fixed merging order (i.e., left/right eye, nose, mouth, other parts) to merge the feature maps.
Wherein, the method in S5 includes:
after the feature tensor fused in S4 is acquired, the image synthesis module converts the feature tensor into a real face image. The module adopts a structure of a condition generation network, takes the feature tensor as condition input, and uses a discriminator model for countermeasure training during training. Our generation network comprises an encoder unit, a residual block and a decoder unit. In this module, we use a multi-scale discriminator network to discriminate between the different scale pairs of the generated images. In training, an input image and a generated image are down-sampled to different image sizes, and the images are discriminated by a discriminator corresponding to the image sizes. We use this to set up to implicitly learn high-level associations between components. During network training, parameter solution optimization is performed by adopting an Adam optimizer, and the feature extraction module and the image synthesis module are trained together through a network, so that generation from feature vectors to real pictures is realized. For the feature extraction module and the image synthesis module, in addition to generating network errors, we also add L1 errors between the real image and the reconstructed image to further guide the generator to ensure the pixel-level quality of the generated image. We use perceptual errors in the training of the discriminator to compare the feature differences between the real image and the generated image.
Wherein, the method in S6 includes:
s61: considering the user's drawing level to variance, it may be difficult for an ordinary user to accurately locate the user's five sense organs. In order to assist a user in drawing, particularly for a user with a low drawing level, a shadow-guided drawing interface is provided, and a shadow map can be updated in real time in the drawing process of the user, so that more layout and detail guidance can be provided for the user. When given the current drawing image
Figure BDA0002503651040000111
The K most similar sketch images in the training set S found in S33 are used. After the K most similar pictures are found, the sketch images are weighted and superposed together, and the weight of each picture is the weight in S3
Figure BDA0002503651040000121
The unitized values were displayed on a panel with 30% transparency. When the drawing is performed by the user,the shadow map is updated in real time. The resulting picture will be displayed to the right of the window.
S62: in consideration of the fact that a user with high drawing level can draw a face image well by himself, and the user can select whether to display a shadow image or not by controlling a control of the shadow above. When the user finishes drawing or wants to view the generated result in real time, the function to be changed or generated in real time can be selected by clicking the 'Convert' control or the 'Realtime' check box.
S62: considering that the difference of the user level may cause the expression difference of the details, the invention designs an adjusting strip to adjust the mixing parameter, and realizes the effect of the user controlling the influence of each part of drawing details on the generation result by adjusting the interpolation weight between the original image characteristics and the optimization characteristics, thereby increasing the controllability of the drawing details. Definition wbcRepresenting the blending parameters of part c, the blended feature vector may be calculated as:
Figure BDA0002503651040000122
will be provided with
Figure BDA0002503651040000123
A new composite image is obtained by inputting to S4 and S5.
The following are system examples corresponding to the above method examples, and this embodiment can be implemented in cooperation with the above embodiments. The related technical details mentioned in the above embodiments are still valid in this embodiment, and are not described herein again in order to reduce repetition. Accordingly, the related-art details mentioned in the present embodiment can also be applied to the above-described embodiments.
The invention also provides a face image generation system based on the sketch, which comprises the following steps:
the method comprises the following steps that a module 1 extracts a plurality of parts of a human face in a hand-drawn sketch through a feature extraction network to obtain a plurality of feature vectors, and reasonable expression of each feature vector in a manifold space is used as an optimization vector;
and the module 2 decodes and maps the optimized vector into a feature tensor by using the feature mapping network, combines all the feature tensors to obtain a complete face feature tensor, and combines the face feature tensor into a face feature image by using the image combining network.
The face image generation system based on sketch, wherein the module 1 includes:
the module 11 sets segmentation areas of a left eye area, a right eye area, a nose area and a mouth area of a face image in the hand-drawn sketch by adopting overlapping windows, obtains other areas by reversely selecting the segmentation areas of the left eye area, the right eye area, the nose area and the mouth area, and independently extracts features of each area c
Figure BDA0002503651040000124
The module 12 trains the automatic coding network by taking the characteristics of each region as training data to obtain a coder model EcThe automatic coding network comprises a multi-layer decoder and a multi-layer encoder, a full connection layer is arranged between the encoder and the decoder, and a residual block is added after each convolution or deconvolution operation of the encoder and the decoder to construct a hidden layer descriptor.
The face image generation system based on sketch, wherein the module 1 includes:
module 13, construct a draft image dataset S ═ S for trainingiTo the training set of the hidden layer, extracting the characteristics of each picture in the training set, and for a certain area c, utilizing the encoder model EcConstructing a set of hidden layer features for the training composition
Figure BDA0002503651040000131
The hidden layer feature set FcAll the characteristic points are distributed in a low-dimensional manifold space McInternal;
module 14, input hand-drawing sketch
Figure BDA00025036510400001316
Then, the encoder die is usedForm EcExtracting the feature vector of the corresponding region c
Figure BDA0002503651040000132
Searching and interpolating the characteristic vector of the c-th part
Figure BDA0002503651040000133
Projection onto manifold space McAbove, is represented as
Figure BDA0002503651040000134
To Mc
The system for retrieving and interpolating specifically comprises:
module 141, feature vector for region c
Figure BDA0002503651040000135
Searching the training set to the feature set F by calculating Euclidean distance between feature vectorscThe most similar K samples
Figure BDA0002503651040000136
Represents a set of K most recent samples to represent
Figure BDA0002503651040000137
In manifold space McAdjacent features of (a);
module 142 solves the interpolation weights by the minimization problem by:
Figure BDA0002503651040000138
wherein
Figure BDA0002503651040000139
Is a sample
Figure BDA00025036510400001310
The weight of each component can be solved by solving the constraint least square of a single component;
module 143, weight given solution
Figure BDA00025036510400001311
Figure BDA00025036510400001312
In manifold space McThe projected points on may be represented as:
Figure BDA00025036510400001313
wherein
Figure BDA00025036510400001314
Is that
Figure BDA00025036510400001315
For image synthesis.
The face image generation system based on sketch, wherein the module 2 includes:
the module 21 maps the optimized vector in the hidden layer space to the multi-channel features to generate a three-dimensional feature tensor, and the three-dimensional feature tensor of each region is mapped according to the freehand sketch
Figure BDA00025036510400001317
And splicing the exact positions of the middle human face components to obtain a complete human face characteristic tensor.
The face image generation system based on sketch still includes:
and a module 3, when a current drawing sketch is given, displaying the sketch image data set and K most similar sketch images of the drawing sketch on a drawing board in a shadow map form with 30% transparency after weighted superposition, and updating the shadow map in real time when a user draws.

Claims (10)

1. A face image generation method based on sketch is characterized by comprising the following steps:
step 1, extracting a plurality of parts of a human face in a hand-drawn sketch through a feature extraction network to obtain a plurality of feature vectors, and reasonably expressing each feature vector in a manifold space as an optimization vector;
and 2, decoding and mapping the optimized vector into a feature tensor by using a feature mapping network, splicing all the feature tensors to obtain a complete face feature tensor, and synthesizing the face feature tensor into a face feature image by using an image synthesis network.
2. The sketch-based human face image generation method as claimed in claim 1, wherein the step 1 comprises:
step 11, setting segmentation areas of a left eye area, a right eye area, a nose area and a mouth area of a face image in the hand-drawn sketch by adopting overlapping windows, obtaining other areas by reversely selecting the segmentation areas of the left eye area, the right eye area, the nose area and the mouth area, and independently extracting characteristics of each area c
Figure FDA0002503651030000018
Step 12, training the automatic coding network by taking the characteristics of each region as training data to obtain a coder model EcThe automatic coding network comprises a multi-layer decoder and a multi-layer encoder, a full connection layer is arranged between the encoder and the decoder, and a residual block is added after each convolution or deconvolution operation of the encoder and the decoder to construct a hidden layer descriptor.
3. The sketch-based human face image generation method as claimed in claim 2, wherein the step 1 comprises:
step 13, constructing a draft image data set s ═ s for trainingiTo the training set of the hidden layer, extracting the characteristics of each picture in the training set, and for a certain area c, utilizing the encoder model EcConstructing a set of hidden layer features for the training composition
Figure FDA0002503651030000011
The hidden layer feature set FcAll the characteristic points are distributed in a low-dimensional manifold space McInternal;
step 14, inputting the hand-drawn sketch
Figure FDA0002503651030000019
Then, the encoder model E is usedcExtracting the feature vector of the corresponding region c
Figure FDA0002503651030000012
Searching and interpolating the characteristic vector of the c-th part
Figure FDA0002503651030000013
Projected onto manifold space Mc, denoted
Figure FDA0002503651030000014
To Mc
The retrieval and interpolation method specifically comprises the following steps:
step 141, feature vector for region c
Figure FDA0002503651030000015
Searching the training set to the feature set F by calculating Euclidean distance between feature vectorscThe most similar K samples
Figure FDA0002503651030000016
Represents a set of K most recent samples to represent
Figure FDA0002503651030000017
In manifold space McAdjacent features of (a);
step 142, solving the interpolation weight by the minimization problem as follows:
Figure FDA0002503651030000021
wherein
Figure FDA0002503651030000022
Is a sample
Figure FDA0002503651030000023
The weight of each component can be solved by solving the constraint least square of a single component;
step 143, weight of given solution
Figure FDA0002503651030000024
Figure FDA0002503651030000025
In manifold space McThe projected points on may be represented as:
Figure FDA0002503651030000026
wherein
Figure FDA0002503651030000027
Is that
Figure FDA0002503651030000028
For image synthesis.
4. A sketch-based human face image generation method as claimed in claim 3, wherein the step 2 comprises:
step 21, mapping the optimized vector in the hidden layer space to multi-channel features to generate a three-dimensional feature tensor, and mapping the three-dimensional feature tensor of each region according to a freehand sketch
Figure FDA00025036510300000210
Exact position of the face componentAnd splicing to obtain a complete face feature tensor.
5. The sketch-based face image generating method as claimed in claim 4, further comprising:
and 3, when a current drawing sketch is given, weighting and superposing the sketch image data set and K most similar sketch images of the drawing sketch together, displaying the sketch images on a drawing board in a form of a shadow image with the transparency of 30%, and updating the shadow image in real time when a user draws.
6. A sketch-based face image generation system, comprising:
the method comprises the following steps that a module 1 extracts a plurality of parts of a human face in a hand-drawn sketch through a feature extraction network to obtain a plurality of feature vectors, and reasonable expression of each feature vector in a manifold space is used as an optimization vector;
and the module 2 decodes and maps the optimized vector into a feature tensor by using the feature mapping network, combines all the feature tensors to obtain a complete face feature tensor, and combines the face feature tensor into a face feature image by using the image combining network.
7. A sketch-based human face image generation system as claimed in claim 6, wherein the module 1 comprises:
the module 11 sets segmentation areas of a left eye area, a right eye area, a nose area and a mouth area of a face image in the hand-drawn sketch by adopting overlapping windows, obtains other areas by reversely selecting the segmentation areas of the left eye area, the right eye area, the nose area and the mouth area, and independently extracts features of each area c
Figure FDA0002503651030000029
The module 12 trains the automatic coding network by taking the characteristics of each region as training data to obtain a coder model EcThe automatic coding network comprises a multi-layer decoder and a multi-layer encoder, and at the encoder and at the decoderA full-concatenation layer is provided in the middle of the encoder, and a residual block is added after each convolution or deconvolution operation of the encoder and decoder to construct the hidden layer descriptor.
8. A sketch-based human face image generation system as claimed in claim 7 wherein the module 1 comprises:
module 13, construct a draft image dataset S ═ S for trainingiTo the training set of the hidden layer, extracting the characteristics of each picture in the training set, and for a certain area c, utilizing the encoder model EcConstructing a set of hidden layer features for the training composition
Figure FDA0002503651030000031
The hidden layer feature set FcAll the characteristic points are distributed in a low-dimensional manifold space McInternal;
module 14, input hand-drawing sketch
Figure FDA0002503651030000032
Then, the encoder model E is usedcExtracting the feature vector of the corresponding region c
Figure FDA0002503651030000033
Searching and interpolating the characteristic vector of the c-th part
Figure FDA0002503651030000034
Projected onto manifold space Mc, denoted
Figure FDA0002503651030000035
To Mc
The system for retrieving and interpolating specifically comprises:
module 141, feature vector for region c
Figure FDA0002503651030000036
By calculating feature vectorsEuclidean distance between them to retrieve the feature set F from the training setcThe most similar K samples
Figure FDA0002503651030000037
Represents a set of K most recent samples to represent
Figure FDA0002503651030000038
In manifold space McAdjacent features of (a);
module 142 solves the interpolation weights by the minimization problem by:
Figure FDA0002503651030000039
wherein
Figure FDA00025036510300000310
Is a sample
Figure FDA00025036510300000311
The weight of each component can be solved by solving the constraint least square of a single component;
module 143, weight given solution
Figure FDA00025036510300000312
Figure FDA00025036510300000313
In manifold space McThe projected points on may be represented as:
Figure FDA00025036510300000314
wherein
Figure FDA00025036510300000315
Is that
Figure FDA00025036510300000316
For image synthesis.
9. A sketch-based human face image generation system as claimed in claim 8 wherein the module 2 comprises:
the module 21 maps the optimized vector in the hidden layer space to the multi-channel features to generate a three-dimensional feature tensor, and the three-dimensional feature tensor of each region is mapped according to the freehand sketch
Figure FDA00025036510300000317
And splicing the exact positions of the middle human face components to obtain a complete human face characteristic tensor.
10. The sketch-based face image generation system of claim 9, further comprising:
and a module 3, when a current drawing sketch is given, displaying the sketch image data set and K most similar sketch images of the drawing sketch on a drawing board in a shadow map form with 30% transparency after weighted superposition, and updating the shadow map in real time when a user draws.
CN202010439641.2A 2020-05-22 2020-05-22 Sketch-based face image generation method and sketch-based face image generation system Active CN111915693B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010439641.2A CN111915693B (en) 2020-05-22 2020-05-22 Sketch-based face image generation method and sketch-based face image generation system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010439641.2A CN111915693B (en) 2020-05-22 2020-05-22 Sketch-based face image generation method and sketch-based face image generation system

Publications (2)

Publication Number Publication Date
CN111915693A true CN111915693A (en) 2020-11-10
CN111915693B CN111915693B (en) 2023-10-24

Family

ID=73237617

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010439641.2A Active CN111915693B (en) 2020-05-22 2020-05-22 Sketch-based face image generation method and sketch-based face image generation system

Country Status (1)

Country Link
CN (1) CN111915693B (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112541958A (en) * 2020-12-21 2021-03-23 清华大学 Parametric modeling method and device for three-dimensional face
CN112686822A (en) * 2020-12-30 2021-04-20 成都信息工程大学 Image completion method based on stack generation countermeasure network
CN112991484A (en) * 2021-04-28 2021-06-18 中国科学院计算技术研究所数字经济产业研究院 Intelligent face editing method and device, storage medium and equipment
CN113112572A (en) * 2021-04-13 2021-07-13 复旦大学 Hidden space search-based image editing method guided by hand-drawn sketch
CN113129447A (en) * 2021-04-12 2021-07-16 清华大学 Three-dimensional model generation method and device based on single hand-drawn sketch and electronic equipment
CN113222058A (en) * 2021-05-28 2021-08-06 新疆爱华盈通信息技术有限公司 Image classification method and device, electronic equipment and storage medium
CN113298097A (en) * 2021-07-27 2021-08-24 电子科技大学 Feature point extraction method and device based on convolutional neural network and storage medium
CN114266946A (en) * 2021-12-31 2022-04-01 智慧眼科技股份有限公司 Feature identification method and device under shielding condition, computer equipment and medium
CN114359034A (en) * 2021-12-24 2022-04-15 北京航空航天大学 Method and system for generating face picture based on hand drawing
WO2022126614A1 (en) * 2020-12-18 2022-06-23 中国科学院深圳先进技术研究院 Manifold optimization-based deep learning method for dynamic magnetic resonance imaging
CN115358917A (en) * 2022-07-14 2022-11-18 北京汉仪创新科技股份有限公司 Method, device, medium and system for transferring non-aligned faces in hand-drawing style
US12008821B2 (en) * 2021-05-07 2024-06-11 Google Llc Machine-learned models for unsupervised image transformation and retrieval

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100134487A1 (en) * 2008-12-02 2010-06-03 Shang-Hong Lai 3d face model construction method
CN102163330A (en) * 2011-04-02 2011-08-24 西安电子科技大学 Multi-view face synthesis method based on tensor resolution and Delaunay triangulation
CN107977629A (en) * 2017-12-04 2018-05-01 电子科技大学 A kind of facial image aging synthetic method of feature based separation confrontation network
CN108520210A (en) * 2018-03-26 2018-09-11 河南工程学院 Based on wavelet transformation and the face identification method being locally linear embedding into
CN108520503A (en) * 2018-04-13 2018-09-11 湘潭大学 A method of based on self-encoding encoder and generating confrontation network restoration face Incomplete image
CN109886869A (en) * 2018-10-15 2019-06-14 武汉工程大学 A kind of unreal structure method of face of the non-linear expansion based on contextual information
CN110175251A (en) * 2019-05-25 2019-08-27 西安电子科技大学 The zero sample Sketch Searching method based on semantic confrontation network

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100134487A1 (en) * 2008-12-02 2010-06-03 Shang-Hong Lai 3d face model construction method
CN102163330A (en) * 2011-04-02 2011-08-24 西安电子科技大学 Multi-view face synthesis method based on tensor resolution and Delaunay triangulation
CN107977629A (en) * 2017-12-04 2018-05-01 电子科技大学 A kind of facial image aging synthetic method of feature based separation confrontation network
CN108520210A (en) * 2018-03-26 2018-09-11 河南工程学院 Based on wavelet transformation and the face identification method being locally linear embedding into
CN108520503A (en) * 2018-04-13 2018-09-11 湘潭大学 A method of based on self-encoding encoder and generating confrontation network restoration face Incomplete image
CN109886869A (en) * 2018-10-15 2019-06-14 武汉工程大学 A kind of unreal structure method of face of the non-linear expansion based on contextual information
CN110175251A (en) * 2019-05-25 2019-08-27 西安电子科技大学 The zero sample Sketch Searching method based on semantic confrontation network

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
SHAOJUN BIAN ET AL.: ""Fully Automatic Facial Deformation Transfer"", 《SYMMETRY 2020》 *
YUWEI LI ET AL.: ""SweepCanvas Sketch-based 3D Prototyping on an RGB-D Image"", 《CONFERENCE:THE 30TH ANNUAL ACM SYMPOSIUM》 *
王智飞: ""低分辨率人脸识别算法研究"", 《中国博士学位论文全文数据库信息科技辑》 *
白万荣: ""人脸识别中基于流形学习的特征提取方法研究"", 《中国优秀硕士学位论文全文数据库信息科技辑》 *
高晨昊: ""基于深度神经网络特征表达的人脸素描合成"", 《中国优秀硕士学位论文全文数据库信息科技辑》 *

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022126614A1 (en) * 2020-12-18 2022-06-23 中国科学院深圳先进技术研究院 Manifold optimization-based deep learning method for dynamic magnetic resonance imaging
CN112541958B (en) * 2020-12-21 2022-09-09 清华大学 Parametric modeling method and device for three-dimensional face
CN112541958A (en) * 2020-12-21 2021-03-23 清华大学 Parametric modeling method and device for three-dimensional face
CN112686822B (en) * 2020-12-30 2021-09-07 成都信息工程大学 Image completion method based on stack generation countermeasure network
CN112686822A (en) * 2020-12-30 2021-04-20 成都信息工程大学 Image completion method based on stack generation countermeasure network
CN113129447A (en) * 2021-04-12 2021-07-16 清华大学 Three-dimensional model generation method and device based on single hand-drawn sketch and electronic equipment
CN113112572B (en) * 2021-04-13 2022-09-06 复旦大学 Hidden space search-based image editing method guided by hand-drawn sketch
CN113112572A (en) * 2021-04-13 2021-07-13 复旦大学 Hidden space search-based image editing method guided by hand-drawn sketch
CN112991484A (en) * 2021-04-28 2021-06-18 中国科学院计算技术研究所数字经济产业研究院 Intelligent face editing method and device, storage medium and equipment
US12008821B2 (en) * 2021-05-07 2024-06-11 Google Llc Machine-learned models for unsupervised image transformation and retrieval
CN113222058A (en) * 2021-05-28 2021-08-06 新疆爱华盈通信息技术有限公司 Image classification method and device, electronic equipment and storage medium
CN113222058B (en) * 2021-05-28 2024-05-10 芯算一体(深圳)科技有限公司 Image classification method, device, electronic equipment and storage medium
CN113298097A (en) * 2021-07-27 2021-08-24 电子科技大学 Feature point extraction method and device based on convolutional neural network and storage medium
CN113298097B (en) * 2021-07-27 2021-10-26 电子科技大学 Feature point extraction method and device based on convolutional neural network and storage medium
CN114359034A (en) * 2021-12-24 2022-04-15 北京航空航天大学 Method and system for generating face picture based on hand drawing
CN114359034B (en) * 2021-12-24 2023-08-08 北京航空航天大学 Face picture generation method and system based on hand drawing
CN114266946A (en) * 2021-12-31 2022-04-01 智慧眼科技股份有限公司 Feature identification method and device under shielding condition, computer equipment and medium
CN115358917A (en) * 2022-07-14 2022-11-18 北京汉仪创新科技股份有限公司 Method, device, medium and system for transferring non-aligned faces in hand-drawing style
CN115358917B (en) * 2022-07-14 2024-05-07 北京汉仪创新科技股份有限公司 Method, equipment, medium and system for migrating non-aligned faces of hand-painted styles

Also Published As

Publication number Publication date
CN111915693B (en) 2023-10-24

Similar Documents

Publication Publication Date Title
CN111915693B (en) Sketch-based face image generation method and sketch-based face image generation system
Frolov et al. Adversarial text-to-image synthesis: A review
Cheng et al. Sdfusion: Multimodal 3d shape completion, reconstruction, and generation
Qi et al. SGUIE-Net: Semantic attention guided underwater image enhancement with multi-scale perception
CN111291212B (en) Zero sample sketch image retrieval method and system based on graph convolution neural network
CN111275518A (en) Video virtual fitting method and device based on mixed optical flow
CN112887698B (en) High-quality face voice driving method based on nerve radiation field
CN113255457A (en) Animation character facial expression generation method and system based on facial expression recognition
CN113807265B (en) Diversified human face image synthesis method and system
CN114529940B (en) Human body image generation method based on gesture guidance
CN111967533B (en) Sketch image translation method based on scene recognition
Zhang et al. Dinet: Deformation inpainting network for realistic face visually dubbing on high resolution video
CN113470182A (en) Face geometric feature editing method and deep face remodeling editing method
CN115618452B (en) Clothing image intelligent generation system with designer style
CN113362422A (en) Shadow robust makeup transfer system and method based on decoupling representation
Yang et al. Controllable sketch-to-image translation for robust face synthesis
Sun et al. Human motion transfer with 3d constraints and detail enhancement
Huang et al. Multi-density sketch-to-image translation network
CN113538662B (en) Single-view three-dimensional object reconstruction method and device based on RGB data
Zeng et al. Expression-tailored talking face generation with adaptive cross-modal weighting
Liu et al. Transformer-based high-fidelity facial displacement completion for detailed 3d face reconstruction
Vuong et al. Language-driven scene synthesis using multi-conditional diffusion model
Wu et al. DeepPortraitDrawing: Generating human body images from freehand sketches
Li et al. MIGT: Multi-modal image inpainting guided with text
CN115578298A (en) Depth portrait video synthesis method based on content perception

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant