CN106023288A - Image-based dynamic substitute construction method - Google Patents
Image-based dynamic substitute construction method Download PDFInfo
- Publication number
- CN106023288A CN106023288A CN201610331428.3A CN201610331428A CN106023288A CN 106023288 A CN106023288 A CN 106023288A CN 201610331428 A CN201610331428 A CN 201610331428A CN 106023288 A CN106023288 A CN 106023288A
- Authority
- CN
- China
- Prior art keywords
- image
- face
- hair
- images
- expression
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000010276 construction Methods 0.000 title claims abstract description 17
- 210000004209 hair Anatomy 0.000 claims abstract description 114
- 230000014509 gene expression Effects 0.000 claims abstract description 57
- 238000000034 method Methods 0.000 claims abstract description 43
- 230000001815 facial effect Effects 0.000 claims abstract description 31
- 230000004927 fusion Effects 0.000 claims abstract description 23
- 238000007781 pre-processing Methods 0.000 claims abstract description 15
- 230000008569 process Effects 0.000 claims abstract description 15
- 230000009471 action Effects 0.000 claims abstract description 10
- 210000003128 head Anatomy 0.000 claims description 34
- 210000001508 eye Anatomy 0.000 claims description 22
- 230000008921 facial expression Effects 0.000 claims description 12
- 230000009466 transformation Effects 0.000 claims description 12
- 238000013507 mapping Methods 0.000 claims description 9
- 230000006870 function Effects 0.000 claims description 7
- 238000013519 translation Methods 0.000 claims description 7
- 238000012937 correction Methods 0.000 claims description 6
- 230000015572 biosynthetic process Effects 0.000 claims description 5
- 238000003786 synthesis reaction Methods 0.000 claims description 5
- 230000011218 segmentation Effects 0.000 claims description 3
- 238000013480 data collection Methods 0.000 abstract description 4
- 230000000875 corresponding effect Effects 0.000 description 28
- 230000033001 locomotion Effects 0.000 description 19
- 210000000515 tooth Anatomy 0.000 description 16
- 230000037303 wrinkles Effects 0.000 description 10
- 238000004364 calculation method Methods 0.000 description 9
- 230000000694 effects Effects 0.000 description 8
- 238000005457 optimization Methods 0.000 description 7
- 238000005516 engineering process Methods 0.000 description 4
- 230000003993 interaction Effects 0.000 description 4
- 238000005070 sampling Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 3
- 210000000887 face Anatomy 0.000 description 3
- 230000005484 gravity Effects 0.000 description 3
- 238000003709 image segmentation Methods 0.000 description 3
- 239000013598 vector Substances 0.000 description 3
- 230000002159 abnormal effect Effects 0.000 description 2
- 210000005252 bulbus oculi Anatomy 0.000 description 2
- 238000007499 fusion processing Methods 0.000 description 2
- 238000003384 imaging method Methods 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- 208000027697 autoimmune lymphoproliferative syndrome due to CTLA4 haploinsuffiency Diseases 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 239000007795 chemical reaction product Substances 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 210000001097 facial muscle Anatomy 0.000 description 1
- 239000011521 glass Substances 0.000 description 1
- 238000009499 grossing Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 210000000056 organ Anatomy 0.000 description 1
- 239000000047 product Substances 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 238000012549 training Methods 0.000 description 1
- 238000012800 visualization Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T13/00—Animation
- G06T13/20—3D [Three Dimensional] animation
- G06T13/40—3D [Three Dimensional] animation of characters, e.g. humans, animals or virtual beings
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Processing Or Creating Images (AREA)
Abstract
本发明公开了一种基于图像的动态替身构造方法,该方法首先进行数据采集和预处理:使用普通的网络摄像头,采集用户的一系列设定动作表情的人脸图像,并对这些图像进行分割、特征点标定等预处理工作;然后基于处理后的图像,生成用户的人脸融合模型和头发形变模型,继而得到用户基于图像的替身表达;在实时人脸动画驱动过程中,根据跟踪得到的人脸动作表情参数,驱动替身表达生成对应的人脸和头发几何;最后基于得到的人脸和头发几何,映射采集图像,并将映射后的图像根据图像置信度进行融合,生成真实的人脸动画图像。利用本发明生成得到的人脸动画结果,具有高真实感、表现力强、细节丰富、还原度高等特点。The invention discloses an image-based dynamic avatar construction method. The method firstly carries out data collection and preprocessing: using a common network camera to collect a series of user face images with set action expressions, and segment these images , feature point calibration and other preprocessing work; then based on the processed image, generate the user's face fusion model and hair deformation model, and then obtain the user's image-based avatar expression; in the real-time face animation driving process, according to the tracking obtained Face action and expression parameters, driving avatar expression to generate corresponding face and hair geometry; finally, based on the obtained face and hair geometry, map and collect images, and fuse the mapped images according to the image confidence to generate a real face Animated images. The facial animation result generated by the present invention has the characteristics of high sense of reality, strong expressive force, rich details, high degree of restoration, and the like.
Description
技术领域technical field
本发明涉及基于表演的人脸动画技术和人脸表情动作重定向技术领域,尤其涉及一种基于图像动态替身的人脸动画方法。The present invention relates to the field of performance-based face animation technology and face expression action redirection technology, and in particular to a face animation method based on image dynamic avatar.
背景技术Background technique
相比于基于人脸肌肉模型(如VENKATARAMANA,K.,LODHAA,S.,AND RAGHAVAN,R.2005.A kinematic-variational model for animating skin withwrinkles.Computer&Graphics 29,5(Oct),756–770.)和过程式参数化模型(如JIMENEZ,J.,ECHEVARRIA,J.I.,OAT,C.,AND GUTIERREZ,D.2011.GPU Pro 2.AK Peters Ltd.,ch.Practical and Realistic Facial Wrinkles Animation.),数据驱动的方法在创建动态替身中更为常见,这得利于该方法可以用非常低的计算代价得到真实的人脸动作。比如,多维线性人脸模型(VLASIC,D.,BRAND,M.,PFISTER,H.,AND POPOVI′C,J.2005.Facetransfer with multilinear models.ACM Trans.Graph.24,3(July),426–433.)使用一个统一的模型,分不同维度来表示人脸形状中的用户个体系数和表情系数。通过拟合输入的深度图或者视频序列,多维线性人脸模型常被用于创建用户特定的人脸融合模型。人脸融合模型中动态的几何变化可以被线性模型化,以用于实时的人脸跟踪与动画系统,如基于深度相机的实时人脸动画方法。这些线性模型因为计算高效被广泛用于实时人脸跟踪与动画方法,但无法展示人脸上的细节,如皱纹。Compared with facial muscle models (such as VENKATARAMANA, K., LODHAA, S., AND RAGHAVAN, R. 2005. A kinematic-variational model for animating skin with wrinkles. Computer & Graphics 29, 5 (Oct), 756–770.) and procedural parametric models (such as JIMENEZ, J., ECHEVARRIA, J.I., OAT, C., AND GUTIERREZ, D.2011. GPU Pro 2. AK Peters Ltd., ch. Practical and Realistic Facial Wrinkles Animation.), data The driving method is more common in the creation of dynamic avatars, which benefits from the fact that this method can obtain real facial movements at a very low computational cost. For example, the multidimensional linear face model (VLASIC, D., BRAND, M., PFISTER, H., AND POPOVI′C, J.2005.Facetransfer with multilinear models. ACM Trans.Graph.24,3 (July), 426 –433.) Use a unified model to represent user individual coefficients and expression coefficients in face shape in different dimensions. Multidimensional linear face models are often used to create user-specific face fusion models by fitting input depth maps or video sequences. The dynamic geometric changes in the face fusion model can be linearly modeled for real-time face tracking and animation systems, such as real-time face animation methods based on depth cameras. These linear models are widely used in real-time face tracking and animation methods because of their computational efficiency, but cannot reveal the details of human faces, such as wrinkles.
在一些高端的产品如电影制作中,一些特殊硬件设备(如LightStage系统)被用于制作高真实感的动态替身,其包含了丰富的人脸细节,如皮肤的褶皱。但是这些特殊硬件设备无法用于面向普通用户的应用中。In some high-end products such as film production, some special hardware devices (such as LightStage system) are used to produce high-realistic dynamic avatars, which contain rich facial details, such as skin folds. But these special hardware devices cannot be used in applications for ordinary users.
另有一些技术基于单帧图像创立动态替身。输入用户和目标替身的各一张人脸图像,如(SARAGIH,J.,LUCEY,S.,AND COHN,J.2011.Real-time avatar animation from asingle image.In AFGR,213–220.)的方法在预处理阶段学习输入用户和目标替身用户之间表情的映射函数。在实时运行阶段,输入视频中用户的人脸动作通过拟合一个可形变的人脸模型来作跟踪,之后基于在预处理阶段学习的映射函数将用户的动作转移到替身上,生成替身相应的动作形状。(CAO,C.,WENG,Y.,ZHOU,S.,TONG,Y.,AND ZHOU,K.2014.Facewarehouse:A 3d facial expression database for visualcomputing.IEEE Transactions on Visualization and Computer Graphics 20,3(Mar.),413–425.)介绍了一种技术,可以令任意用户通过表演驱动一张静态的人脸图像做出各种人脸动画效果。该技术首先使用多维线性人脸模型为静态图像中的人脸拟合一个人脸融合模型,并从图像抽取人脸模型的纹理。在实时运行过程中,跟踪得到的头部刚性运动参数和非刚性表情系数被转移到该人脸融合模型上,生成全新视角下的人脸动画。对于头发,该技术使用一种基于单视角的头发建模技术来创立了一个三维头发发丝模型,在实时运行过程中与头部一起运动和渲染,生成最终的结果。因为只有一张图像被使用,因此得到的替身动画效果表现力并不强,特别是不能产生表情皱纹的细节。此外,大幅度的头部旋转和夸张人脸表情也是这一系列方法无法解决的难题。There are also techniques for creating animatronics based on a single frame of image. Input a face image of the user and the target avatar, such as (SARAGIH, J., LUCEY, S., AND COHN, J.2011. Real-time avatar animation from asingle image. In AFGR, 213–220.) The method learns the expression mapping function between the input user and the target avatar user in the preprocessing stage. In the real-time operation stage, the user's face motion in the input video is tracked by fitting a deformable face model, and then based on the mapping function learned in the preprocessing stage, the user's motion is transferred to the avatar, and the avatar corresponding action shape. (CAO,C.,WENG,Y.,ZHOU,S.,TONG,Y.,AND ZHOU,K.2014.Facewarehouse: A 3d facial expression database for visualcomputing.IEEE Transactions on Visualization and Computer Graphics 20,3(Mar .), 413–425.) introduce a technique that allows any user to drive a static face image to make various facial animation effects by acting. The technology first uses a multi-dimensional linear face model to fit a face fusion model for the face in the static image, and extracts the texture of the face model from the image. During real-time operation, the tracked head rigid motion parameters and non-rigid expression coefficients are transferred to the face fusion model to generate a face animation from a new perspective. For hair, the technology uses a single-view-based hair modeling technique to create a 3D model of hair strands that are moved and rendered with the head in real-time to produce the final result. Because only one image is used, the obtained stand-in animation effect is not very expressive, especially the details of expression wrinkles cannot be produced. In addition, large head rotation and exaggerated facial expressions are also difficult problems that this series of methods cannot solve.
近期有一些基于用户多张图像的动态替身方法被提出,致力于帮助普通用户生成完整的,富有细节特征的人脸动画替身。(ICHIM,A.E.,BOUAZIZ,S.,AND PAULY,M.2015.Dynamic 3d avatar creation from hand-held video input.ACMTrans.Graph.34,4(July),45:1–45:14.)提出了一个基于手持设备视频的,以两个尺度表达的三维动态替身方法。通过用户输入的一系列自然表情图像,该方法首先使用运动结构算法得到人脸形状的点云,并针对该点云数据拟合该用户的自然表情模型。之后,通过对该自然表情模型和采集视频中人脸表情的拟合,生成一个中等尺度的用户特定的人脸融合模型。之后,精密尺度的人脸细节,如人脸皱纹等信息则通过明暗成形算法获得,并使用法向贴图和环境光遮蔽贴图来表达。(GARRIDO,P.,ZOLLHOFER,M.,CASAS,D.,VALGAERTS,L.,VARANASI,K.,PEREZ,P.,AND THEOBALT,C.2016.Reconstruction of personalized 3dface rigs from monocular video.ACM Trans.Graph.)提出了一种全自动的方法,可以从单目视频人脸视频数据(如传统电影)中自动构建三维人脸动态替身。该方法的替身基于三个尺度的几何层次表达,从稀疏几何尺度描述的人脸基本动作,到精密几何尺度描述的人脸细节,如皮肤皱纹。上述的两个方法分别使用多层的几何表达来表示替身,运算较为复杂。此外,更为重要的是,头发的运动效果在之前的工作中都无法被有效处理。Recently, some dynamic avatar methods based on multiple images of users have been proposed, which are dedicated to helping ordinary users generate complete and detailed facial animation avatars. (ICHIM,A.E.,BOUAZIZ,S.,AND PAULY,M.2015.Dynamic 3d avatar creation from hand-held video input.ACMTrans.Graph.34,4(July),45:1–45:14.) proposed A method for 3D dynamic stand-ins expressed at two scales based on handheld device video. Through a series of natural expression images input by the user, the method first uses the motion structure algorithm to obtain the point cloud of the face shape, and then fits the user's natural expression model to the point cloud data. Afterwards, a medium-scale user-specific face fusion model is generated by fitting the natural expression model and facial expressions in the captured video. Afterwards, fine-scale facial details, such as facial wrinkles, are obtained through shading algorithms, and expressed using normal maps and ambient light occlusion maps. (GARRIDO, P., ZOLLHOFER, M., CASAS, D., VALGAERTS, L., VARANASI, K., PEREZ, P., AND THEOBALT, C. 2016. Reconstruction of personalized 3dface rigs from monocular video. ACM Trans. Graph.) proposed a fully automatic method to automatically construct 3D face dynamic avatars from monocular video face video data such as traditional movies. The stand-in of this method is based on three scales of geometric hierarchical expression, from basic facial movements described by sparse geometric scales to facial details such as skin wrinkles described by precise geometric scales. The above two methods respectively use multi-layer geometric expressions to represent avatars, and the calculation is relatively complicated. In addition, and more importantly, the motion effect of hair cannot be effectively handled in previous work.
发明内容Contents of the invention
本发明的目的在于针对现有技术的不足,提出了一种新的基于图像的人脸动态替身构造方法,在人脸表情跟踪和动画系统的驱动下,该替身表达可以给出具有真实感、有表现力的人脸动画结果。本发明中提出的替身模型包含了用户完整的头部表达,如人脸、头发、牙齿、眼睛,甚至于头发上的头饰,相比于其它的替身表达本发明的表达更加完整,因此得到的动画效果更具有真实说服力。本发明中在预处理过程需要的数据都可以由普通用户使用普通设备,如网络摄像头、家庭电脑等得到,而运行过程中则可以由任意用户的表演对替身进行人脸动画驱动,非常适合各于种虚拟现实的应用,如网络游戏、视频聊天、远程教育等等。The purpose of the present invention is to address the deficiencies in the prior art, and propose a new image-based method for constructing a dynamic substitute for a human face. Driven by a facial expression tracking and animation system, the substitute expression can give realistic, Expressive facial animation results. The stand-in model proposed in the present invention includes the user's complete head expression, such as human face, hair, teeth, eyes, and even the headgear on the hair. Compared with other stand-in expressions, the expression of the present invention is more complete, so the obtained The animation effect is more realistic and convincing. In the present invention, the data required in the preprocessing process can be obtained by ordinary users using ordinary equipment, such as network cameras, home computers, etc., and in the running process, the avatar can be driven by the performance of any user, which is very suitable for various For a variety of virtual reality applications, such as online games, video chat, distance education and so on.
本发明的目的是通过以下技术方案来实现的:The purpose of the present invention is achieved through the following technical solutions:
一种基于图像的人脸动态替身构造方法,其主要包括以下几个步骤:An image-based method for constructing dynamic substitutes for human faces, which mainly includes the following steps:
1、数据采集和预处理:使用普通的网络摄像头,采集用户的一系列设定动作表情的人脸图像,并对这些图像进行分割、特征点标定等预处理工作。1. Data collection and preprocessing: Use an ordinary webcam to collect a series of facial images of users with set action expressions, and perform preprocessing work such as segmentation and feature point calibration on these images.
2、基于图像的替身构造:基于处理后的图像,生成用户的人脸融合模型和头发形变模型,继而得到用户基于图像的替身表达。2. Image-based avatar construction: Based on the processed image, the user's face fusion model and hair deformation model are generated, and then the user's image-based avatar expression is obtained.
3、实时人脸和头发几何生成:在实时人脸动画驱动过程中,根据跟踪得到的人脸动作表情参数,驱动替身表达生成对应的人脸和头发几何。3. Real-time face and hair geometry generation: In the real-time face animation driving process, according to the tracked facial movement and expression parameters, the avatar expression is driven to generate the corresponding face and hair geometry.
4、实时人脸动画合成:基于得到的人脸和头发几何,映射采集图像,并将映射后的图像根据图像置信度进行融合,生成真实的人脸动画图像。4. Real-time facial animation synthesis: Based on the obtained face and hair geometry, map and collect images, and fuse the mapped images according to image confidence to generate real facial animation images.
本发明的有益效果是,本发明替身表达完整,相比于之前的动态替身表达方法,本发明提出的基于图像的动态替身不仅包括人脸、眼睛、牙齿等器官,还包括难以建模的头发和头发上的头饰,因此该替身表达更加完整;此外,人脸表情变化的细节,如皮肤的折叠、皱纹,在本替身表达中并不需要重建对应的三维几何,而是完全隐式的包含在采集图像中。综上所述,相比于其他的人脸替身动画方法,本发明提出基于该替身的人脸动画更具有真实性和表现力,且因为设备要求低,易于使用、适用于任意用户等特点,因此可以应用于如网络游戏、视频聊天等各种应用中,具有广阔的应用前景。The beneficial effect of the present invention is that the avatar expression of the present invention is complete. Compared with the previous dynamic avatar expression method, the image-based dynamic avatar proposed by the present invention includes not only human faces, eyes, teeth and other organs, but also hair that is difficult to model. and the headgear on the hair, so the stand-in expression is more complete; in addition, the details of facial expression changes, such as skin folds and wrinkles, do not need to reconstruct the corresponding 3D geometry in this stand-in expression, but are completely implicitly included in the captured image. To sum up, compared with other face substitute animation methods, the present invention proposes that the face animation based on the substitute is more authentic and expressive, and because of low equipment requirements, easy to use, and suitable for any user, etc., Therefore, it can be used in various applications such as online games, video chats, etc., and has broad application prospects.
附图说明Description of drawings
图1是本发明数据采集和预处理的一个示例图,从左到右分别是:采集图像、图像分割层次,头发层透明通道和标记的特征点。Fig. 1 is an example diagram of data acquisition and preprocessing of the present invention, from left to right are: acquisition image, image segmentation level, hair layer transparent channel and marked feature points.
图2是本发明构造的人脸融合模型和头发形变模型的示例图,图中展示了一个用户三个表情的示例,图中第一行为输入的图像,第二行为重建的人脸、头发网格模型。Fig. 2 is an example figure of the face fusion model and the hair deformation model constructed by the present invention, in which an example of three facial expressions of a user is shown, the first line in the figure is the input image, and the second line is the reconstructed face and hair mesh grid model.
图3是本发明中基于图像的动态替身的动画结果示例图,图中展示了7个不同用户的例子,每一行从左到右,分别是:采集图像中的正面图像,重建的人脸、头发网格模型,5个不同姿态、不同表情的合成人脸动画结果。Fig. 3 is an example diagram of an animation result of an image-based dynamic avatar in the present invention, in which 7 examples of different users are shown, each line from left to right is respectively: the frontal image in the collected image, the reconstructed face, Hair mesh model, synthetic face animation results of 5 different poses and expressions.
图4是本发明中基于图像的动态替身的动画结果与真实图像的对比示例图,图中第一行是利用本发明生成的人脸动画,第二行是驱动生成人脸动画的真实图像。Fig. 4 is a comparison example diagram between the animation result of the image-based dynamic avatar and the real image in the present invention, the first row in the figure is the facial animation generated by the present invention, and the second row is the real image driven to generate the facial animation.
具体实施方式detailed description
在动态替身构造过程中,首先,本发明要采集用户一组特定动作和表情的图像中,然后,需要对这些图像进行预处理,包括分层和特征点标定;基于这些处理后的图像,建立一个三维人脸、头发模型。其中人脸模型可以通过匹配图像中而为特征点而拟合人脸容和模型;而对于头发几何模型,并没有一个通用的头发模型,这是因为对于不同的用户,头发的形状区别很大,很难用一个全局的模型来表示头发的形状。因此,本发明只能依赖于输入的图像来建立头发的三维模型。此外,对于头发,特别是长发,在不同的头部姿势下,由于重力和与身体交互等因素影响,会产生非刚性的变形。为了处理这些复杂的问题,本发明建立了一个可变形的头发模型,可以近似的模拟头部旋转中动态的头发变化。为了建立这一可变形的头发模型,本发明首先在每张采集图像中单独估算头发区域的深度;之后结合所有图像,执行一个头发深度的联合优化;最后基于这些优化后的深度构造一个全局的可变形头发模型。此外,在图像替身构造的另一方面,人体的其他部位,如眼睛,牙齿和身体被分别建模,分别用一个平板来表示。In the process of constructing dynamic avatars, firstly, the present invention collects a set of images of specific actions and expressions of the user, and then needs to preprocess these images, including layering and feature point calibration; based on these processed images, establish A 3D face and hair model. Among them, the face model can fit the face and model by matching the feature points in the image; for the hair geometric model, there is no general hair model, because the shape of the hair is very different for different users , it is difficult to use a global model to represent the shape of hair. Therefore, the present invention can only rely on the input image to build a three-dimensional model of the hair. In addition, for hair, especially long hair, under different head poses, due to factors such as gravity and interaction with the body, it will produce non-rigid deformation. In order to deal with these complicated problems, the present invention establishes a deformable hair model, which can approximate the dynamic hair changes in head rotation. In order to build this deformable hair model, the present invention first estimates the depth of the hair region separately in each captured image; then combines all images to perform a joint optimization of the hair depth; finally constructs a global model based on these optimized depths Deformable hair model. In addition, on the other hand of image avatar construction, other parts of the human body, such as eyes, teeth, and body, are modeled separately and represented by a flat panel, respectively.
在系统实时动画驱动的过程中,对于输入的每一帧图像,本发明首先使用现成的实时人脸跟踪方法得到图像中人脸动作参数,包括头部刚性变换参数和脸部非刚性表情系数,这些参数之后被转移到动态图像替身中生成该替身对应的三维网格。然后,在这个三维网格帮助下,系统映射和融合输入的采集图像,以生成替身在新视角下的人脸动画图像。其中,在融合过程中,本发明依赖三维几何计算融合过程中每张映射图像在最终图像中每个像素的权重,以保证最后最终结果中不同区域够光滑,无缝连接。本发明被用于拥有不同发型的不同用户上,得到令人信服的结果。In the process of real-time animation driving of the system, for each input frame image, the present invention first uses the ready-made real-time face tracking method to obtain the face movement parameters in the image, including head rigid transformation parameters and facial non-rigid expression coefficients, These parameters are then transferred to the dynamic image avatar to generate the 3D mesh corresponding to the avatar. Then, with the help of this 3D mesh, the system maps and fuses the input captured images to generate an animated image of the avatar's face from a new perspective. Among them, in the fusion process, the present invention relies on three-dimensional geometry to calculate the weight of each pixel in the final image of each mapped image in the fusion process, so as to ensure that different regions in the final final result are smooth enough and seamlessly connected. The invention was used on different users with different hairstyles with convincing results.
1、数据采集和预处理1. Data collection and preprocessing
1.1 数据采集1.1 Data collection
对每一个用户,本发明使用普通网络摄像头采集32张图像:包括15张不同头部姿势,17张不同表情。第一组的15张图像记录的是用户保持自然表情的不同头部姿势,这些头部姿势包括不同的旋转。这些头部姿势动作用旋转的欧拉角来表示,分别是:yaw方向从-60°到60°,间隔20°采样(保持pitch和roll方向为0°);pitch方向从-30°到30°,间隔15°采样但去除0°(保持其他方向为0°);roll方向的采样与pitch方向相同。用户做的这些头部旋转并不需要准确满足设定的角度,只需要近似的角度即可。For each user, the present invention uses a common webcam to collect 32 images: including 15 different head poses and 17 different expressions. The first set of 15 images records different head poses of the user maintaining a natural expression, and these head poses include different rotations. These head poses are represented by the Euler angles of rotation, which are: the yaw direction is from -60° to 60°, and the sampling interval is 20° (keep the pitch and roll directions at 0°); the pitch direction is from -30° to 30° °, sampling at intervals of 15° but removing 0° (keeping other directions as 0°); sampling in the roll direction is the same as the pitch direction. These head rotations made by the user do not need to meet the set angle exactly, only an approximate angle is required.
接下来,需要对采集图像进行两步的预处理:图像分割和特征点标定。Next, two steps of preprocessing are required for the collected images: image segmentation and feature point calibration.
1.2 图像分割1.2 Image Segmentation
预处理的第一步是对采集的图像进行分割,将每张图像分割为不同的层次:脸部,头发(包括头发上的头饰),眼镜,牙齿,身体和背景。本发明使用Lazy Snapping工具,在少量人工交互的基础上对每张图像进行分割。在本发明中,人工交互只需要在图像上简单的划定几笔就可以完成各层次的分割。此外,头发因为其半透明的特性而在边界存在复杂性,因此本发明在头发层进一步执行图像抽取算法,对头发层做进一步的处理,为头发层增加一个透明通道(alpha channel)。The first step of preprocessing is to segment the acquired images, segmenting each image into different layers: face, hair (including headgear on the hair), glasses, teeth, body and background. The present invention uses a Lazy Snapping tool to segment each image on the basis of a small amount of manual interaction. In the present invention, human interaction only needs to simply draw a few strokes on the image to complete the segmentation of each level. In addition, hair has complexity at the boundary due to its translucent characteristics, so the present invention further executes an image extraction algorithm on the hair layer, further processes the hair layer, and adds a transparent channel (alpha channel) to the hair layer.
1.3特征点标定1.3 Feature point calibration
在预处理的第二步中,本发明需要对采集的每张图像Ii,半自动的标定其特征点Si。这些特征点描述了图像中人脸一系列特征的二维位置,这些特征包括嘴巴、眼睛的轮廓,脸部轮廓等。本发明首先使用(CAO,C.,HOU,Q.,AND ZHOU,K.2014.Displaced dynamicexpression regression for real-time facial tracking and animation.ACMTrans.Graph.33,4(July),43:1–43:10.)中描述的实时人脸跟踪算法自动标定图像中的特征点,然后使用一个拖拽工具进行人工修正。In the second step of preprocessing, the present invention needs to semi-automatically mark the feature points S i of each collected image I i . These feature points describe the two-dimensional position of a series of features of the face in the image, these features include the mouth, the outline of the eyes, the outline of the face, etc. The present invention first uses (CAO, C., HOU, Q., AND ZHOU, K.2014. Displaced dynamic expression regression for real-time facial tracking and animation. ACMTrans. Graph. 33, 4 (July), 43: 1–43 The real-time face tracking algorithm described in :10.) automatically marks the feature points in the image, and then manually corrects them using a drag-and-drop tool.
2、基于图像的替身构造2. Image-based stand-in construction
2.1 人脸融合模型的构造2.1 Construction of face fusion model
基于标定后的采集图像{(Ii,Si)},本发明构造一个人脸融合模型,用于表示其低分辨率、动态的人脸几何。首先,基于FaceWarehouse人脸数据库,本发明为每张图像拟合生成一个初始的人脸网格然后本发明使用网格变形对每个进行修正,得到Fi;最后,本发明基于{Fi}计算该用户的表情融合模型{Bj}。Based on the calibrated collected images {(I i , S i )}, the present invention constructs a face fusion model to represent its low-resolution, dynamic face geometry. First, based on the FaceWarehouse face database, the present invention generates an initial face grid for each image fitting The present invention then uses mesh deformation for each Make corrections to obtain F i ; finally, the present invention calculates the user's facial expression fusion model {B j } based on {F i }.
首先,对每张图像(Ii,Si),本发明使用人脸仓库中的双线性模型,一个3阶的数据张量C,计算该用户全局的个体系数wid和每张图像中的表情系数来拟合人脸初始网格拟合的方法沿用(CAO,C.,WENG,Y.,LIN,S.,AND ZHOU,K.2013.3d shaperegression for real-time facial animation.ACM Trans.Graph.32,4(July),41:1–41:10.),首先基于人脸数据张量C生成该用户特定的融合模型,再根据该融合模型生成与每张图像匹配的网格 First, for each image (I i , S i ), the present invention uses a bilinear model in the face warehouse, a 3rd-order data tensor C, to calculate the user's global individual coefficient w id and in each image expression coefficient to fit the initial mesh of the face The fitting method follows (CAO, C., WENG, Y., LIN, S., AND ZHOU, K.2013.3d shape regression for real-time facial animation. ACM Trans.Graph.32, 4 (July), 41: 1–41:10.), first generate the user-specific fusion model based on the face data tensor C, and then generate a grid matching each image based on the fusion model
通过人脸仓库拟合得到的初始网格只是一个粗略的近似,为了让网格与图像匹配的更准确,本发明使用网格变形算法对网格做进一步的修正。网格修正的目的,是令每个图像Ii对应的网格Fi,其相应的顶点与图像二维特征点吻合,与本发明描述这一匹配能量为:The initial mesh obtained by fitting the face warehouse It is only a rough approximation. In order to match the grid with the image more accurately, the present invention uses a grid deformation algorithm to further correct the grid. The purpose of grid correction is to make the corresponding vertices of the grid F i corresponding to each image I i coincide with the two-dimensional feature points of the image, and the matching energy described in the present invention is:
其中Π(·)是相机的投影算子,将相机坐标系中的一个三维点投影得到图像中的二维点位置,是上一步中拟合的初始人脸网格,si,k是Si中第k个特征点的二维位置,vk是该特征点在网格中对应的顶点序号。Where Π(·) is the projection operator of the camera, which projects a 3D point in the camera coordinate system to obtain the 2D point position in the image, is the initial face grid fitted in the previous step, s i,k is the two-dimensional position of the kth feature point in S i , and v k is the corresponding vertex number of the feature point in the grid.
为了保证网格变形中网格的光滑,本发明在网格变形中加入一个拉普拉斯正则能量项:In order to ensure the smoothness of the grid in the grid deformation, the present invention adds a Laplace regular energy term in the grid deformation:
其中Δ是网格上基于余弦公式的离散拉普拉斯算子,δk是初始网格第k个顶点上的拉普拉斯向量的长度。经过网格变形修正后,每张图像得到准确匹配的网格Fi。where Δ is the discrete Laplacian operator on the grid based on the cosine formula, and δ k is the initial grid The length of the Laplacian vector at the kth vertex. After grid deformation correction, each image gets an exact matching grid F i .
基于这些修正的网格{Fi},结合拟合初始网格中得到的每张图像的表情系数使用(LI,H.,WEISE,T.,AND PAULY,M.2010.Example-based facial rigging.ACMTrans.Graph.29,4(July),32:1–32:6.)中描述的基于实例的人脸骨骼算法,就可以得到该用户修正的人脸融合模型{Bj}。Based on these modified grids {F i }, combined with the expression coefficients of each image obtained in fitting the initial grid Using the example-based facial rigging described in (LI, H., WEISE, T., AND PAULY, M. 2010. Example-based facial rigging. ACMTrans. Graph. 29, 4 (July), 32:1–32:6.) The face skeleton algorithm of the user can get the corrected face fusion model {B j }.
2.2 头发形变模型的构造2.2 Construction of hair deformation model
构造动画替身中头发的模型相比于构造人脸模型要困难的多。首先,不同人的发型迥异,很难使用一个全局的头发模型来表示各种不同的发型,只能完全从图像中重建头发的模型。此外,对于头发,特别是长头发,随着头的旋转,由于重力、与身体交互等因素的影响,会发生非刚性的运动。因此如果将头发看做一个刚性的物体,随着头部一起运动,就与采集的图像无法匹配,生成的人脸动画效果也会缺乏真实感。因此,本发明中为图像替身创建了一个形变的头发模型,用于近似模拟头发的动态变化。Modeling hair for an avatar is much more difficult than modeling a human face. First of all, the hairstyles of different people are very different, it is difficult to use a global hair model to represent various hairstyles, and only the hair model can be completely reconstructed from the image. Also, for hair, especially long hair, as the head rotates, there will be non-rigid motion due to gravity, interaction with the body, etc. Therefore, if the hair is regarded as a rigid object and moves with the head, it cannot match the collected image, and the generated facial animation effect will also lack realism. Therefore, in the present invention, a deformed hair model is created for the image stand-in to approximate the dynamic changes of the hair.
本发明只需要构建一个低分辨率的头发模型,用于在实时运行中帮助映射和融合图像。为了构建这样的头发模型,本发明首先在每张图像的头发区域估算深度值;所有的深度图之后被联合优化,其中为了联合优化发生非刚性运动的长发,本发明需要找到不同图像头发像素之间的对应关系;最后基于这些深度图,为每张图像生成一个拓扑结构一致的头发网格。The present invention only requires building a low-resolution hair model to help map and fuse images in real-time. In order to construct such a hair model, the present invention first estimates the depth value in the hair region of each image; all depth maps are then jointly optimized, and in order to jointly optimize the long hair with non-rigid motion, the present invention needs to find hair pixels in different images The correspondence between them; finally, based on these depth maps, a topologically consistent hair mesh is generated for each image.
本发明使用(CHAI,M.,WANG,L.,WENG,Y.,YU,Y.,GUO,B.,AND ZHOU,K.2012.Single-view hair modeling for portrait manipulation.ACMTrans.Graph.31,4(July),116:1–116:8.)中单视角头发建模的方法,来估算每张图像中头发的深度。该方法结合了边界和光滑两个能量项用于优化。首先,本发明基于分割出来的头发区域Ωh计算头发的轮廓而该轮廓上像素的深度值可以直接根据上一步中生成的人脸网格Fi进行初始化。初始深度D0设置方法如下:对于内部轮廓像素点(在图像上与脸部有重合),则直接将该像素点的深度设为人脸网格Fi上的深度;而对于外部的轮廓像素点,它们的深度设置为人脸网格Fi外部轮廓点深度的平均值。这样,边界能量项就可以被描述成:The present invention uses (CHAI, M., WANG, L., WENG, Y., YU, Y., GUO, B., AND ZHOU, K.2012.Single-view hair modeling for portrait manipulation.ACMTrans.Graph.31 , 4 (July), 116:1–116:8.) for single-view hair modeling to estimate the depth of hair in each image. This method combines both boundary and smooth energy terms for optimization. First, the present invention calculates the contour of the hair based on the segmented hair region Ω h And the depth value of the pixels on the contour can be directly initialized according to the face mesh F i generated in the previous step. The setting method of the initial depth D0 is as follows: for the inner contour pixel point (which coincides with the face on the image), the depth of the pixel point is directly set as the depth on the face grid F i ; and for the outer contour pixel point , and their depths are set as the average of the depths of the outer contour points of the face mesh F i . In this way, the boundary energy term can be described as:
其中,Dp是本发明需要求解的每个像素的深度值,是初始的深度,np是该像素的法向,而表示在二维图像上沿着头发轮廓的图像梯度。Wherein, D p is the depth value of each pixel that the present invention needs to solve, is the initial depth, n p is the normal direction of the pixel, and Represents the image gradient along a hair contour on a 2D image.
对于第二项的光滑能量,其用于保证求解得到的头发深度和法向尽量的平滑,可以描述为:For the smooth energy of the second term, it is used to ensure that the obtained hair depth and normal direction are as smooth as possible, which can be described as:
其中,p是头发区域Ωh中的一个像素,N(p)是像素p的四邻域像素集合,q是其中一个邻域像素,Dp和Dq分别是p和q的深度,np和nq分别是p和q的法向,ωd和ωn分别是用于控制深度和法向平滑的权重。通过联合优化能量Esil+Esm,即可得到每张图像Ii的深度Di。Among them, p is a pixel in the hair region Ω h , N(p) is the four-neighborhood pixel set of pixel p, q is one of the neighbor pixels, D p and D q are the depths of p and q respectively, n p and n q are the normals of p and q, respectively, and ω d and ω n are the weights used to control depth and normal smoothing, respectively. By jointly optimizing the energy E sil +E sm , the depth D i of each image I i can be obtained.
上述的深度计算方法对每张图像单独处理,没有考虑不同图像头发深度的一致性。以此生成的头发模型无法匹配每张图像。因此,本发明需要考虑不同图像之间的头发深度一致性,使用一个全局优化的方法,联合求解所有图像中的头发深度。该联合优化过程以一种交替方式迭代执行,在每次迭代过程中,依次处理每一张深度图;而在修正一张深度图Di时,会固定其他的深度图,并将其他的深度图作为约束对Di进行优化。具体来说,本发明首先将其他的深度图{Dj}j≠i转换到Di的相机坐标系中,表示为之后不同图像之间深度的一致性可以表示为Di和之间像素差异的总和,即:The above-mentioned depth calculation method processes each image separately, without considering the consistency of hair depth in different images. The resulting hair model cannot match every image. Therefore, the present invention needs to consider the hair depth consistency between different images, and use a global optimization method to jointly solve the hair depth in all images. The joint optimization process is iteratively executed in an alternating manner. In each iteration, each depth map is processed in turn; when a depth map D i is corrected, other depth maps will be fixed, and other depth maps will be fixed. The graph acts as a constraint to optimize D i . Specifically, the present invention first transforms other depth maps {D j } j≠i into the camera coordinate system of D i , expressed as Then the consistency of depth between different images can be expressed as D i and The sum of pixel differences between , that is:
其中,di,p和分别是Di和在像素p处的深度值,Di和分别是第i张图像的深度图和转换后的深度图。将这一联合约束能量与上述的边界能量项Esil和平滑能量项Esm结合起来联合优化,就可以得到联合优化后的深度图Di。Among them, d i, p and are D i and The depth value at pixel p, D i and are the depth map of the i-th image and the transformed depth map, respectively. Combining this joint constraint energy with the above-mentioned boundary energy item E sil and smooth energy item E sm for joint optimization, the jointly optimized depth map D i can be obtained.
接下来本发明将描述如何将一张深度图Dj转换到Di的相机坐标系中以生成对于用户短发的情况,可以假设头发随着头部刚性运动,因此可以通过两张图像中拟合的头部网格模型的刚性变换参数,将Dj通过刚性变换转换到Di的相机坐标系中,具体公式为:Next, the present invention will describe how to transform a depth map D j into the camera coordinate system of D i to generate For the case of the user with short hair, it can be assumed that the hair moves rigidly with the head, so D j can be converted to the camera coordinate system of D i through rigid transformation through the rigid transformation parameters of the head mesh model fitted in the two images , the specific formula is:
其中,P(·)表示将深度图中的一个像素点转换到相机坐标系下的一个三维点位置,R和T是在拟合人脸网格时,将网格从物体坐标系转换到相机坐标系的旋转和平移参数,di,p和分别是Di和在像素p处的深度值。Among them, P( ) means to convert a pixel point in the depth map to a three-dimensional point position in the camera coordinate system, and R and T are to convert the grid from the object coordinate system to the camera when fitting the face grid The rotation and translation parameters of the coordinate system, d i, p and are D i and Depth value at pixel p.
但对于长发的情况,头发的运动不能简单的被描述成与头部一起做刚性运动,因为头发会因为重力或与身体接触等因素发生非刚性的变化。因此,本发明需要计算Di和Dj之间的对应关系Cij,在后面内容中本发明详细介绍如何计算两张图像中头发的对应关系。基于这一对应关系,本发明对Dj做网格变形得到变形优化的能量描述如下:But in the case of long hair, the movement of the hair cannot simply be described as a rigid movement with the head, because the hair will undergo non-rigid changes due to factors such as gravity or contact with the body. Therefore, the present invention needs to calculate the corresponding relationship C ij between D i and D j . In the following content, the present invention introduces in detail how to calculate the corresponding relationship of hair in two images. Based on this correspondence, the present invention performs grid deformation on D j to obtain The energy description for deformation optimization is as follows:
其中,(ci,cj)是Cij中一组对应关系,和分别是Di和上对应像素点的深度值,顶点vk是根据深度图构造的网格上第k个顶点,拉普拉斯能量项中,Δ是网格上基于余弦公式的离散拉普拉斯算子,δk是初始网格第k个顶点上的拉普拉斯向量的长度,ωl用于控制拉普拉斯能量的权重,在本发明中被设置为10。Among them, (c i ,c j ) is a set of corresponding relations in C ij , and are D i and The depth value of the corresponding pixel on the vertex v k is based on the depth map The kth vertex on the constructed grid, in the Laplacian energy term, Δ is the discrete Laplacian operator based on the cosine formula on the grid, and δ k is the initial grid The length of the Laplacian vector on the kth vertex, ω l , is used to control the weight of the Laplacian energy, and is set to 10 in the present invention.
根据之前的描述,头发在随着头部旋转移动时可能会包含一些非刚性的运动。因此,在进行深度联合优化之前,本发明需要计算深度图Di和另外一张深度图Dj之间的对应关系。寻找对应关系的算法包括三个步骤:图像空间对应关系计算,粗略匹配修建和对应关系修正。According to the previous description, the hair may contain some non-rigid motion when moving with the head rotation. Therefore, before performing joint depth optimization, the present invention needs to calculate the corresponding relationship between the depth map D i and another depth map D j . The algorithm for finding correspondence includes three steps: image space correspondence calculation, rough matching construction and correspondence correction.
在第一个步骤中,本发明首先利用PatchMatch算法(BARNES,C.,SHECHTMAN,E.,FINKELSTEIN,A.,AND GOLDMAN,D.B.2009.Patchmatch:A randomized correspondencealgorithm for structural image editing.ACM Trans.Graph.28,3(July),24:1–24:11.)计算图像Ii和Ij中头发区域的对应关系。但这样生成的对应关系在某些像素上不够准确,因此在第二步中,本发明使用一个网格变形算法来计算一个粗略的匹配关系。本发明首先在图像Ii的头发区域构造一个规则的网格Pi,其中每个像素,结合Di中的深度,构建了网格中的一个顶点。之后,对于Pi中的每个顶点,如果该顶点周围的3×3邻域内的所有像素在PatchMatch算法计算中的误差小于给定的阈值0.05,且这些像素经过PatchMatch计算的偏移类似,本发明计算这些领域像素偏移的平均值,将其应用到原顶点中,移动到新的位置;而对于不满足上述的顶点则认为PatchMatch算法得到的偏移不合理。将找到Pi中的合理偏移作为位置约束,本发明使用拉普拉斯网格变形算法对Pi进行变形操作。变形后的网格Pi'被渲染到Ij的图像空间中,这样根据Pi和和Pi'在两张图像中的渲染投影,就可以得到Ii和Ij像素之间一个粗略的匹配。在最后一步中,基于这样一个粗略匹配,在Ii图像头发区域的每个像素,在Ij中对应像素中的9×9邻域中,进一步使用PatchMatch算法找到最吻合的像素。如果经过这一步的修正后,PatchMatch给出的误差依然大于阈值,则这个像素的对应关系被标定为不合法,在联合约束能量中将这个像素点约束去除即可。In the first step, the present invention first utilizes the PatchMatch algorithm (BARNES, C., SHECHTMAN, E., FINKELSTEIN, A., AND GOLDMAN, DB2009.Patchmatch: A randomized correspondence algorithm for structural image editing. ACM Trans.Graph.28 , 3 (July), 24:1–24:11.) Compute the correspondence of hair regions in images Ii and Ij . But the corresponding relationship generated in this way is not accurate enough on some pixels, so in the second step, the present invention uses a grid deformation algorithm to calculate a rough matching relationship. The present invention firstly constructs a regular grid P i in the hair area of image I i , where each pixel, combined with the depth in D i , constructs a vertex in the grid. Afterwards, for each vertex in P i , if the error of all pixels in the 3×3 neighborhood around the vertex in the PatchMatch algorithm calculation is less than the given threshold 0.05, and the offsets of these pixels after PatchMatch calculation are similar, this The invention calculates the average value of pixel offset in these areas, applies it to the original vertex, and moves to a new position; and for the vertex that does not meet the above, the offset obtained by the PatchMatch algorithm is considered unreasonable. Taking finding a reasonable offset in P i as a position constraint, the present invention uses a Laplacian grid deformation algorithm to deform P i . The deformed grid P i ' is rendered into the image space of I j , so that according to the rendering projections of P i and P i ' in the two images, a rough distance between the pixels of I i and I j can be obtained match. In the last step, based on such a rough match, for each pixel in the hair region of the I i image, in the 9×9 neighborhood of the corresponding pixel in I j , further use the PatchMatch algorithm to find the most matching pixel. If after this step of correction, the error given by PatchMatch is still greater than the threshold, the corresponding relationship of this pixel is calibrated as illegal, and the constraint of this pixel point can be removed in the joint constraint energy.
基于这些得到的头发深度图,就可以构造该用户的可变形头发模型。具体来说,本发明首先将所有图像的深度图转换、变形到同一个相机坐标系下,即第一张图像I1的相机空间中。对于任意一个深度图Dj(j≠1),首先基于像素的深度建立一个规则的网格,然后利用在计算对应关系时相同的算法对该规则网格进行变形。将变形后的网格每个顶点当做一个三维点,就得到了头发模型的三维点云。在该点云中,本发明使用以下规则移除点云中的异常点:如果某个点的法向的z方向上的值小于给定的阈值0.5,则标记该点为异常点。基于该三维点云,本发明使用泊松平面重建算法生成图像I1对应的头发网格H1。Based on these obtained hair depth maps, a deformable hair model of the user can be constructed. Specifically, the present invention first converts and warps the depth maps of all images into the same camera coordinate system, that is, the camera space of the first image I 1 . For any depth map D j (j≠1), a regular grid is established based on the depth of pixels, and then the regular grid is deformed by using the same algorithm when calculating the corresponding relationship. Taking each vertex of the deformed mesh as a 3D point, the 3D point cloud of the hair model is obtained. In the point cloud, the present invention uses the following rule to remove the abnormal points in the point cloud: if the value of the normal direction of a certain point in the z direction is less than a given threshold of 0.5, then mark the point as an abnormal point. Based on the 3D point cloud, the present invention uses a Poisson plane reconstruction algorithm to generate a hair mesh H 1 corresponding to the image I 1 .
由于其他图像Ij(j≠1)到I1的刚性旋转和对应关系在之前的计算中都已知,本发明最后将H1进行刚性变换和非刚性网格变形,将其转化到其他14张不同头部姿势的图像中,得到头发在不同头部姿势图像中的网格形状{Hi}i=1,2,...,15。本发明将这15个头发网格集合作为该用户的可变形头发模型,其表达了头发在不同头部姿势下的非刚性运动空间。需要注意的是,本发明假设脸部表情对头发的形状没有影响。Since the rigid rotation and corresponding relationship between other images I j (j≠1) and I 1 are known in previous calculations, the present invention finally performs rigid transformation and non-rigid grid deformation on H 1 , and transforms it into other 14 From the images of different head poses, the grid shape {H i } i=1,2,...,15 of the hair in the images of different head poses is obtained. The present invention uses these 15 hair grid sets as the user's deformable hair model, which expresses the non-rigid movement space of the hair under different head poses. It should be noted that the present invention assumes that facial expressions have no effect on hair shape.
2.3眼睛、牙齿和身体模型的构造2.3 Construction of eyes, teeth and body models
为了让替身更加完整,本发明继续描述构造其他部分的模型,包括眼睛、牙齿和身体。与脸部和头发不同,这些部分的动作相对来说较简单,且随着不同表情不同差异不大。因此,本发明基于正面自然表情图像(即I1)来构造眼睛和身体的模型,基于露牙齿表情的图像来构造牙齿的模型。To make the avatar more complete, the invention goes on to describe constructing models of other parts, including the eyes, teeth, and body. Unlike the face and hair, the animations of these parts are relatively simple and do not vary much with different expressions. Therefore, the present invention constructs eye and body models based on the natural facial expression image (i.e., I 1 ), and constructs a tooth model based on the tooth-showing expression image.
本发明中使用两个平板来表达眼睛:其中一个用于表达虹膜,另一个用于表达眼白。本发明首先基于每张图像中拟合的人脸网格,计算网格上相应眼睛顶点的包围盒,作为眼睛的矩形区域。虹膜的位置和大小则通过在该矩形中自动检测到的最大的椭圆,而这个检测到的虹膜之后被拷贝到虹膜的平板模型中。而对于眼白平板,本发明首先将眼睛图像拷贝到眼白平板中,并移除其中虹膜的区域。而对于移除虹膜后的缺失像素,本发明使用PatchMatch算法,将眼白区域作为来源来合成这些缺失像素的颜色。In this invention, two plates are used to express the eyes: one is used to express the iris, and the other is used to express the white of the eye. The present invention first calculates the bounding box of the corresponding eye vertices on the grid based on the fitted face grid in each image, as the rectangular area of the eye. The position and size of the iris is determined by automatically detecting the largest ellipse in the rectangle, and this detected iris is then copied to the tablet model of the iris. As for the white plate, the present invention first copies the eye image into the white plate, and removes the area of the iris. For the missing pixels after removing the iris, the present invention uses the PatchMatch algorithm to synthesize the colors of these missing pixels using the white area of the eye as a source.
本发明使用两个平板,分别表示替身中的上颚牙齿和下颚牙齿。在露牙齿的表情图像中,与眼睛模型类似,本发明首先根据人脸网格中相应的顶点位置计算包围盒,确定牙齿区域。牙齿平板的大小则由人脸网格中相应顶点来决定。在牙齿构造中,本发明还提供了一个拖拽工具用于手工修正不准确的平板位置。The present invention uses two plates representing the upper and lower teeth in the stand-in, respectively. In the tooth-showing expression image, similar to the eye model, the present invention first calculates the bounding box according to the corresponding vertex positions in the face grid to determine the tooth area. The size of the tooth plate is determined by the corresponding vertices in the face mesh. In tooth construction, the present invention also provides a drag tool for manually correcting inaccurate plate positions.
本发明用一个平板来近似上身的模型,该平板的颜色直接由I1中身体层的图像来填充。身体平板的深度则定义为人脸网格外轮廓点深度的平均值。The present invention approximates the model of the upper body with a flat plate whose color is directly filled by the image of the body layer in I1 . The depth of the body plate is defined as the average depth of the outline points of the face grid.
3、实时人脸和头发几何生成3. Real-time face and hair geometry generation
在实时运行过程中,本发明使用(CAO,C.,HOU,Q.,AND ZHOU,K.2014.Displaceddynamic expression regression for real-time facial tracking and animation.ACMTrans.Graph.33,4(July),43:1–43:10.)中的基于单目相机的人脸跟踪系统,来捕获用户的人脸动作并驱动图像替身。具体来说,在运行过程中,对于输入的每一个视频帧,人脸跟踪系统得到人脸动作的参数,包括刚性头部变换R,T和人脸表情系数e。基于这些参数,本发明生成当前帧替身人脸和头发的几何模型。During real-time operation, the present invention uses (CAO, C., HOU, Q., AND ZHOU, K.2014.Displaceddynamic expression regression for real-time facial tracking and animation.ACMTrans.Graph.33,4(July), 43:1–43:10.) for a monocular camera-based face tracking system to capture user facial motions and drive image avatars. Specifically, during the running process, for each input video frame, the face tracking system obtains the parameters of the face movement, including the rigid head transformation R, T and the facial expression coefficient e. Based on these parameters, the present invention generates a geometric model of the face and hair of the current frame stand-in.
3.1 人脸几何生成3.1 Face geometry generation
基于在预计算的替身人脸融合模型{Bj},本发明使用以下公式生成当前帧的人脸几何网格F:Based on the pre-calculated stand-in face fusion model {B j }, the present invention uses the following formula to generate the face geometric grid F of the current frame:
其中,R和T是人脸跟踪系统给出的当前帧人脸的刚性旋转和平移参数,B0是自然表情网格,ej是表情系数e第j项数值,Bj是融合模型中的表情网格。Among them, R and T are the rigid rotation and translation parameters of the current frame face given by the face tracking system, B 0 is the natural expression grid, e j is the value of the jth item of the expression coefficient e, and B j is the fusion model Emoticon grid.
为了使头部与脖子可以无缝连接,本发明需要固定头部模型中脖子位置点的位置,将其与身体平板固定,然后通过基于拉普拉斯的网格变形算法来更新其他点的位置。更新后的头部几何网格记做最后系统需要通过F和之间的三维配准来重新计算的刚性变换参数 In order to make the head and neck seamlessly connected, the present invention needs to fix the position of the neck position point in the head model, fix it with the body plate, and then update the positions of other points through the grid deformation algorithm based on Laplace . The updated head geometry mesh is denoted as Finally the system needs to pass F and 3D registration between to recalculate The rigid transformation parameter of
3.2 头发几何生成3.2 Hair Geometry Generation
基于上述生成的人脸几何网格本发明继续生成头发的几何网格H。与人脸网格构造方法类似,头发几何网格也是通过在预采集图像的头发网格{Hi}中插值得到:Based on the face geometric mesh generated above The present invention proceeds to generate the geometric mesh H of the hair. Similar to the face mesh construction method, the hair geometric mesh is also obtained by interpolating in the hair mesh {H i } of the pre-acquired image:
其中,和分别是上一步中计算得到的人脸刚性变换参数,ri是预采集图像的头发网格Hi的权重,该权重基于当前帧头部旋转和预采集图像中头部的旋转{Ri}计算:in, and are the face rigid transformation parameters calculated in the previous step, r i is the weight of the hair grid H i of the pre-acquired image, and the weight is based on the head rotation of the current frame and the rotation {R i } of the head in the pre-acquisition image is computed:
其中,ωr是插值参数,在本发明中被设为10,e是自然底数,当前帧头部旋转和预采集图像中头部的旋转{Ri}均是四元数表达。Among them, ω r is an interpolation parameter, which is set to 10 in the present invention, e is a natural base number, and the head of the current frame is rotated and the head rotation {R i } in the pre-acquisition image are expressed by quaternions.
4、实时人脸动画合成4. Real-time facial animation synthesis
4.1 映射图像4.1 Mapping images
基于生成的当前帧几何网格和H,结合预采集图像中的网格人脸网格{Fi}和头发网格{Hi},本发明映射预采集的图像{Ii},得到一系列映射后的图像,表示为 Based on generated current frame geometry mesh and H, combined with the grid face grid {F i } and hair grid {H i } in the pre-acquisition image, the present invention maps the pre-acquisition image {I i } to obtain a series of mapped images, expressed as
最终替身驱动的动画图像,它的每个像素都是通过映射图像中相对应像素加权平均得到。为了得到每张映射图像中每个像素的权重,本发明首先根据计算对应网格顶点上的权重,然后基于径向基函数来插值得到每个像素的权重。The final avatar-driven animation image, each pixel of which is obtained by mapping the weighted average of the corresponding pixels in the image. In order to obtain the weight of each pixel in each mapped image, the present invention first calculates the weight on the corresponding grid vertex, and then interpolates based on the radial basis function to obtain the weight of each pixel.
4.2 顶点权重计算4.2 Vertex weight calculation
具体来说,对每一个映射图像本发明首先计算网格和H上每个顶点vk的权重w(vk)。核心思想在于,当采集图像中与当前帧网格的法向/表情近似,且采集图像中网格顶点更朝向视线法向,则赋予该采集图像中该顶点较大的权重,用数学表达如下:Specifically, for each mapped image The present invention first calculates grid and the weight w(v k ) of each vertex v k on H. The core idea is that when the normal direction/expression of the grid in the captured image is similar to that of the current frame, and the grid vertex in the captured image is more towards the normal of the line of sight, then the vertex in the captured image is given a greater weight, expressed mathematically as follows :
其中,vk是网格中的某一顶点,vi,k是采集图像Ii中对应网格Fi/Hi中对应的顶点,ni,k/nk是顶点vi,k/vk对应的法向,是法向ni,k在z方向上的分量,ωz,ωn和ωe分别是控制各分量的相对重要性,在本发明中设为5,10和30。αi,k是为特定表情手工标定的网格上的二值模板,在与表情语义相关的区域设为1,而其余部分设为0。需要注意的是,手工标定二值模板的过程是一次性的,与替身本身没有关系。只需要在一个通用的表情融合模型上标定相应的模板,就可以用在所有的图像替身中。而ψ(ei,e)的计算方法如下:where v k is the mesh A certain vertex in , v i, k is the corresponding vertex in the corresponding grid F i /H i in the collected image I i , n i, k / n k is the normal direction corresponding to the vertex v i, k / v k , is the component of the normal n i, k in the z direction, ω z , ω n and ω e are the relative importance of the control components, which are set to 5, 10 and 30 in the present invention. α i,k is a binary template on a grid manually calibrated for a specific expression, set to 1 in regions related to expression semantics and 0 in the rest. It should be noted that the process of manually calibrating the binary template is a one-time process and has nothing to do with the avatar itself. It only needs to calibrate the corresponding template on a general expression fusion model, and it can be used in all image avatars. And ψ(e i ,e) is calculated as follows:
其中,(·)表示两个向量的点乘,ei和e分别是采集图像Ii和当前帧I的表情系数。Among them, (·) represents the dot product of two vectors, e i and e are the expression coefficients of the captured image I i and the current frame I respectively.
4.3 像素权重计算和图像合成4.3 Pixel weight calculation and image synthesis
在得到每张图像中每个网格顶点的权重后,系统就可以通过径向基函数插值计算每张图像Ii中单像素p上的权重wi,p:After obtaining the weight of each grid vertex in each image, the system can calculate the weight w i,p on a single pixel p in each image I i through radial basis function interpolation:
其中,up是像素p的二维坐标,ui,k是顶点vi,k在图像上的投影位置,wu用于控制每个顶点vi,k的图像影响区域大小,βk是可见性项,如果当前帧网格中的顶点vk在当前图像中可见则视为1,否则设为0。计算了所有图像的单像素权重后,这些权重被标准化,使得所有图像同一像素的权重和为1,即∑iwi,p=1。Among them, u p is the two-dimensional coordinates of pixel p, u i,k is the projected position of vertex v i,k on the image, w u is used to control the image influence area size of each vertex v i,k , β k is Visibility item, if the vertex v k in the grid of the current frame is visible in the current image, it is regarded as 1, otherwise it is set to 0. After calculating the single-pixel weights of all images, these weights are standardized so that the sum of the weights of the same pixel of all images is 1, that is, ∑ i w i,p =1.
需要注意的是,在实践中本发明并没有利用网格的所有顶点,而是通过在网格上均匀采样得到原顶点数1/10的采样顶点进行径向基函数插值。通过实验发现,使用这些小数量的点,不仅可以加快运算速度,还会帮助不同图像之间的融合更加平滑,可以得到令人满意的结果。It should be noted that in practice, the present invention does not use all the vertices of the grid, but performs radial basis function interpolation by sampling vertices uniformly on the grid to obtain 1/10 of the original number of vertices. Through experiments, it is found that using these small number of points can not only speed up the calculation speed, but also help the fusion between different images to be smoother, and can get satisfactory results.
在人脸和头发之外,在人脸动画合成的过程中,还需要产生替身其他部分的图像,包括眼睛、牙齿和身体,并将它们与脸部、头发自然的结合起来,生成最后的结果。In addition to the face and hair, in the process of face animation synthesis, it is also necessary to generate images of other parts of the avatar, including eyes, teeth and body, and combine them with the face and hair naturally to generate the final result .
对于眼睛,本发明首先在(CAO,C.,HOU,Q.,AND ZHOU,K.2014.Displaced dynamicexpression regression for real-time facial tracking and animation.ACMTrans.Graph.33,4(July),43:1–43:10.)中所有训练数据中增加两个指示虹膜位置的特征点,使得在实时运行过程中可以跟踪虹膜的位置,描述眼珠的转动。在生成替身的眼睛图像过程中,对于眼白平板,本发明直接让其与头部一起做刚性运动,即对眼白平板直接应用刚性变换转换到相应位置。而对于虹膜平板,除了刚性变换以外,本发明需要根据跟踪得到的虹膜特征点位置计算虹膜的平移,将该平移加到该虹膜平板上,实现眼珠的转动。For eyes, the present invention is first described in (CAO, C., HOU, Q., AND ZHOU, K.2014.Displaced dynamic expression regression for real-time facial tracking and animation.ACMTrans.Graph.33,4(July),43: 1–43:10.), adding two feature points indicating the position of the iris to all the training data, so that the position of the iris can be tracked during real-time operation, and the rotation of the eyeball can be described. In the process of generating the eye image of the stand-in, for the white plate of the eye, the present invention directly makes it perform rigid motion together with the head, that is, directly applies rigid transformation to the white plate of the eye switch to the corresponding position. And for the iris slab, in addition to the rigid transformation In addition, the present invention needs to calculate the translation of the iris according to the positions of iris feature points obtained by tracking, and add the translation to the iris plate to realize the rotation of the eyeball.
对于牙齿,本发明将上颚牙齿的平板与头部直接连接,且随着头部一起做刚性运动,利用刚性变换参数直接对齐做刚性变换即可;而下颚牙齿则与下巴一起运动。For the teeth, the present invention directly connects the flat plate of the upper jaw teeth with the head, and performs rigid movement together with the head, and utilizes rigid transformation parameters Direct alignment is done with a rigid transformation; the lower teeth move with the jaw.
对于身体,本发明令其与头部一起平移,即将平移参数应用在身体平板,将其变换到相应位置并渲染图像空间中,作为其他部分(人脸、头发、牙齿和眼睛)的背景。For the body, the present invention makes it translate together with the head, that is, the translation parameter Applies to the body slab, transforms it into position and renders it in image space as a background for other parts (face, hair, teeth and eyes).
本发明中只使用一个低分辨率网格来表达替身,而人脸上的细节信息如皮肤折叠、皱纹等信息,被隐式的表达在本发明中的基于图像的替身表达中。此外,更为重要的是,本发明中基于图像的替身表达,可以有效处理头发和头发上的头饰,而这些效果在之前的已有工作中都无法处理。In the present invention, only a low-resolution grid is used to express the avatar, and detailed information on the human face, such as skin folds and wrinkles, is implicitly expressed in the image-based avatar expression in the present invention. In addition, more importantly, the image-based stand-in representation in the present invention can effectively deal with hair and headgear on the hair, and these effects cannot be dealt with in the previous work.
实施实例Implementation example
本发明描述的方法在一台普通台式电脑(Intel i7 3.6GHz CPU,32GB内存,NVidia Gtx 760显卡)上被实现。所有的采集图像都是用一个普通的网络摄像头采集的,它可以提供1280×720分辨率的图像。为了构造一个图像动画替身,通常需要10分钟用于图像采集,40分钟用于图像预处理以及15分钟用于计算人脸融合模型和头发形变模型。在实时运行时,本发明结合(CAO,C.,HOU,Q.,AND ZHOU,K.2014.Displaced dynamic expressionregression for real-time facial tracking and animation.ACM Trans.Graph.33,4(July),43:1–43:10.)实时人脸跟踪系统,需要约30毫秒为每一帧输入图像驱动图像替身生成动画效果。总体来说,本发明在普通台式电脑上越能达到25帧/秒的速度。The method described in the present invention is realized on an ordinary desktop computer (Intel i7 3.6GHz CPU, 32GB memory, NVidia Gtx 760 graphics card). All captured images were captured with a common webcam, which can provide images with 1280×720 resolution. In order to construct an image animation stand-in, it usually takes 10 minutes for image acquisition, 40 minutes for image preprocessing and 15 minutes for calculating the face fusion model and hair deformation model. When running in real time, the present invention combines (CAO, C., HOU, Q., AND ZHOU, K.2014.Displaced dynamic expression regression for real-time facial tracking and animation.ACM Trans.Graph.33,4(July), 43:1–43:10.) A real-time face tracking system that takes about 30 milliseconds for each frame of the input image to drive the image avatar to generate animation effects. In general, the present invention can achieve a speed of 25 frames per second on a common desktop computer.
从实践结果中可以看到,本发明可以为拥有不同发型、头饰的不同用户创建真实的动态替身。与之前工作中为脸部细节如皱纹生成几何层次不同,得益于基于图像的替身表达方法,本发明的结果可以很自然的包含用户的各种皱纹细节。而本发明也同样可以处理大幅度的旋转和因此带来的头发变化,充分体现了本发明中人脸融合模型和头发形变模型的作用。It can be seen from the practical results that the present invention can create real dynamic avatars for different users with different hairstyles and headgears. Different from generating geometric layers for facial details such as wrinkles in previous work, thanks to the image-based stand-in expression method, the result of the present invention can naturally contain various wrinkle details of the user. And the present invention can also deal with large-scale rotation and the hair changes brought about by it, which fully embodies the functions of the human face fusion model and the hair deformation model in the present invention.
Claims (5)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610331428.3A CN106023288B (en) | 2016-05-18 | 2016-05-18 | An Image-Based Dynamic Stand-In Construction Method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610331428.3A CN106023288B (en) | 2016-05-18 | 2016-05-18 | An Image-Based Dynamic Stand-In Construction Method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN106023288A true CN106023288A (en) | 2016-10-12 |
CN106023288B CN106023288B (en) | 2019-11-15 |
Family
ID=57097457
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610331428.3A Active CN106023288B (en) | 2016-05-18 | 2016-05-18 | An Image-Based Dynamic Stand-In Construction Method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106023288B (en) |
Cited By (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106599811A (en) * | 2016-11-29 | 2017-04-26 | 叶飞 | Facial expression tracking method of VR heat-mounted display |
CN107705276A (en) * | 2017-09-11 | 2018-02-16 | 广东欧珀移动通信有限公司 | Image processing method and device, electronic installation and computer-readable recording medium |
CN107707839A (en) * | 2017-09-11 | 2018-02-16 | 广东欧珀移动通信有限公司 | Image processing method and device |
CN107784689A (en) * | 2017-10-18 | 2018-03-09 | 唐越山 | Method for running, system and collecting device based on data acquisition avatar service |
CN108154550A (en) * | 2017-11-29 | 2018-06-12 | 深圳奥比中光科技有限公司 | Face real-time three-dimensional method for reconstructing based on RGBD cameras |
CN109215131A (en) * | 2017-06-30 | 2019-01-15 | Tcl集团股份有限公司 | The driving method and device of conjecture face |
CN110060205A (en) * | 2019-05-08 | 2019-07-26 | 北京迈格威科技有限公司 | Image processing method and device, storage medium and electronic equipment |
CN110163957A (en) * | 2019-04-26 | 2019-08-23 | 李辉 | A kind of expression generation system based on aestheticism face program |
CN110443885A (en) * | 2019-07-18 | 2019-11-12 | 西北工业大学 | Three-dimensional number of people face model reconstruction method based on random facial image |
CN111182350A (en) * | 2019-12-31 | 2020-05-19 | 广州华多网络科技有限公司 | Image processing method, image processing device, terminal equipment and storage medium |
CN111292276A (en) * | 2018-12-07 | 2020-06-16 | 北京字节跳动网络技术有限公司 | Image processing method and device |
CN111382634A (en) * | 2018-12-29 | 2020-07-07 | 河南中原大数据研究院有限公司 | Three-dimensional face recognition method based on depth video stream |
CN111445384A (en) * | 2020-03-23 | 2020-07-24 | 杭州趣维科技有限公司 | Universal portrait photo cartoon stylization method |
CN111460872A (en) * | 2019-01-18 | 2020-07-28 | 北京市商汤科技开发有限公司 | Image processing method and apparatus, image device, and storage medium |
CN111583367A (en) * | 2020-05-22 | 2020-08-25 | 构范(厦门)信息技术有限公司 | Hair simulation method and system |
CN113014832A (en) * | 2019-12-19 | 2021-06-22 | 志贺司 | Image editing system and image editing method |
CN113066156A (en) * | 2021-04-16 | 2021-07-02 | 广州虎牙科技有限公司 | Expression redirection method, device, equipment and medium |
CN113269888A (en) * | 2021-05-25 | 2021-08-17 | 山东大学 | Hairstyle three-dimensional modeling method, character three-dimensional modeling method and system |
CN116012497A (en) * | 2023-03-29 | 2023-04-25 | 腾讯科技(深圳)有限公司 | Animation redirection method, device, equipment and medium |
CN117808943A (en) * | 2024-02-29 | 2024-04-02 | 天度(厦门)科技股份有限公司 | Three-dimensional cartoon face reconstruction method, device, equipment and storage medium |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP4377896A1 (en) * | 2021-07-29 | 2024-06-05 | Digital Domain Virtual Human (US), Inc. | System and method for animating secondary features |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102800129A (en) * | 2012-06-20 | 2012-11-28 | 浙江大学 | Hair modeling and portrait editing method based on single image |
CN103093490A (en) * | 2013-02-02 | 2013-05-08 | 浙江大学 | Real-time facial animation method based on single video camera |
CN103606186A (en) * | 2013-02-02 | 2014-02-26 | 浙江大学 | Virtual hair style modeling method of images and videos |
CN103942822A (en) * | 2014-04-11 | 2014-07-23 | 浙江大学 | Facial feature point tracking and facial animation method based on single video vidicon |
-
2016
- 2016-05-18 CN CN201610331428.3A patent/CN106023288B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102800129A (en) * | 2012-06-20 | 2012-11-28 | 浙江大学 | Hair modeling and portrait editing method based on single image |
CN103093490A (en) * | 2013-02-02 | 2013-05-08 | 浙江大学 | Real-time facial animation method based on single video camera |
CN103606186A (en) * | 2013-02-02 | 2014-02-26 | 浙江大学 | Virtual hair style modeling method of images and videos |
CN103942822A (en) * | 2014-04-11 | 2014-07-23 | 浙江大学 | Facial feature point tracking and facial animation method based on single video vidicon |
Non-Patent Citations (1)
Title |
---|
CHEN CAO等: "Face Warehouse: a 3D Facial Expression Database for Visual Computing", 《 IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS》 * |
Cited By (30)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106599811A (en) * | 2016-11-29 | 2017-04-26 | 叶飞 | Facial expression tracking method of VR heat-mounted display |
CN106599811B (en) * | 2016-11-29 | 2019-11-05 | 苏州虚现数字科技有限公司 | A kind of VR aobvious facial expression method for tracing |
CN109215131A (en) * | 2017-06-30 | 2019-01-15 | Tcl集团股份有限公司 | The driving method and device of conjecture face |
CN109215131B (en) * | 2017-06-30 | 2021-06-01 | Tcl科技集团股份有限公司 | Virtual face driving method and device |
CN107705276A (en) * | 2017-09-11 | 2018-02-16 | 广东欧珀移动通信有限公司 | Image processing method and device, electronic installation and computer-readable recording medium |
CN107707839A (en) * | 2017-09-11 | 2018-02-16 | 广东欧珀移动通信有限公司 | Image processing method and device |
CN107784689A (en) * | 2017-10-18 | 2018-03-09 | 唐越山 | Method for running, system and collecting device based on data acquisition avatar service |
CN108154550B (en) * | 2017-11-29 | 2021-07-06 | 奥比中光科技集团股份有限公司 | RGBD camera-based real-time three-dimensional face reconstruction method |
CN108154550A (en) * | 2017-11-29 | 2018-06-12 | 深圳奥比中光科技有限公司 | Face real-time three-dimensional method for reconstructing based on RGBD cameras |
CN111292276A (en) * | 2018-12-07 | 2020-06-16 | 北京字节跳动网络技术有限公司 | Image processing method and device |
CN111382634B (en) * | 2018-12-29 | 2023-09-26 | 河南中原大数据研究院有限公司 | Three-dimensional face recognition method based on depth video stream |
CN111382634A (en) * | 2018-12-29 | 2020-07-07 | 河南中原大数据研究院有限公司 | Three-dimensional face recognition method based on depth video stream |
CN111460872A (en) * | 2019-01-18 | 2020-07-28 | 北京市商汤科技开发有限公司 | Image processing method and apparatus, image device, and storage medium |
CN111460872B (en) * | 2019-01-18 | 2024-04-16 | 北京市商汤科技开发有限公司 | Image processing method and device, image equipment and storage medium |
CN110163957A (en) * | 2019-04-26 | 2019-08-23 | 李辉 | A kind of expression generation system based on aestheticism face program |
CN110060205B (en) * | 2019-05-08 | 2023-08-08 | 北京迈格威科技有限公司 | Image processing method and device, storage medium and electronic equipment |
CN110060205A (en) * | 2019-05-08 | 2019-07-26 | 北京迈格威科技有限公司 | Image processing method and device, storage medium and electronic equipment |
CN110443885A (en) * | 2019-07-18 | 2019-11-12 | 西北工业大学 | Three-dimensional number of people face model reconstruction method based on random facial image |
CN113014832A (en) * | 2019-12-19 | 2021-06-22 | 志贺司 | Image editing system and image editing method |
CN113014832B (en) * | 2019-12-19 | 2024-03-12 | 志贺司 | Image editing system and image editing method |
CN111182350A (en) * | 2019-12-31 | 2020-05-19 | 广州华多网络科技有限公司 | Image processing method, image processing device, terminal equipment and storage medium |
CN111445384A (en) * | 2020-03-23 | 2020-07-24 | 杭州趣维科技有限公司 | Universal portrait photo cartoon stylization method |
CN111583367B (en) * | 2020-05-22 | 2023-02-10 | 构范(厦门)信息技术有限公司 | Hair simulation method and system |
CN111583367A (en) * | 2020-05-22 | 2020-08-25 | 构范(厦门)信息技术有限公司 | Hair simulation method and system |
CN113066156A (en) * | 2021-04-16 | 2021-07-02 | 广州虎牙科技有限公司 | Expression redirection method, device, equipment and medium |
CN113269888A (en) * | 2021-05-25 | 2021-08-17 | 山东大学 | Hairstyle three-dimensional modeling method, character three-dimensional modeling method and system |
CN116012497B (en) * | 2023-03-29 | 2023-05-30 | 腾讯科技(深圳)有限公司 | Animation redirection method, device, equipment and medium |
CN116012497A (en) * | 2023-03-29 | 2023-04-25 | 腾讯科技(深圳)有限公司 | Animation redirection method, device, equipment and medium |
CN117808943A (en) * | 2024-02-29 | 2024-04-02 | 天度(厦门)科技股份有限公司 | Three-dimensional cartoon face reconstruction method, device, equipment and storage medium |
CN117808943B (en) * | 2024-02-29 | 2024-07-05 | 天度(厦门)科技股份有限公司 | Three-dimensional cartoon face reconstruction method, device, equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN106023288B (en) | 2019-11-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106023288B (en) | An Image-Based Dynamic Stand-In Construction Method | |
CN109377557B (en) | Real-time three-dimensional face reconstruction method based on single-frame face image | |
CN101916454B (en) | Method for reconstructing high-resolution human face based on grid deformation and continuous optimization | |
Ichim et al. | Dynamic 3D avatar creation from hand-held video input | |
CN105427385B (en) | A kind of high-fidelity face three-dimensional rebuilding method based on multilayer deformation model | |
Plänkers et al. | Tracking and modeling people in video sequences | |
Hasler et al. | Multilinear pose and body shape estimation of dressed subjects from image sets | |
CN103606186B (en) | The virtual hair style modeling method of a kind of image and video | |
CN100416612C (en) | 3D Dynamic Facial Expression Modeling Method Based on Video Stream | |
CN101739719B (en) | 3D meshing method for 2D frontal face images | |
CN103473801B (en) | A kind of human face expression edit methods based on single camera Yu movement capturing data | |
CN113744374B (en) | Expression-driven 3D virtual image generation method | |
US11928778B2 (en) | Method for human body model reconstruction and reconstruction system | |
CN105913416A (en) | Method for automatically segmenting three-dimensional human face model area | |
CN114842136B (en) | A single-image 3D face reconstruction method based on differentiable renderer | |
Cong | Art-directed muscle simulation for high-end facial animation | |
CN110796719A (en) | Real-time facial expression reconstruction method | |
Ichim et al. | Building and animating user-specific volumetric face rigs | |
CN108564619B (en) | Realistic three-dimensional face reconstruction method based on two photos | |
CN115951784B (en) | A motion capture and generation method for clothed human body based on dual neural radiation field | |
CN117036620B (en) | Three-dimensional face reconstruction method based on single image | |
Pighin et al. | Realistic facial animation using image-based 3D morphing | |
JP7251003B2 (en) | Face mesh deformation with fine wrinkles | |
Tejera et al. | Animation control of surface motion capture | |
US20240169635A1 (en) | Systems and Methods for Anatomically-Driven 3D Facial Animation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |