[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN107622257A - A kind of neural network training method and three-dimension gesture Attitude estimation method - Google Patents

A kind of neural network training method and three-dimension gesture Attitude estimation method Download PDF

Info

Publication number
CN107622257A
CN107622257A CN201710954487.0A CN201710954487A CN107622257A CN 107622257 A CN107622257 A CN 107622257A CN 201710954487 A CN201710954487 A CN 201710954487A CN 107622257 A CN107622257 A CN 107622257A
Authority
CN
China
Prior art keywords
gesture
mrow
neural network
depth
network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710954487.0A
Other languages
Chinese (zh)
Inventor
王好谦
李达
方璐
王兴政
张永兵
戴琼海
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Weilai Media Technology Research Institute
Shenzhen Graduate School Tsinghua University
Original Assignee
Shenzhen Weilai Media Technology Research Institute
Shenzhen Graduate School Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Weilai Media Technology Research Institute, Shenzhen Graduate School Tsinghua University filed Critical Shenzhen Weilai Media Technology Research Institute
Priority to CN201710954487.0A priority Critical patent/CN107622257A/en
Publication of CN107622257A publication Critical patent/CN107622257A/en
Pending legal-status Critical Current

Links

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a kind of neural network training method and three-dimension gesture Attitude estimation method, including:S1:The data set of multiple gesture depth maps is included by depth camera collection;S2:Random forest learner is trained using step S1 data set;S3:Multiple gesture depth maps in step S1 data set are split using random forest learner, it is partitioned into gesture subgraph, again the gesture subgraph is handled to obtain processing figure, multiple gesture depth maps in the processing figure and step S1 data set are carried out out of order to be divided into training set and test set;S4:The obtained training sets of step S3 and test set are used for training convolutional neural networks, training obtains network model.Three-dimension gesture Attitude estimation method is that the three-dimension gesture posture in individual depth picture is estimated using the network model.The present invention can accurately identify the particular location and posture of palm finger in gesture.

Description

Neural network training method and three-dimensional gesture attitude estimation method
Technical Field
The invention relates to the field of computer vision and deep learning, in particular to a neural network training method and a three-dimensional gesture posture estimation method.
Background
In recent years, with the rapid development of computer vision and deep learning, virtual reality and augmented reality technologies are gradually popularized, and still have immeasurable development prospects. As an important means of human-computer interaction, the gesture recognition technology has been highly concerned by the field of computer vision, and because the human hand has more joints, complicated shape, higher degree of freedom and easy shielding phenomenon, the rapid and accurate recognition of the gesture position and the hand action is always a difficult problem.
Conventional gesture pose estimation methods can be generally divided into two categories: sensor based and image based. The gesture posture estimation technology based on the sensor is that the sensors such as an accelerometer, an angular velocity meter and the like are fixed at specific parts of a palm and a finger of a person; the position and motion state information of a specific part of a human hand are acquired through the worn sensor equipment, and then the states of the palm and the fingers of the human hand are calculated by using a kinematics method, so that the gesture posture estimation effect is achieved; the method has great limitation on gesture detection due to the fact that sensor equipment needs to be worn, and detection errors are generally large under the influence of factors such as the accuracy of the sensor and the change of the wearing position. Another gesture posture estimation method based on images is generally to use edge or region detection-based methods such as edge detection and skin color detection on images including human hands shot by an RGB camera, firstly determine approximate regions of the human hands in the images, and then segment detailed information such as fingers and wrists by means of image segmentation and the like; because a picture containing a hand is shot by a common camera, the picture can only reflect plane information of a scene generally, if occlusion occurs between fingers, action details of the occluded fingers cannot be identified, and therefore a large error exists.
The above background disclosure is only for the purpose of assisting understanding of the concept and technical solution of the present invention and does not necessarily belong to the prior art of the present patent application, and should not be used for evaluating the novelty and inventive step of the present application in the case that there is no clear evidence that the above content is disclosed at the filing date of the present patent application.
Disclosure of Invention
In order to solve the technical problems, the invention provides a neural network training method and a three-dimensional gesture attitude estimation method, which can accurately identify the specific position and attitude of a palm finger in a gesture.
In order to achieve the purpose, the invention adopts the following technical scheme:
the invention discloses a neural network training method, which comprises the following steps:
s1: acquiring, by a depth camera, a data set comprising a plurality of gesture depth maps;
s2: training a random forest learner by using the data set of the step S1;
s3: adopting a random forest learner to segment the gesture depth maps in the data set of the step S1 to obtain gesture sub-maps, processing the gesture sub-maps to obtain a processing map, and disordering the processing map and the gesture depth maps in the data set of the step S1 to divide the processing map and the gesture depth maps into a training set and a testing set;
s4: and (5) using the training set and the test set obtained in the step (S3) to train the convolutional neural network, and training to obtain a network model.
Preferably, the step S3 of processing the gesture sub-graph to obtain a processing graph includes: s32: projecting the gesture subgraph in the direction of three axes X, Y, Z to obtain three single-channel projection graphs; wherein the processing map includes the projection map in step S32.
Preferably, the step S3 of processing the gesture sub-graph to obtain a processing graph further includes: s33: respectively carrying out down-sampling on the three projection drawings to obtain a plurality of down-sampling drawings with different sizes; wherein the processing map includes the projection map in step S32 and the down-sampling map in step S33.
Preferably, step S1 specifically includes:
s11: acquiring a plurality of gesture depth maps of different people by adopting a plurality of depth cameras;
s12: labeling each gesture depth map, and storing a plurality of gesture depth maps and corresponding labeling information in a data set.
Preferably, the labeling of each gesture depth map in step S12 specifically includes: and labeling coordinate information (x, y, d) on the preset position of the finger and the palm in each gesture depth map, wherein x and y are horizontal and vertical coordinates on the gesture depth map, and d is the pixel depth.
Preferably, the predetermined position of the finger comprises all joint points of the finger.
Preferably, step S4 specifically includes:
s41: randomly selecting m pictures and corresponding label information from the training set, and randomly selecting n pictures and corresponding label information from the testing set;
s42: the pictures are rolled and laminated in the network;
s43: pictures are passed through a pooling layer in the network;
s44: the output layer restores the picture;
s45: calculating the error between the network output and the label information, learning the network, and updating the network parameters;
s46: repeatedly iterating the steps S42-S45, and continuously updating the parameters until the parameters are converged; and storing the trained parameters to finally obtain the trained network model.
Preferably, step S45 is specifically: the formula for calculating the error between the network output and the tag information is as follows:
wherein,for predicted tag coordinates, byComposition, J is original label, consisting of (J)1,j2,...,jn) The composition, n is the number of labels,
assuming that the network parameter of the neuron in the network is ω, updating the network parameter according to the following formula:
the invention also discloses a three-dimensional gesture posture estimation method, and the three-dimensional gesture posture in a single depth picture is estimated by adopting the network model obtained by training of the neural network training method.
Compared with the prior art, the invention has the beneficial effects that: according to the neural network training method, the gesture depth map acquired by the depth camera can accurately identify the gesture and position information of the palm and each finger, then the gesture depth map is segmented by the random forest learner, so that the feature information of the gesture in the picture can be beneficially explored, the residual neural network is trained through the set of the picture, and the convolution pooling layer of the neural network can learn the features in the regions with different scales in the picture, so that the trained network model is applied to three-dimensional gesture posture estimation, the shielding influence can be weakened, and the image-based method is not restricted by wearable equipment; by using the residual convolutional neural network, the problem of gradient dispersion in the process of back propagation and parameter updating is avoided, so that the network training effect is better; the three-dimensional gesture posture estimation method combines the use of a depth learning method and a depth camera and is applied to gesture recognition, and the influence of factors such as illumination change, object shielding and the like can be reduced for the gesture recognition.
In a further scheme, by performing horizontal, vertical and depth triaxial projection on the gesture subgraph, a picture with a three-dimensional view angle in the picture can be obtained, and the feature information of the gesture can be conveniently found; furthermore, the gesture subgraphs are downsampled to obtain multiscale pictures with different sizes, so that the pixel characteristics and the region characteristics with different sizes of the pictures can be better developed, and the specific positions and postures of the palms and fingers in the gestures can be more accurately identified by a trained network model; and can accurately recognize the detailed information of the occluded gesture.
Drawings
FIG. 1 is a flow chart diagram of a three-dimensional gesture pose estimation method in accordance with a preferred embodiment of the present invention;
FIG. 2 is a schematic diagram of the marker points of the fingers and palm of the hand of the preferred embodiment of the present invention;
FIG. 3 is a schematic diagram of the steps of the neural network training method according to the preferred embodiment of the present invention.
Detailed Description
The invention will be further described with reference to the accompanying drawings and preferred embodiments.
As shown in fig. 1, the three-dimensional gesture posture estimation method of the preferred embodiment of the present invention includes the following steps:
s1: collecting a data set of a gesture depth map; the method specifically comprises the following steps:
s11: acquiring gesture depth pictures of different people by using a plurality of depth cameras, acquiring a plurality of pictures containing a plurality of gestures with different angles and different postures for each gesture of each person, and sorting the acquired pictures into a picture library;
s12: labeling each picture in the picture library; in order to accurately position the detailed information of the positions and postures of the joints of the gesture, coordinate information (x, y, d) is marked at specific positions of fingers and palms in the embodiment, wherein x and y are horizontal and vertical coordinates of a foreground image (on a gesture depth image), and d is pixel depth which is the embodiment of the gesture depth on the image; as shown in fig. 2, a plurality of key points are set at specific positions of fingers and palms as specific mark points, the hand on each picture is marked as a label of a picture library, and the picture name and the corresponding label are stored in a file form; wherein the key points marked in fig. 2 include all the joint points of the five fingers and the important position point of the palm, the posture of the current hand can be accurately estimated by accurately predicting the position of each joint point.
S2: training a random forest learner by using the data set obtained in the step S1;
s3: preprocessing a gesture depth map in the data set; as shown in fig. 3, the method specifically includes the following steps:
s31: a random forest learner is adopted to segment the gesture depth map in the data set in the step S1 to obtain a gesture sub-map;
s32: projecting the gesture subgraph in the direction of three axes X, Y, Z to obtain three single-channel projection graphs;
s33: respectively carrying out down-sampling on the three projection drawings to obtain a plurality of down-sampling drawings with different sizes;
s34: sorting all the gesture depth maps in the data set in the step S1, the projection map obtained in the step S32 and the downsampling map obtained in the step S33, and dividing the data into a training set and a testing set according to the proportion of 90% and 10%.
S4: using the training set and the test set obtained in the step S34 to train a convolutional neural network, and training to obtain a network model; the method specifically comprises the following steps:
s41: randomly selecting m pictures and corresponding label information from the training set, and randomly selecting n pictures and corresponding label information from the testing set;
s42: the pictures are rolled and laminated in the network; assuming that the original size of the picture is l × l, k square matrices with the same size and different pixel values are selected as convolution kernels, and the size of the convolution kernels can be represented as k × c. Wherein k is the number of convolution kernels, and c is the number of parameters of each dimension of the convolution kernels; and (3) performing convolution operation on each picture and k convolution kernels respectively to obtain k pictures with the same size but not the same pixel point. New dimension lc*lcThe size is shown in the following formula:
lc*lc=(l-c+1)*(l-c+1)
s43: pictures pass through a pooling layer in a network; assuming that the size of the picture before entering the pooling layer is l x l, pooling is a sliding of an area of size p x p over the picture each time in f steps; each sliding time, one pixel in the area is selected to represent all pixels of the area, and the size of each picture is changed into l after passing through the pooling layerp*lpAs shown in the following formula:
lp*lp=((lp-(f-p))/f)*((lp-(f-p))/f)
s44: after the picture is processed by convolution pooling and the like in a network, an output layer restores the predicted picture; assuming that h convolution kernels are output after network training, and when the h convolution kernels reach the inlet of an output layer through the network, assuming that the size of each picture is le*le(le<li),liThe size of the picture entering the output layer is h x le*leDimension l is coupled through the output layero*loIs recovered to li*li
S45: calculating the difference between the network output and the standard label, learning the network, and updating the network parameters; the Euclidean distance of the calculated error is shown as follows:
wherein,for predicted tag coordinates, byAnd (4) forming. J is the original label, consisting of (J)1,j2,...,jn) The composition, n is the number of labels,therefore, it is
Assuming that the network parameter of the neuron in the network is ω, updating the network parameter according to the following formula:
s46: repeatedly iterating the steps S42-S45, and continuously updating the parameters until the parameters are converged; and storing the trained parameters to finally obtain the trained convolutional neural network model.
S5: and estimating the three-dimensional gesture attitude of the single depth picture by adopting the convolutional neural network model obtained by training in the step S4.
The preferred embodiment of the invention also discloses a neural network training method, which comprises the steps S1 to S4.
According to the three-dimensional gesture posture estimation method in the preferred embodiment of the invention, a depth camera is used for collecting a large number of pictures; segmenting the gesture foreground by using a random forest classifier; manually marking gesture joint point information; training a convolutional neural network using a data set; storing the trained convolutional network, and directly using the convolutional network for three-dimensional gesture attitude estimation of a single depth map; the method applies the use of the depth learning method and the depth camera to gesture recognition, and the recognition of the gesture can reduce the influence of factors such as illumination change, object shielding and the like.
Acquiring a gesture posture image through a depth camera, displaying the depth information of a gesture in the form of a single-channel gray-scale image which represents the distance between an object and the camera by the pixel value, and restoring a gesture posture skeleton in the form of joint points according to the acquired gesture posture depth image; because the convolution pooling layer of the convolutional neural network can learn features in regions of different scales in the picture, occlusion effects can be reduced, and the image-based method is not constrained by the wearable device.
The three-dimensional gesture posture estimation method of the preferred embodiment of the invention overcomes the bottleneck of traditional gesture posture estimation, and realizes the accurate recognition of the specific position and posture of the palm and the fingers in the gesture by using a detection method of a human hand depth map shot by a depth camera and a new convolutional neural network method based on deep learning.
The foregoing is a more detailed description of the invention in connection with specific preferred embodiments and it is not intended that the invention be limited to these specific details. For those skilled in the art to which the invention pertains, several equivalent substitutions or obvious modifications can be made without departing from the spirit of the invention, and all the properties or uses are considered to be within the scope of the invention.

Claims (9)

1. A neural network training method is characterized by comprising the following steps:
s1: acquiring, by a depth camera, a data set comprising a plurality of gesture depth maps;
s2: training a random forest learner by using the data set of the step S1;
s3: adopting a random forest learner to segment the gesture depth maps in the data set of the step S1 to obtain gesture sub-maps, processing the gesture sub-maps to obtain a processing map, and disordering the processing map and the gesture depth maps in the data set of the step S1 to divide the processing map and the gesture depth maps into a training set and a testing set;
s4: and (5) using the training set and the test set obtained in the step (S3) to train the convolutional neural network, and training to obtain a network model.
2. The neural network training method of claim 1, wherein the processing the gesture sub-graph in step S3 to obtain a processing graph comprises:
s32: projecting the gesture subgraph in the direction of three axes X, Y, Z to obtain three single-channel projection graphs;
wherein the processing map includes the projection map in step S32.
3. The neural network training method of claim 2, wherein the processing the gesture sub-graph in step S3 to obtain a processing graph further comprises:
s33: respectively carrying out down-sampling on the three projection drawings to obtain a plurality of down-sampling drawings with different sizes;
wherein the processing map further includes the down-sampling map in step S33.
4. The neural network training method according to claim 1, wherein step S1 specifically includes:
s11: acquiring a plurality of gesture depth maps of different people by adopting a plurality of depth cameras;
s12: labeling each gesture depth map, and storing a plurality of gesture depth maps and corresponding labeling information in a data set.
5. The neural network training method of claim 4, wherein labeling each gesture depth map in step S12 specifically includes: and labeling coordinate information (x, y, d) on the preset position of the finger and the palm in each gesture depth map, wherein x and y are horizontal and vertical coordinates on the gesture depth map, and d is the pixel depth.
6. The neural network training method of claim 5, wherein the predetermined positions of the finger include all joint points of the finger.
7. The neural network training method according to claim 4, wherein the step S4 specifically includes:
s41: randomly selecting m pictures and corresponding label information from the training set, and randomly selecting n pictures and corresponding label information from the testing set;
s42: the pictures are rolled and laminated in the network;
s43: pictures are passed through a pooling layer in the network;
s44: the output layer restores the picture;
s45: calculating the error between the network output and the label information, learning the network, and updating the network parameters;
s46: repeatedly iterating the steps S42-S45, and continuously updating the parameters until the parameters are converged; and storing the trained parameters to finally obtain the trained network model.
8. The neural network training method according to claim 7, wherein the step S45 specifically comprises: the formula for calculating the error between the network output and the tag information is as follows:
<mrow> <mi>E</mi> <mi>r</mi> <mi>r</mi> <mi>o</mi> <mi>r</mi> <mo>=</mo> <mfrac> <mn>1</mn> <mn>2</mn> </mfrac> <mo>|</mo> <mo>|</mo> <mover> <mi>J</mi> <mo>^</mo> </mover> <mo>-</mo> <mi>J</mi> <mo>|</mo> <msup> <mo>|</mo> <mn>2</mn> </msup> <mo>=</mo> <mfrac> <mn>1</mn> <mn>2</mn> </mfrac> <mrow> <mo>|</mo> <mrow> <msup> <mrow> <mo>(</mo> <mover> <mi>x</mi> <mo>^</mo> </mover> <mo>-</mo> <mi>x</mi> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>+</mo> <msup> <mrow> <mo>(</mo> <mover> <mi>y</mi> <mo>^</mo> </mover> <mo>-</mo> <mi>y</mi> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>+</mo> <msup> <mrow> <mo>(</mo> <mover> <mi>d</mi> <mo>^</mo> </mover> <mo>-</mo> <mi>d</mi> <mo>)</mo> </mrow> <mn>2</mn> </msup> </mrow> <mo>|</mo> </mrow> </mrow>
wherein,for predicted tag coordinates, byComposition, J is original label, consisting of (J)1,j2,...,jn) Composition, n is the number of labels, ji=(xi,yi,di),
Assuming that the network parameter of the neuron in the network is ω, updating the network parameter according to the following formula:
<mrow> <msup> <mi>&amp;omega;</mi> <mo>&amp;prime;</mo> </msup> <mo>=</mo> <mi>&amp;omega;</mi> <mo>+</mo> <mfrac> <mrow> <mo>&amp;part;</mo> <mi>E</mi> <mi>r</mi> <mi>r</mi> <mi>o</mi> <mi>r</mi> </mrow> <mrow> <mo>&amp;part;</mo> <mi>&amp;omega;</mi> </mrow> </mfrac> <mo>.</mo> </mrow>
9. a three-dimensional gesture posture estimation method is characterized in that a network model obtained by training through the neural network training method according to any one of claims 1 to 8 is adopted to estimate the three-dimensional gesture posture in a single depth picture.
CN201710954487.0A 2017-10-13 2017-10-13 A kind of neural network training method and three-dimension gesture Attitude estimation method Pending CN107622257A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710954487.0A CN107622257A (en) 2017-10-13 2017-10-13 A kind of neural network training method and three-dimension gesture Attitude estimation method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710954487.0A CN107622257A (en) 2017-10-13 2017-10-13 A kind of neural network training method and three-dimension gesture Attitude estimation method

Publications (1)

Publication Number Publication Date
CN107622257A true CN107622257A (en) 2018-01-23

Family

ID=61092146

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710954487.0A Pending CN107622257A (en) 2017-10-13 2017-10-13 A kind of neural network training method and three-dimension gesture Attitude estimation method

Country Status (1)

Country Link
CN (1) CN107622257A (en)

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108681402A (en) * 2018-05-16 2018-10-19 Oppo广东移动通信有限公司 Identify exchange method, device, storage medium and terminal device
CN108717524A (en) * 2018-04-28 2018-10-30 天津大学 It is a kind of based on double gesture recognition systems and method for taking the photograph mobile phone and artificial intelligence system
CN108960178A (en) * 2018-07-13 2018-12-07 清华大学 A kind of manpower Attitude estimation method and system
CN108960036A (en) * 2018-04-27 2018-12-07 北京市商汤科技开发有限公司 3 D human body attitude prediction method, apparatus, medium and equipment
CN109035327A (en) * 2018-06-25 2018-12-18 北京大学 Panorama camera Attitude estimation method based on deep learning
CN109215080A (en) * 2018-09-25 2019-01-15 清华大学 6D Attitude estimation network training method and device based on deep learning Iterative matching
CN109858380A (en) * 2019-01-04 2019-06-07 广州大学 Expansible gesture identification method, device, system, gesture identification terminal and medium
CN110197156A (en) * 2019-05-30 2019-09-03 清华大学 Manpower movement and the shape similarity metric method and device of single image based on deep learning
WO2019196099A1 (en) * 2018-04-09 2019-10-17 深圳大学 Method for positioning boundaries of target object in medical image, storage medium, and terminal
CN110555335A (en) * 2018-05-30 2019-12-10 深圳市掌网科技股份有限公司 Gesture recognition artificial intelligence training system and method thereof
CN111104820A (en) * 2018-10-25 2020-05-05 中车株洲电力机车研究所有限公司 Gesture recognition method based on deep learning
CN111325166A (en) * 2020-02-26 2020-06-23 南京工业大学 Sitting posture identification method based on projection reconstruction and multi-input multi-output neural network
CN111368733A (en) * 2020-03-04 2020-07-03 电子科技大学 Three-dimensional hand posture estimation method based on label distribution learning, storage medium and terminal
CN111428555A (en) * 2020-01-17 2020-07-17 大连理工大学 Joint-divided hand posture estimation method
CN112085161A (en) * 2020-08-20 2020-12-15 清华大学 Graph neural network method based on random information transmission
CN112836597A (en) * 2021-01-15 2021-05-25 西北大学 Multi-hand posture key point estimation method based on cascade parallel convolution neural network
CN113408443A (en) * 2021-06-24 2021-09-17 齐鲁工业大学 Gesture posture prediction method and system based on multi-view images
TWI777153B (en) * 2020-04-21 2022-09-11 和碩聯合科技股份有限公司 Image recognition method and device thereof and ai model training method and device thereof

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105005769A (en) * 2015-07-08 2015-10-28 山东大学 Deep information based sign language recognition method
CN106326860A (en) * 2016-08-23 2017-01-11 武汉闪图科技有限公司 Gesture recognition method based on vision
CN106815578A (en) * 2017-01-23 2017-06-09 重庆邮电大学 A kind of gesture identification method based on Depth Motion figure Scale invariant features transform
CN107103613A (en) * 2017-03-28 2017-08-29 深圳市未来媒体技术研究院 A kind of three-dimension gesture Attitude estimation method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105005769A (en) * 2015-07-08 2015-10-28 山东大学 Deep information based sign language recognition method
CN106326860A (en) * 2016-08-23 2017-01-11 武汉闪图科技有限公司 Gesture recognition method based on vision
CN106815578A (en) * 2017-01-23 2017-06-09 重庆邮电大学 A kind of gesture identification method based on Depth Motion figure Scale invariant features transform
CN107103613A (en) * 2017-03-28 2017-08-29 深圳市未来媒体技术研究院 A kind of three-dimension gesture Attitude estimation method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
贺芳姿: "基于Kinect深度信息的手势识别研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019196099A1 (en) * 2018-04-09 2019-10-17 深圳大学 Method for positioning boundaries of target object in medical image, storage medium, and terminal
CN108960036A (en) * 2018-04-27 2018-12-07 北京市商汤科技开发有限公司 3 D human body attitude prediction method, apparatus, medium and equipment
CN108717524A (en) * 2018-04-28 2018-10-30 天津大学 It is a kind of based on double gesture recognition systems and method for taking the photograph mobile phone and artificial intelligence system
CN108717524B (en) * 2018-04-28 2022-05-06 天津大学 Gesture recognition system based on double-camera mobile phone and artificial intelligence system
CN108681402A (en) * 2018-05-16 2018-10-19 Oppo广东移动通信有限公司 Identify exchange method, device, storage medium and terminal device
WO2019218880A1 (en) * 2018-05-16 2019-11-21 Oppo广东移动通信有限公司 Interaction recognition method and apparatus, storage medium, and terminal device
CN110555335B (en) * 2018-05-30 2022-12-27 深圳市掌网科技股份有限公司 Gesture recognition artificial intelligence training system and method thereof
CN110555335A (en) * 2018-05-30 2019-12-10 深圳市掌网科技股份有限公司 Gesture recognition artificial intelligence training system and method thereof
CN109035327B (en) * 2018-06-25 2021-10-29 北京大学 Panoramic camera attitude estimation method based on deep learning
CN109035327A (en) * 2018-06-25 2018-12-18 北京大学 Panorama camera Attitude estimation method based on deep learning
CN108960178A (en) * 2018-07-13 2018-12-07 清华大学 A kind of manpower Attitude estimation method and system
CN109215080A (en) * 2018-09-25 2019-01-15 清华大学 6D Attitude estimation network training method and device based on deep learning Iterative matching
US11200696B2 (en) 2018-09-25 2021-12-14 Tsinghua University Method and apparatus for training 6D pose estimation network based on deep learning iterative matching
CN111104820A (en) * 2018-10-25 2020-05-05 中车株洲电力机车研究所有限公司 Gesture recognition method based on deep learning
CN109858380A (en) * 2019-01-04 2019-06-07 广州大学 Expansible gesture identification method, device, system, gesture identification terminal and medium
CN110197156A (en) * 2019-05-30 2019-09-03 清华大学 Manpower movement and the shape similarity metric method and device of single image based on deep learning
CN111428555A (en) * 2020-01-17 2020-07-17 大连理工大学 Joint-divided hand posture estimation method
CN111428555B (en) * 2020-01-17 2022-09-20 大连理工大学 Joint-divided hand posture estimation method
CN111325166A (en) * 2020-02-26 2020-06-23 南京工业大学 Sitting posture identification method based on projection reconstruction and multi-input multi-output neural network
CN111325166B (en) * 2020-02-26 2023-07-07 南京工业大学 Sitting posture identification method based on projection reconstruction and MIMO neural network
CN111368733A (en) * 2020-03-04 2020-07-03 电子科技大学 Three-dimensional hand posture estimation method based on label distribution learning, storage medium and terminal
TWI777153B (en) * 2020-04-21 2022-09-11 和碩聯合科技股份有限公司 Image recognition method and device thereof and ai model training method and device thereof
CN112085161A (en) * 2020-08-20 2020-12-15 清华大学 Graph neural network method based on random information transmission
CN112085161B (en) * 2020-08-20 2022-12-13 清华大学 Graph neural network method based on random information transmission
CN112836597A (en) * 2021-01-15 2021-05-25 西北大学 Multi-hand posture key point estimation method based on cascade parallel convolution neural network
CN112836597B (en) * 2021-01-15 2023-10-17 西北大学 Multi-hand gesture key point estimation method based on cascade parallel convolution neural network
CN113408443A (en) * 2021-06-24 2021-09-17 齐鲁工业大学 Gesture posture prediction method and system based on multi-view images

Similar Documents

Publication Publication Date Title
CN107622257A (en) A kind of neural network training method and three-dimension gesture Attitude estimation method
CN107103613B (en) A kind of three-dimension gesture Attitude estimation method
JP7178396B2 (en) Method and computer system for generating data for estimating 3D pose of object included in input image
CN108345869B (en) Driver posture recognition method based on depth image and virtual data
CN108717531B (en) Human body posture estimation method based on Faster R-CNN
CN106055091B (en) A kind of hand gestures estimation method based on depth information and correcting mode
Hasan et al. RETRACTED ARTICLE: Static hand gesture recognition using neural networks
CN112906604B (en) Behavior recognition method, device and system based on skeleton and RGB frame fusion
Prisacariu et al. 3D hand tracking for human computer interaction
CN108256504A (en) A kind of Three-Dimensional Dynamic gesture identification method based on deep learning
CN112560741A (en) Safety wearing detection method based on human body key points
CN104794737B (en) A kind of depth information Auxiliary Particle Filter tracking
WO2021238548A1 (en) Region recognition method, apparatus and device, and readable storage medium
WO2015000286A1 (en) Three-dimensional interactive learning system and method based on augmented reality
CN111401266B (en) Method, equipment, computer equipment and readable storage medium for positioning picture corner points
CN113034652A (en) Virtual image driving method, device, equipment and storage medium
CN104537705A (en) Augmented reality based mobile platform three-dimensional biomolecule display system and method
CN111046734A (en) Multi-modal fusion sight line estimation method based on expansion convolution
CN110941996A (en) Target and track augmented reality method and system based on generation of countermeasure network
CN114641799A (en) Object detection device, method and system
JP2016014954A (en) Method for detecting finger shape, program thereof, storage medium of program thereof, and system for detecting finger shape
EP3185212B1 (en) Dynamic particle filter parameterization
CN116883588A (en) Method and system for quickly reconstructing three-dimensional point cloud under large scene
CN111709268A (en) Human hand posture estimation method and device based on human hand structure guidance in depth image
Zou et al. Automatic reconstruction of 3D human motion pose from uncalibrated monocular video sequences based on markerless human motion tracking

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20180123

RJ01 Rejection of invention patent application after publication