CN107977636A - Method for detecting human face and device, terminal, storage medium - Google Patents
Method for detecting human face and device, terminal, storage medium Download PDFInfo
- Publication number
- CN107977636A CN107977636A CN201711305306.8A CN201711305306A CN107977636A CN 107977636 A CN107977636 A CN 107977636A CN 201711305306 A CN201711305306 A CN 201711305306A CN 107977636 A CN107977636 A CN 107977636A
- Authority
- CN
- China
- Prior art keywords
- pixel
- detection
- window
- feature
- face
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
- G06V40/166—Detection; Localisation; Normalisation using acquisition arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- General Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
Abstract
The disclosure is directed to a kind of method for detecting human face and device, terminal, storage medium, belong to Face datection field.This method includes:Capture the depth image of human body by the stereo camera shooting component of terminal, depth image includes multiple pixels, the distance between the correspondence position point and stereo camera shooting component of each pixel of depth value expression of each pixel in multiple pixels on human body;The depth value of each pixel in multiple pixels, determines the corresponding window size of each pixel;According to position of each pixel in depth image and the corresponding window size of each pixel, the corresponding detection window of each pixel is determined;Using the corresponding detection window of each pixel, Face datection is carried out in depth image.The disclosure helps to solve the problem of Face datection process is complicated, and detection speed is slower, simplifies Face datection process, improves detection speed.The disclosure is used for Face datection.
Description
Technical field
This disclosure relates to Face datection field, more particularly to a kind of method for detecting human face and device, terminal, storage medium.
Background technology
With the rise of the full frame technology of terminal, fingerprint recognition component is gradually removed from the front panel of terminal, recognition of face
Component gradually replaces fingerprint recognition component, becomes the important identity verification component of terminal.Terminal based on recognition of face component into
During row authentication, it is necessary first to carry out Face datection.
In correlation technique, recognition of face component mainly includes the front camera of terminal, terminal when carrying out Face datection,
Human body image is captured by front camera first, then in each pixel of human body image, using various sizes of detection
Window extracts characteristics of image from human body image, is judged to whether there is face in detection window according to the characteristics of image extracted,
To carry out Face datection.
During the disclosure is realized, inventor has found that correlation technique has at least the following problems:
Correlation technique needs each pixel in human body image to carry out Face datection using various sizes of detection window,
Therefore, Face datection process is complicated, and detection speed is slower.
The content of the invention
Present disclose provides a kind of method for detecting human face and device, terminal, storage medium, Face datection process can be solved
The problem of complexity, detection speed is slower.The technical solution of the disclosure is as follows:
According to the first aspect of the disclosure, there is provided a kind of method for detecting human face, the described method includes:
The depth image of human body is captured by the stereo camera shooting component of terminal, the depth image includes multiple pixels,
The depth value of each pixel in the multiple pixel represent each correspondence position point of the pixel on human body with
The distance between described stereo camera shooting component;
The depth value of each pixel in the multiple pixel, determines the corresponding window of each pixel
Size;
According to position of each pixel in the depth image and the corresponding window ruler of each pixel
It is very little, determine the corresponding detection window of each pixel;
Using the corresponding detection window of each pixel, Face datection is carried out in the depth image.
Alternatively, the detection window is positive K sides shape window, and the K is the integer more than or equal to 3, the window ruler
The very little length of side for the detection window, the depth value of each pixel in the multiple pixel, determines described
The corresponding window size of each pixel, including:According to the depth value of each pixel, determine that formula determines using size
The corresponding window size of each pixel;
Wherein, the size determines that formula is f=ax2+ bx+c, the f represent the window size, and the x represents institute
Depth value is stated, a, the b and the c are constant.
Alternatively, the method further includes:
In the depth value according to each pixel, determine that formula determines that each pixel is corresponding using size
Before window size, window size sample set is obtained, the window size sample set is a including p window size and with the p
The one-to-one p depth value of window size, the p are the integer more than or equal to 1;
According to the p window size and the p depth value, the value of a, the b and the c are determined.
Alternatively, it is described using the corresponding detection window of each pixel, face is carried out in the depth image
Detection, including:
The each corresponding region of the corresponding detection window of pixel in the depth image is determined as described every
The corresponding detection zone of a pixel;
From the corresponding detection zone of each pixel, characteristics of image is extracted;
Detect whether described image feature is people's face feature.
Alternatively, whether the detection described image feature is people's face feature, including:Using described in the detection of character separation device
Whether characteristics of image is face characteristic;
Wherein, the feature classifiers be grader generating means using feature samples collection as training data, according to classification
What Algorithm for Training obtained be used to determine characteristics of image whether be face characteristic grader, the feature samples concentrate that to include n a
Face characteristic sample and m non-face feature samples, the n and the m are the integer more than 0.
Alternatively, it is described to use character separation device whether to detect described image feature for people's face feature, including:
Receive the feature classifiers that the grader generating means are sent;
The described image feature extracted is inputted into the feature classifiers, to pass through the feature classifiers output characteristic
Label, the feature tag are used to indicating the classification of described image feature, the classification of described image feature include face characteristic or
Non-face feature;
According to the feature tag, determine whether described image feature is people's face feature.
Alternatively, the depth value of each pixel in the multiple pixel, determines each pixel
The corresponding window size of point, including:At least one detection pixel is determined from the multiple pixel, according to described at least one
The depth value of each detection pixel in a detection pixel, determines the corresponding window size of each detection pixel;
The position and the corresponding window of each pixel according to each pixel in the depth image
Mouth size, determines the corresponding detection window of each pixel, including:According to each detection pixel in the depth
Position and the corresponding window size of each detection pixel in image, determine the corresponding inspection of each detection pixel
Survey window;
It is described to use the corresponding detection window of each pixel, Face datection is carried out in the depth image, is wrapped
Include:Using the corresponding detection window of each detection pixel, Face datection is carried out in the depth image.
Alternatively, the method further includes:
Before the depth image of human body is captured by the stereo camera shooting component of terminal, detect whether the terminal is in quilt
Grip state;
When the terminal is held state described in, the stereo camera shooting component of the terminal is opened.
Above-mentioned all optional technical solutions, can use any combination to form the alternative of the disclosure, not another herein
One repeats.
According to the second aspect of the disclosure, there is provided a kind of human face detection device, described device include:
Trapping module, is configured as capturing the depth image of human body, the depth map by the stereo camera shooting component of terminal
As including multiple pixels, the depth value of each pixel in the multiple pixel represents each pixel in human body
On correspondence position point and the distance between the stereo camera shooting component;
First determining module, the depth value for each pixel being configured as in the multiple pixel, determines institute
State the corresponding window size of each pixel;
Second determining module, is configured as according to position of each pixel in the depth image and described every
The corresponding window size of a pixel, determines the corresponding detection window of each pixel;
First detection module, is configured as using the corresponding detection window of each pixel, in the depth image
Middle carry out Face datection.
Alternatively, the detection window is positive K sides shape window, and the K is the integer more than or equal to 3, the window ruler
The very little length of side for the detection window, first determining module, is configured as the depth value according to each pixel, adopts
Determine that formula determines the corresponding window size of each pixel with size;
Wherein, the size determines that formula is f=ax2+ bx+c, the f represent the window size, and the x represents institute
Depth value is stated, a, the b and the c are constant.
Alternatively, described device further includes:
Acquisition module, is configured as, in the depth value according to each pixel, determining that formula determines institute using size
Before stating the corresponding window size of each pixel, window size sample set is obtained, the window size sample set includes p window
Mouthful size and be the integer more than or equal to 1 with the one-to-one p depth value of the p window size, the p;
3rd determining module, is configured as, according to the p window size and the p depth value, determining a, institute
State the value of b and the c.
Alternatively, the first detection module, including:
Determination sub-module, is configured as each pair of the corresponding detection window of pixel in the depth image
Region is answered to be determined as the corresponding detection zone of each pixel;
Extracting sub-module, is configured as from the corresponding detection zone of each pixel, extracts characteristics of image;
Detection sub-module, is configured as whether detection described image feature is people's face feature.
Alternatively, the detection sub-module, is configured as using character separation device whether to detect described image feature for people
Face feature;
Wherein, the feature classifiers be grader generating means using feature samples collection as training data, according to classification
What Algorithm for Training obtained be used to determine characteristics of image whether be face characteristic grader, the feature samples concentrate that to include n a
Face characteristic sample and m non-face feature samples, the n and the m are the integer more than 0.
Alternatively, the detection sub-module, is configured as:
Receive the feature classifiers that the grader generating means are sent;
The described image feature extracted is inputted into the feature classifiers, to pass through the feature classifiers output characteristic
Label, the feature tag are used to indicating the classification of described image feature, the classification of described image feature include face characteristic or
Non-face feature;
According to the feature tag, determine whether described image feature is people's face feature.
Alternatively, first determining module, is configured as determining at least one detection picture from the multiple pixel
Vegetarian refreshments, the depth value of each detection pixel at least one detection pixel, determines each detection picture
The corresponding window size of vegetarian refreshments;
Second determining module, is configured as the position in the depth image according to each detection pixel
Window size corresponding with each detection pixel, determines the corresponding detection window of each detection pixel;
The first detection module, is configured as using the corresponding detection window of each detection pixel, described
Face datection is carried out in depth image.
Alternatively, described device further includes:
Second detection module, is configured as before the depth image of human body is captured by the stereo camera shooting component of terminal,
The terminal is detected whether in being held state;
Opening module, is configured as the terminal and is in described when being held state, the solid for opening the terminal is taken the photograph
As component.
According to the third aspect of the disclosure, there is provided a kind of terminal, including:
Processor;
Memory for the executable instruction for storing the processor;
Wherein, the processor is configured as:
The depth image of human body is captured by the stereo camera shooting component of terminal, the depth image includes multiple pixels,
The depth value of each pixel in the multiple pixel represent each correspondence position point of the pixel on human body with
The distance between described stereo camera shooting component;
The depth value of each pixel in the multiple pixel, determines the corresponding window of each pixel
Size;
According to position of each pixel in the depth image and the corresponding window ruler of each pixel
It is very little, determine the corresponding detection window of each pixel;
Using the corresponding detection window of each pixel, Face datection is carried out in the depth image.
According to the fourth aspect of the disclosure, there is provided a kind of computer-readable recording medium, the computer-readable storage medium
Instruction is stored with matter, when described instruction is being run on the processing component of terminal so that the processing component performs first party
The method for detecting human face that any optional mode of face or first aspect is provided.
The technical solution that the disclosure provides can include the following benefits:
The method for detecting human face and device, terminal, storage medium that the disclosure provides, due to can be according to each pixel
Depth value determines the corresponding window size of each pixel, according to position of each pixel in depth image and each pixel
The corresponding window size of point, determines the corresponding detection window of each pixel, using the corresponding detection window of each pixel,
Face datection is carried out in depth image, it is therefore not necessary to carry out face inspection using various sizes of detection window in each pixel
Survey, help to solve the problem of Face datection process is complicated, and detection speed is slower, simplify Face datection process, improve detection speed
Degree.
It should be appreciated that the general description and following detailed description of the above are only exemplary and explanatory, not
The disclosure can be limited.
Brief description of the drawings
In order to illustrate more clearly of embodiment of the disclosure, attached drawing needed in embodiment description will be made below
Simply introduce, it should be apparent that, drawings in the following description are only some embodiments of the present disclosure, common for this area
For technical staff, without creative efforts, other attached drawings can also be obtained according to these attached drawings.
Fig. 1 is a kind of method flow diagram of method for detecting human face according to an exemplary embodiment;
Fig. 2 is a kind of method flow diagram of the method for detecting human face shown according to another exemplary embodiment;
Fig. 3 is the method flow diagram of the value of a kind of definite a, b and c according to an exemplary embodiment;
Fig. 4 is that one kind according to an exemplary embodiment uses the corresponding detection window of pixel in depth image
Carry out the method flow diagram of Face datection;
Fig. 5 is a kind of schematic diagram of detection zone according to an exemplary embodiment;
Fig. 6 be a kind of detection image feature according to an exemplary embodiment whether be people's face feature method flow
Figure;
Fig. 7 is a kind of block diagram of human face detection device according to an exemplary embodiment;
Fig. 8 is a kind of block diagram of the human face detection device shown according to another exemplary embodiment;
Fig. 9 is a kind of block diagram of first detection module according to an exemplary embodiment;
Figure 10 is a kind of block diagram of human face detection device according to another exemplary embodiment;
Figure 11 is a kind of structure diagram of terminal according to an exemplary embodiment.
Attached drawing herein is merged in specification and forms the part of this specification, shows the implementation for meeting the disclosure
Example, and be used to together with specification to explain the principle of the disclosure.
Embodiment
Here exemplary embodiment will be illustrated in detail, its example is illustrated in the accompanying drawings.Following description is related to
During attached drawing, unless otherwise indicated, the same numbers in different attached drawings represent the same or similar key element.Following exemplary embodiment
Described in embodiment do not represent all embodiments consistent with the disclosure.On the contrary, they be only with it is such as appended
The example of the consistent apparatus and method of some aspects be described in detail in claims, the disclosure.
The front camera of terminal is usually RGB cameras, and the recognition of face component of the front camera based on terminal is
Through being applied in each flagship terminal (such as mobile phone), but from effect, the recognition of face based on front camera
Component, the effect when resisting light change, face prosthese (such as recognition of face is carried out from photo) are poor.The disclosure is implemented
The method for detecting human face that example provides, Face datection is carried out using the 3D cameras based on structure light, can be multiple in half-light, unglazed etc.
The depth image of human body is obtained under heterocycle border, and the depth image got can greatly strengthen from the influence of face dressing
The reliability of Face datection algorithm.The detailed description for the method for detecting human face that the embodiment of the present disclosure provides refer to following implementations
Example.
The method for detecting human face that the embodiment of the present disclosure provides can be performed by terminal, which can be smart mobile phone, put down
Plate computer, smart television, intelligent watch, car-mounted terminal, intelligent wearable device, dynamic image expert's compression standard audio layer
(the English of face 5:Moving Picture Experts Group Audio Layer V;Referred to as:MP5) player, it is on knee just
Take computer or desktop computer etc..
During the method for detecting human face that the embodiment of the present disclosure provides is performed, terminal can also be obtained from server
Related data, determines formula, obtains feature classifiers etc. for example, obtaining size.Certainly, size determines formula and feature classifiers
Can also be stored by terminal itself, and can be configured when terminal is dispatched from the factory or terminal it is true by machine learning algorithm
Fixed.Wherein, server can be a server, or the server cluster being made of some servers, or one
A cloud computing service center.Terminal and server can by cable network or wireless network connection, the cable network for example but
It is not limited to, Universal Serial Bus (English:Universal Serial Bus;Referred to as:USB), wireless network is such as, but not limited to:
Wireless Fidelity (English:Wireless Fidelity;Referred to as:WIFI), bluetooth, infrared, purple honeybee (English:Zigbee), data
Deng.
Fig. 1 is a kind of method flow diagram of method for detecting human face according to an exemplary embodiment, the Face datection
Method can be performed by terminal, and the method for detecting human face can simplify Face datection process, improve detection speed.Referring to figure
1, which includes:
In a step 101, the depth image of human body is captured by the stereo camera shooting component of terminal, depth image includes multiple
Pixel, the depth value of each pixel in multiple pixels represent correspondence position point of each pixel on human body with it is vertical
The distance between body camera assembly.
In a step 102, the depth value of each pixel in multiple pixels, determines that each pixel is corresponding
Window size.
In step 103, according to position of each pixel in depth image and the corresponding window ruler of each pixel
It is very little, determine the corresponding detection window of each pixel.
At step 104, using the corresponding detection window of each pixel, Face datection is carried out in depth image.
In conclusion the method for detecting human face that the embodiment of the present disclosure provides, due to can be according to the depth of each pixel
Value determines the corresponding window size of each pixel, according to position of each pixel in depth image and each pixel pair
The window size answered, determines the corresponding detection window of each pixel, using the corresponding detection window of each pixel, in depth
Face datection is carried out in image, it is therefore not necessary to carry out Face datection using various sizes of detection window in each pixel, is had
Help solve the problem of Face datection process is complicated, and detection speed is slower, simplify Face datection process, improve detection speed.
Fig. 2 is a kind of method flow diagram of the method for detecting human face shown according to another exemplary embodiment, which examines
The method of survey can be performed by terminal, and the method for detecting human face can simplify Face datection process, improve detection speed.Referring to
Fig. 2, the method for detecting human face include:
In step 201, whether detection terminal, which is in, is held state.
In the embodiments of the present disclosure, user, which can grasp terminal, makes terminal in state is held, so that terminal detection is eventually
Whether end that is to say whether detection terminal is grasped by user in being held state.Alternatively, touch can be provided with terminal
Inductive component, the touch sensible component is by detecting whether touch signal is grasped come detection terminal by user, when touch sensible group
When part detects touch signal, terminal determines that terminal is grasped by user, so that it is determined that terminal, which is in, is held state, feels when touching
When answering the component to be not detected by touch signal, terminal determines that terminal is not grasped by user, so that it is determined that terminal is not in being held shape
State.Wherein, touch sensible component can include at least one of touch-screen and touch sensor, and touch sensor can be set
On the shell of terminal, for detecting touch signal.
Alternatively, in the state that is held and when being not in being held state, the posture of terminal is typically different, terminal
In posture inductive component can be set, which can obtain the posture of terminal, and terminal can sense according to posture
The posture for the terminal that component obtains, determines whether terminal is in and is held state.Wherein, posture inductive component can include accelerating
At least one of sensor, gyro sensor and gravity sensor are spent, which obtains the posture of terminal
Process may be referred to correlation technique, and details are not described herein for the embodiment of the present disclosure.
It should be noted that the scheme whether detection terminal that the step 201 is provided is in the state that is held is only
Exemplary, in practical application, other scheme detection terminals can also be used whether to be in and be held state, the embodiment of the present disclosure
Detection terminal is intended to whether in state is held, for being held state using which kind of scheme detection terminal whether be in, this
Open embodiment does not limit.
In step 202, when terminal, which is in, is held state, the stereo camera shooting component of terminal is opened.
If terminal determines that terminal is in and is held state in step 202, terminal can open the stereo camera shooting of terminal
Component.In the embodiments of the present disclosure, which is 3D cameras, and 3D cameras can be the 3D based on structure light
Camera.3D cameras based on structure light can overcome the influence of the complex environment such as half-light, unglazed in capture images, and
From the influence of face dressing, the authenticity of the image of acquisition is improved, so as to improve the reliability of Face datection algorithm.
It should be noted that the embodiment of the present disclosure be by stereo camera shooting component be the 3D cameras based on structure light exemplified by into
Row explanation, in practical application, stereo camera shooting component can also be that camera assembly, the binocular cameras such as binocular camera usually wrap
Two cameras for being in diverse location are included, since the position of two cameras is different, binocular camera can be same
One moment carried out image capture with different view, so as to obtain the image of two different visual angles.
It should also be noted that, when terminal is in and is held state, the embodiment of the present disclosure thinks that user is needed into pedestrian
Face detects, so that terminal opens the stereo camera shooting component of terminal, it will be appreciated by those skilled in the art that, when terminal is not in being held
When holding state, user may be also required to carry out Face datection, so that when terminal is not in being held state, terminal can also be opened
The stereo camera shooting component of terminal is opened, certainly, when terminal is not in being held state, terminal can not also open stereo camera shooting group
Part, the embodiment of the present disclosure are not construed as limiting this.In addition, the step 201 to step 202 for user need carry out Face datection can
Select step, in practical application, terminal can not perform the step 201 to step 202, directly under when carrying out Face datection
Step 203 is stated to start to perform.
In step 203, the depth image of human body is captured by the stereo camera shooting component of terminal, depth image includes multiple
Pixel, the depth value of each pixel in multiple pixels represent correspondence position point of each pixel on human body with it is vertical
The distance between body camera assembly.
Terminal can capture the image of human body by the stereo camera shooting component of terminal, and the image of the human body that is to say human body
Depth image, the depth image can include multiple pixels, and each pixel in multiple pixels has depth value.At this
In open embodiment, each pixel in depth image can be the point in three-dimensional system of coordinate, so that depth image can be adopted
Represented with the point set in three-dimensional coordinate, two reference axis of three-dimensional system of coordinate can be located at depth image institute planar, separately
One reference axis can be vertical with plane where depth image, and the depth value of each pixel can be should be with depth image place
Numerical value in the vertical reference axis of plane, the depth value of each pixel represent correspondence position point of each pixel on human body
The distance between stereo camera shooting component, in other words for, the depth value of each pixel is represented when capturing the depth image,
Distance between each correspondence position point and stereo camera shooting component of the pixel on human body.
Alternatively, when stereo camera shooting component is 3D cameras, directly human body is shot using 3D cameras, so that it may
To obtain the depth image of human body.When stereo camera shooting component is binocular camera, two cameras of binocular camera can be with
Human body is shot with different view in synchronization, obtains the human body image of different visual angles, terminal can be to different visual angles
Human body image synthesized to obtain the depth image of human body.Certainly, can also be adopted according to the difference of stereo camera shooting component, terminal
The depth image of human body is obtained in other ways, and details are not described herein for the embodiment of the present disclosure.
In step 204, the depth value of each pixel in multiple pixels, determines that each pixel is corresponding
Window size.
Wherein, the corresponding window size of each pixel that is to say the size of the corresponding detection window of each pixel.
In the embodiment of the present disclosure, detection window can be positive K sides shape window, and K is the integer more than or equal to 3, and window size can be
The length of side of detection window.Alternatively, detection window can be square window.
Alternatively, the depth value of each pixel in multiple pixels, determines the corresponding window of each pixel
Size can include:According to the depth value of each pixel, determine that formula determines the corresponding window of each pixel using size
Size.Wherein, size determines that formula can be f=ax2+ bx+c, f represent window size, and x represents depth value, and a, b and c are
Constant.The depth value x of each pixel can be substituted into above-mentioned size and determine that formula is calculated each pixel and corresponds to by terminal
Window size.
It should be noted that before determining that formula determines the corresponding window size of each pixel using size, terminal
It can first determine that above-mentioned size determines the value of a, b and c in formula.Alternatively, terminal can be determined using linear regression algorithm
A, the value of b and c.Illustratively, Fig. 3 is the method flow of the value of a kind of definite a, b and c according to an exemplary embodiment
Figure, referring to Fig. 3, this method includes:
In sub-step 2041, window size sample set is obtained, window size sample set includes p window size and and p
A one-to-one p depth value of window size, p are the integer more than or equal to 1.
Alternatively, terminal can determine p sampled pixel point at least one sample facial image, in each sample picture
Vegetarian refreshments carries out Face datection using the detection window of different windows size, and it is every that window size when detecting face is determined as this
The corresponding window size of a sampled pixel point, and according to the depth value and each sampled pixel point pair of each sampled pixel point
The window size answered, determines one group of sample parameter, obtains p group sample parameters, so as to obtain window size sample set.Alternatively, eventually
End can according to it is conventional carry out Face datection when, window size when each pixel of facial image detects face and
The depth value of each pixel, determines one group of sample parameter, obtains p group sample parameters, so as to obtain window size sample set.Or
Person, terminal directly can obtain window size sample set from server, and the embodiment of the present disclosure is not construed as limiting this.
In the embodiments of the present disclosure, every group of sample parameter includes a depth value x and a window corresponding with depth value x
Mouth size f, window size sample set can be:(x1, f1), (x2, f2), (x3, f3) ..., (xp, fp), x1With f1It is corresponding, x2With
f2It is corresponding, x3With f3It is corresponding, and so on, xpWith fpIt is corresponding.
In sub-step 2042, according to p window size and p depth value, the value of a, b and c are determined.
Alternatively, p groups sample parameter can be substituted into formula f=ax by terminal using a, b and c as unknown number2+ bx+c is obtained
To p equation, the value of a, b and c are obtained according to the p equation calculation.Illustratively, which can be as follows:
The above-mentioned p equation of terminal-pair is solved, and can obtain the value of a, b and c.Illustratively, the above-mentioned p side of terminal-pair
Cheng Jinhang is solved, and can obtain a=2.135 × 10-3, b=0.3326, c=32.345.
It should be noted that the embodiment of the present disclosure is so that terminal determines the value of a, b and c using linear regression algorithm as an example
Illustrate, in practical application, terminal can also determine the value of a, b and c using other algorithms.In addition, terminal can be with
Directly the size of value comprising a, b and c is obtained from server determine formula, so that, the value of a, b and c is determined without terminal,
The embodiment of the present disclosure is not construed as limiting this.
It should also be noted that, the embodiment of the present disclosure is to determine that each pixel in multiple pixels is corresponding with terminal
Illustrated exemplified by window size, in practical application, terminal can determine at least one detection pixel from multiple pixels
Point, and the depth value of each detection pixel at least one detection pixel determines that each detection pixel is corresponding
Window size, it can be appreciated that compared to the corresponding window size of each pixel is determined, determines at least one from multiple pixels
A detection pixel, and determine that the corresponding window size of each detection pixel can reduce the calculation amount of definite window size,
Improve Face datection speed.
In step 205, according to position of each pixel in depth image and the corresponding window ruler of each pixel
It is very little, determine the corresponding detection window of each pixel.
, can be according to each pixel in depth image after terminal determines the corresponding window size of each pixel
Position and the corresponding window size of each pixel, determine the corresponding detection window of each pixel, wherein, each pixel exists
Position in depth image can be that the central point position of the corresponding detection window of each pixel (that is to say that pixel can
Think the central point of corresponding detection window).In the embodiments of the present disclosure, detection window can be positive K sides shape window, positive K sides
All diagonal of shape are met at a bit, the central point of positive K shapes when cornerwise intersection point of shape is positive K, therefore detection window
Central point can be detection window cornerwise intersection point.Alternatively, when detection window is square window, detection window
Central point be the square window two cornerwise intersection points.
It should be noted that in practical application, terminal can be according to each detection pixel determined in step 204 in depth
The position in image and the corresponding window size of each detection pixel are spent, determines each to detect the corresponding detection window of pixel
Mouthful, to reduce the calculation amount for determining detection window, improve Face datection speed.
In step 206, using the corresponding detection window of each pixel, Face datection is carried out in depth image.
After terminal determines the corresponding detection window of each pixel, the corresponding detection window of each pixel can be used,
Face datection is carried out in depth image.Alternatively, Fig. 4 is that a kind of terminal according to an exemplary embodiment uses each
The corresponding detection window of pixel carries out the method flow diagram of Face datection in depth image, and referring to Fig. 4, this method includes:
In step 2061, corresponding region of the corresponding detection window of each pixel in depth image is determined as often
The corresponding detection zone of a pixel.
Corresponding region of the corresponding detection window of each pixel in depth image can be determined as each picture by terminal
The corresponding detection zone of vegetarian refreshments.It that is to say, the area that terminal limits the corresponding detection window of each pixel in depth image
Domain is determined as the corresponding detection zone of each pixel.Illustratively, Fig. 5 is a kind of detection according to an exemplary embodiment
The schematic diagram in region, which illustrates so that detection window is square window as an example, referring to Fig. 5, the picture in depth image S
The corresponding detection windows of vegetarian refreshments E are detection window W, and pixel E is located at the central point position of detection window W, detects window
Corresponding regions of the mouth W in depth image S is region G, and region G can be the corresponding detection zones of pixel E, so that, it is right
In the depth image S shown in Fig. 5, region G can be determined as the corresponding detection zones of pixel E by terminal.
In step 2062, from the corresponding detection zone of each pixel, characteristics of image is extracted.
After terminal determines the corresponding detection zone of each pixel, feature extraction algorithm can be used from each pixel pair
Characteristics of image is extracted in the detection zone answered.Illustratively, terminal uses extracts in the region G shown in feature extraction algorithm from Fig. 5
Characteristics of image.Wherein, feature extraction algorithm is such as, but not limited to, Local Binary Pattern (English:local binary
pattern;Referred to as:LBP) algorithm, Scale invariant features transform (English:Scale-invariant feature
transform;Referred to as:SIFT) algorithm, gradient orientation histogram (English:Histograms of Oriented
Gradients;Referred to as:HOG) algorithm and deep neural network learning algorithm etc..Terminal is from feature extraction algorithm is used from image
The process of middle extraction characteristics of image is well known to those skilled in the art, and details are not described herein for the embodiment of the present disclosure.
In step 2063, whether detection image feature is people's face feature.
Terminal can with Detection and Extraction to characteristics of image whether be face characteristic.Wherein, face characteristic can be included but not
It is limited to nose, the nasion, left eye angle, right eye angle, the left corners of the mouth, the right corners of the mouth, chin central point, left ear-lobe, auris dextra hang down, left cheek and the right side
Cheek etc..
Alternatively, whether terminal uses character separation device detection image feature for people's face feature, and this feature separator can be with
For algorithm (English:Adaboost) grader, this feature separator can be that terminal is just configured when dispatching from the factory, can also
It is that terminal is obtained from server, can also be what terminal was trained by machine learning algorithm, the embodiment of the present disclosure is not right
The source of character separation device limits.
Alternatively, Fig. 6 is that one kind according to an exemplary embodiment is characterized in using character separation device detection image
The no method flow diagram for face characteristic, referring to Fig. 6, this method includes:
In step 20631, the feature classifiers that grader generating means are sent are received.
Wherein, feature classifiers can be grader generating means using feature samples collection as training data, according to classification
Algorithm for Training obtain be used for determine characteristics of image whether be face characteristic grader, feature samples concentrate include n face
Feature samples and m non-face feature samples, n and m are the integer more than 0.
Alternatively, grader generating means obtain spy using feature samples collection as training data according to classification algorithm training
Sign grader can include:Grader generating means extract n face feature samples from multiple sample images and m non-face
N face feature samples and m non-face feature samples are merged into feature samples collection by feature samples;With this feature sample set
As training data, feature classifiers are obtained according to classification algorithm training.In the embodiments of the present disclosure, grader generating means can
The desktop computer thinking a processing component in terminal or server or be separately configured, notebook,
Server or server cluster etc..Wherein, sample image is used for grader generating means extraction face characteristic sample and non-face
Feature samples, and the quantity of sample image is more, the formation speed of feature samples collection is faster, and then the generation speed of feature classifiers
Degree also can be faster.
Grader generating means can extract multiple face characteristic samples and multiple non-face feature samples from sample image
This, any face characteristic sample in the plurality of face characteristic sample can include q characteristic parameter and for indicating that this is any
The label of the classification of face characteristic sample, the classification of any face characteristic sample is face characteristic, and q is the integer more than 0, should
Any non-face feature samples in multiple non-face feature samples include q characteristic parameter and for indicating that this is any non-face
The label of the classification of feature samples, the classification of any non-face feature samples is non-face feature;It is special to multiple faces respectively
Levy sample and multiple non-face feature samples carry out format check processing and missing values completion processing, obtain n face characteristic sample
Sheet and m face feature samples.
It should be noted that q characteristic parameter can include:The parameter of nose, the parameter of the nasion, the parameter at left eye angle,
The parameter at right eye angle, the parameter of the left corners of the mouth, the parameter of the right corners of the mouth, the parameter of chin central point, the parameter of left ear-lobe, auris dextra hang down
Parameter of parameter, the parameter of left cheek and right cheek etc., q is bigger, and the accuracy of the feature classifiers of generation is higher.
In step 20632, the characteristics of image input feature vector grader that will extract is special to pass through feature classifiers output
Label is levied, feature tag is used for the classification for indicating characteristics of image, and the classification of characteristics of image includes face characteristic or non-face feature.
The characteristics of image input feature vector grader that terminal will can be extracted, which, which can include multiple features, joins
Number, any face characteristic sample or any non-face spy that feature classifiers obtain this feature parameter and grader generating means
The characteristic parameter of sign sample is compared, and obtains the feature tag of the characteristics of image, and exports the feature tag of the characteristics of image,
This feature label is used for the classification for indicating characteristics of image, and the classification of the characteristics of image indicated according to this feature label can determine to scheme
As whether feature is people's face feature.
In step 20633, according to feature tag, determine whether characteristics of image is face characteristic.
Whether the feature tag that terminal can be exported according to feature classifiers, the characteristics of image for determining to extract are face spy
Sign.In the embodiments of the present disclosure, feature tag is used for the classification for indicating characteristics of image, and the classification of characteristics of image can be face
Feature or non-face feature, when the classification of the characteristics of image for the feature tag instruction that feature classifiers export in sub-step 20632
For face characteristic when, terminal determines that characteristics of image is face characteristic, when in sub-step 20632 feature classifiers export feature mark
When the classification for signing the characteristics of image of instruction is non-face feature, terminal determines that characteristics of image is non-face feature.
It should be noted that in practical application, terminal can use each detection pixel determined in step 205 to correspond to
Detection window, Face datection is carried out in depth image, to improve Face datection speed.In addition, if terminal determines to extract
Characteristics of image be face characteristic, then terminal determines to detect face, and terminal can carry out face knowledge according to the face detected
Not etc., to carry out authentication to user, which is well known to those skilled in the art, and the embodiment of the present disclosure is no longer superfluous herein
State.
It should also be noted that, the sequencing of method for detecting human face step provided in an embodiment of the present invention can be fitted
Work as adjustment, step according to circumstances can also accordingly be increased and decreased, and any one skilled in the art takes off in the present invention
In the technical scope of dew, the method that can readily occur in change should be covered by the protection scope of the present invention, therefore no longer superfluous
State.
In conclusion the method for detecting human face that the embodiment of the present disclosure provides, due to can be according to the depth of each pixel
Value determines the corresponding window size of each pixel, according to position of each pixel in depth image and each pixel pair
The window size answered, determines the corresponding detection window of each pixel, using the corresponding detection window of each pixel, in depth
Face datection is carried out in image, it is therefore not necessary to carry out Face datection using various sizes of detection window in each pixel, is had
Help solve the problem of Face datection process is complicated, and detection speed is slower, simplify Face datection process, improve detection speed.
In traditional Face datection algorithm, due to the corresponding window size of unpredictable pixel, it is therefore desirable in human body
Each pixel position of image carries out Face datection, detection process complexity and speed using various sizes of detection window
It is slower;The method for detecting human face that the embodiment of the present disclosure provides, can use the corresponding window of depth value prediction pixel point of pixel
Mouth size, only needs to carry out Face datection using a detection window, without in pixel position in each pixel
Face datection is carried out using various sizes of detection window, therefore detection process is simple and speed, compared to traditional people
Detection speed can be lifted 10 times or so by face detection algorithm, the method for detecting human face that the embodiment of the present disclosure provides.
Following is embodiment of the present disclosure, can be used for performing embodiments of the present disclosure.It is real for disclosure device
The details not disclosed in example is applied, refers to embodiments of the present disclosure.
Fig. 7 is a kind of block diagram of human face detection device 700 according to an exemplary embodiment, the human face detection device
700 can be implemented in combination with as some or all of of terminal by software, hardware or both.Referring to Fig. 7, face inspection
Surveying device 700 can include:
Trapping module 710, is configured as capturing the depth image of human body, depth image by the stereo camera shooting component of terminal
Including multiple pixels, the depth value of each pixel in multiple pixels represents correspondence position of each pixel on human body
Put a little the distance between with stereo camera shooting component;
First determining module 720, the depth value for each pixel being configured as in multiple pixels, determines each
The corresponding window size of pixel;
Second determining module 730, is configured as according to position of each pixel in depth image and each pixel
Corresponding window size, determines the corresponding detection window of each pixel;
First detection module 740, is configured as using the corresponding detection window of each pixel, is carried out in depth image
Face datection.
In conclusion the human face detection device that the embodiment of the present disclosure provides, due to can be according to the depth of each pixel
Value determines the corresponding window size of each pixel, according to position of each pixel in depth image and each pixel pair
The window size answered, determines the corresponding detection window of each pixel, using the corresponding detection window of each pixel, in depth
Face datection is carried out in image, it is therefore not necessary to carry out Face datection using various sizes of detection window in each pixel, is had
Help solve the problem of Face datection process is complicated, and detection speed is slower, simplify Face datection process, improve detection speed.
Alternatively, detection window is positive K sides shape window, and K is the integer more than or equal to 3, and window size is detection window
The length of side, the first determining module 720, is configured as the depth value according to each pixel, determines that formula determines often using size
The corresponding window size of a pixel;
Wherein, size determines that formula is:F=ax2+ bx+c, f represent window size, and x represents depth value, and a, b and c are
Constant.
Further, Fig. 8 is refer to, another human face detection device 700 provided it illustrates the embodiment of the present disclosure
Block diagram, referring to Fig. 8, on the basis of Fig. 7, which further includes:
Acquisition module 750, is configured as in the depth value according to each pixel, and it is each to determine that formula determines using size
Before the corresponding window size of pixel, obtain window size sample set, window size sample set include p window size and with
The one-to-one p depth value of p window size, p are the integer more than or equal to 1;
3rd determining module 760, is configured as, according to p window size and p depth value, determining the value of a, b and c.
Alternatively, Fig. 9 is refer to, a kind of block diagram of the first detection module 740 provided it illustrates the embodiment of the present disclosure,
Referring to Fig. 9, which includes:
Determination sub-module 741, is configured as the correspondence area in depth image by the corresponding detection window of each pixel
Domain is determined as the corresponding detection zone of each pixel;
Extracting sub-module 742, is configured as from the corresponding detection zone of each pixel, extracts characteristics of image;
Detection sub-module 743, is configured as whether detection image feature is people's face feature.
Alternatively, whether detection sub-module 743, be configured as using character separation device detection image feature for people's face spy
Sign;
Wherein, feature classifiers be grader generating means using feature samples collection as training data, according to sorting algorithm
Training obtain be used for determine characteristics of image whether be face characteristic grader, feature samples concentrate include n face characteristic
Sample and m non-face feature samples, n and m are the integer more than 0.
Alternatively, detection sub-module 743, are configured as:
Receive the feature classifiers that grader generating means are sent;
The characteristics of image input feature vector grader that will be extracted, to pass through feature classifiers output characteristic label, feature mark
Label are used for the classification for indicating characteristics of image, and the classification of characteristics of image includes face characteristic or non-face feature;
According to feature tag, determine whether characteristics of image is face characteristic.
Alternatively, the first determining module 720, is configured as determining at least one detection pixel from multiple pixels;
The depth value of each detection pixel at least one detection pixel, determines the corresponding window of each detection pixel
Size;
Second determining module 730, is configured as according to each position for detecting pixel in depth image and each inspection
The corresponding window size of pixel is surveyed, determines the corresponding detection window of each detection pixel;
First detection module 740, is configured as using the corresponding detection window of each detection pixel, in depth image
Carry out Face datection.
Further, 0 is please referred to Fig.1, another human face detection device 700 provided it illustrates the embodiment of the present disclosure
Block diagram, referring to Figure 10, on the basis of Fig. 8, which further includes:
Second detection module 770, be configured as captured by the stereo camera shooting component of terminal human body depth image it
Before, whether detection terminal, which is in, is held state;
Opening module 780, is configured as terminal in when being held state, opens the stereo camera shooting component of terminal.
Alternatively, stereo camera shooting component is 3D cameras.
Alternatively, 3D cameras are the 3D cameras based on structure light.
In conclusion the human face detection device that the embodiment of the present disclosure provides, due to can be according to the depth of each pixel
Value determines the corresponding window size of each pixel, according to position of each pixel in depth image and each pixel pair
The window size answered, determines the corresponding detection window of each pixel, using the corresponding detection window of each pixel, in depth
Face datection is carried out in image, it is therefore not necessary to carry out Face datection using various sizes of detection window in each pixel, is had
Help solve the problem of Face datection process is complicated, and detection speed is slower, simplify Face datection process, improve detection speed.
On the device in above-described embodiment, wherein modules perform the concrete mode of operation in related this method
Embodiment in be described in detail, explanation will be not set forth in detail herein.
The embodiment of the present disclosure additionally provides a kind of terminal, including:
Processor;
Memory for the executable instruction for storing processor;
Wherein, processor is configured as:
The depth image of human body is captured by the stereo camera shooting component of terminal, depth image includes multiple pixels, multiple
The depth value of each pixel in pixel represents correspondence position point of each pixel on human body and stereo camera shooting component
The distance between;
The depth value of each pixel in multiple pixels, determines the corresponding window size of each pixel;
According to position of each pixel in depth image and the corresponding window size of each pixel, each picture is determined
The corresponding detection window of vegetarian refreshments;
Using the corresponding detection window of each pixel, Face datection is carried out in depth image.
Figure 11 is a kind of structure diagram of terminal 1100 according to an exemplary embodiment.For example, terminal 1100
Can be smart mobile phone, tablet computer, smart television, intelligent watch, car-mounted terminal, intelligent wearable device, MP5 broadcastings
Device, computer, digital broadcast terminal, messaging devices, game console, tablet device, Medical Devices, body-building equipment or a
Personal digital assistant etc..
Referring to Figure 11, terminal 1100 can include following one or more assemblies:Processing component 1102, memory 1104,
Power supply module 1106, multimedia component 1108, audio component 1110, input/output (I/O) interface 1112, sensor component
1114 and communication component 1116.
The integrated operation of the usual control terminal 1100 of processing component 1102, such as with display, call, data communication,
The operation that positioning, camera operation and record operation are associated.Processing component 1102 can include one or more processors 1120
Carry out execute instruction, to complete all or part of step of above-mentioned method for detecting human face.In addition, processing component 1102 can include one
A or multiple modules, easy to the interaction between processing component 1102 and other assemblies.For example, processing component 1102 can include it is more
Media module, to facilitate the interaction between multimedia component 1108 and processing component 1102.
Memory 1104 is configured as storing various types of data to support the operation in terminal 1100.These data
Example include being used for any application for being operated in terminal 1100 or the instruction of method, contact data, telephone book data, disappear
Breath, picture, video etc..Memory 1104 can be by any kind of volatibility or non-volatile memory device or their group
Close and realize, such as static RAM (English:Dynamic Random Access Memory;Referred to as:SRAM), electricity can
Erasable programmable read-only memory (EPROM) (English:Electrically Erasable Programmable Read-Only
Memory;Referred to as:EEPROM), Erasable Programmable Read Only Memory EPROM (English:Erasable Programmable Read
Only Memory;Referred to as:EPROM), programmable read only memory (English:Programmable Read Only Memory;
Referred to as:PROM), read-only storage (English:Read-Only Memory;Referred to as:ROM), magnetic memory, flash memory, magnetic
Disk or CD.
Power supply module 1106 provides electric power for the various assemblies of terminal 1100.Power supply module 1106 can include power management
System, one or more power supplys and other components associated with generating, managing and distributing electric power for terminal 1100.
Multimedia component 1108 is included in the screen that an output interface is provided between terminal 1100 and user.In some realities
Apply in example, screen can include liquid crystal display (English:LiquidCrystalDisplay;Referred to as:) and touch panel LCD
(English:TouchPanle;Referred to as:TP).If screen includes touch panel, screen may be implemented as touch-screen, to receive
Input signal from the user.Touch panel includes one or more touch sensors to sense touch, slip and touch panel
On gesture.Touch sensor can not only sense the boundary of a touch or slide action, but also detect and grasped with touching or sliding
Make relevant duration and pressure.In certain embodiments, multimedia component 1108 include a front camera and/or after
Put camera.When terminal 1100 is in operator scheme, during such as screening-mode or video mode, front camera and/or postposition are taken the photograph
As head can receive the multi-medium data of outside.Each front camera and rear camera can be a fixed optical lens
Mirror system has focusing and optical zoom capabilities.
Audio component 1110 is configured as output and/or input audio signal.For example, audio component 1110 includes a wheat
Gram wind (English:Microphone;Referred to as:MIC), when terminal 1100 is in operator scheme, such as call model, logging mode and language
During sound recognition mode, microphone is configured as receiving external audio signal.The received audio signal can be further stored
Sent in memory 1104 or via communication component 1116.In certain embodiments, audio component 1110 further includes one and raises one's voice
Device, for exports audio signal.
I/O interfaces 1112 provide interface, above-mentioned peripheral interface module between processing component 1102 and peripheral interface module
Can be keyboard, click wheel, button etc..These buttons can include but is not limited to:Home button, volume button, start button
And locking press button.
Sensor component 1114 includes one or more sensors, and the state for providing various aspects for terminal 1100 is commented
Estimate.For example, sensor component 1114 can detect opening/closed mode of terminal 1100, the relative positioning of component, such as group
Part is the display and keypad of terminal 1100, and sensor component 1114 can be with 1,100 1 groups of detection terminal 1100 or terminal
The position of part changes, the existence or non-existence that user contacts with terminal 1100,1100 orientation of terminal or acceleration/deceleration and terminal
1100 temperature change.Sensor component 1114 can include proximity sensor, be configured to connect in no any physics
Presence of nearby objects is detected when tactile.Sensor component 1114 can also include optical sensor, as complementary metal oxide is partly led
Body (English:Complementary Metal Oxide Semiconductor;Referred to as:) or charge coupled cell (English CMOS
Text:Charge-coupled Device;Referred to as:CCD) imaging sensor, for being used in imaging applications.In some implementations
In example, which can also include acceleration transducer, gyro sensor, Magnetic Sensor, pressure sensor
Or temperature sensor.
Communication component 1116 is configured to facilitate the communication of wired or wireless way between terminal 1100 and other equipment.Eventually
End 1100 can access the wireless network based on communication standard, such as WIFI, 2G, 3G or combinations thereof.In an exemplary reality
Apply in example, communication component 1116 receives broadcast singal or the related letter of broadcast from external broadcasting management system via broadcast channel
Breath.In one exemplary embodiment, communication component 1116 further includes near-field communication (English:Near Field
Communication;Referred to as:NFC) module, to promote junction service.For example, radio frequency identification (English can be based in NFC module:
Radio Frequency Identification;Referred to as:RFID) technology, Infrared Data Association's (English:Infrared Data
Association;Referred to as:IrDA) technology, ultra wide band (English:Ultra Wideband;Referred to as:UWB) technology, bluetooth (English
Text:Bluetooth;Referred to as:BT) technology and other technologies are realized.
In the exemplary embodiment, terminal 1100 can be by one or more application application-specific integrated circuit (English:
Application Specific Integrated Circuit;Referred to as:ASIC), digital signal processor (English:
Digital Signal Processing;Referred to as:DSP), digital signal processing appts (English:Digital Signal
Processing Device;Referred to as:DSPD), programmable logic device (English:Programable Logic Device;Letter
Claim:PLD), field programmable gate array (English:Field-Programmable Gate Array;Referred to as:FPGA), control
Device, microcontroller, microprocessor or other electronic components are realized, for performing above-mentioned method for detecting human face.
In the exemplary embodiment, a kind of non-transitorycomputer readable storage medium including instructing, example are additionally provided
Such as include the memory 1104 of instruction, above-metioned instruction can be performed by the processor 1120 of terminal 1100 to complete above-mentioned Face datection
Method.For example, non-transitorycomputer readable storage medium can be ROM, random access memory (English:Random
Access Memory;Referred to as:RAM), laser record read only memory (English:Compact Disk Read-Only
Memory;Referred to as:CD-ROM), tape, floppy disk and optical data storage devices etc..
A kind of non-transitorycomputer readable storage medium, is stored with instruction in the computer-readable recording medium, when this
Instruction on the processing component (such as processor) of terminal 1100 when performing so that the processing component of terminal 1100 is able to carry out
State method for detecting human face.
Those skilled in the art will readily occur to the disclosure its after considering specification and putting into practice invention disclosed herein
Its embodiment.This application is intended to cover any variations, uses, or adaptations of the disclosure, these modifications, purposes or
Person's adaptive change follows the general principle of the disclosure and including the undocumented common knowledge in the art of the disclosure
Or conventional techniques.Description and embodiments are considered only as exemplary, and the true scope and spirit of the disclosure are by following
Claim is pointed out.
It should be appreciated that the present disclosure is not limited to the precise structures that have been described above and shown in the drawings, and
And various modifications and changes may be made without departing from the scope thereof.The scope of the present disclosure is only limited by appended claim.
Claims (18)
- A kind of 1. method for detecting human face, it is characterised in that the described method includes:The depth image of human body is captured by the stereo camera shooting component of terminal, the depth image includes multiple pixels, described The depth value of each pixel in multiple pixels represent each correspondence position point of the pixel on human body with it is described The distance between stereo camera shooting component;The depth value of each pixel in the multiple pixel, determines the corresponding window ruler of each pixel It is very little;According to position of each pixel in the depth image and the corresponding window size of each pixel, really The corresponding detection window of fixed each pixel;Using the corresponding detection window of each pixel, Face datection is carried out in the depth image.
- 2. according to the method described in claim 1, it is characterized in that, the detection window is positive K sides shape window, the K is big In or equal to 3 integer, the window size be the detection window the length of side, it is described every in the multiple pixel The depth value of a pixel, determines the corresponding window size of each pixel, including:According to the depth value of each pixel, determine that formula determines the corresponding window ruler of each pixel using size It is very little;Wherein, the size determines that formula is f=ax2+ bx+c, the f represent the window size, and the x represents the depth Value, a, the b and the c are constant.
- 3. according to the method described in claim 2, it is characterized in that, the method further includes:In the depth value according to each pixel, determine that formula determines the corresponding window of each pixel using size Before size, obtain window size sample set, the window size sample set include p window size and with the p window The one-to-one p depth value of size, the p are the integer more than or equal to 1;According to the p window size and the p depth value, the value of a, the b and the c are determined.
- 4. method according to any one of claims 1 to 3, it is characterised in that described to be corresponded to using each pixel Detection window, Face datection is carried out in the depth image, including:The each corresponding region of the corresponding detection window of pixel in the depth image is determined as each picture The corresponding detection zone of vegetarian refreshments;From the corresponding detection zone of each pixel, characteristics of image is extracted;Detect whether described image feature is people's face feature.
- 5. according to the method described in claim 4, it is characterized in that, it is described detection described image feature whether be people's face feature, Including:Character separation device is used whether to detect described image feature for people's face feature;Wherein, the feature classifiers be grader generating means using feature samples collection as training data, according to sorting algorithm What training obtained be used to determine characteristics of image whether be face characteristic grader, the feature samples, which are concentrated, includes n face Feature samples and m non-face feature samples, the n and the m are the integer more than 0.
- 6. according to the method described in claim 5, it is characterized in that, described be characterized in using character separation device detection described image No is face characteristic, including:Receive the feature classifiers that the grader generating means are sent;The described image feature extracted is inputted into the feature classifiers, to pass through the feature classifiers output characteristic mark Label, the feature tag are used for the classification for indicating described image feature, and the classification of described image feature includes face characteristic or non- Face characteristic;According to the feature tag, determine whether described image feature is people's face feature.
- 7. method according to any one of claims 1 to 3, it is characterised in thatThe depth value of each pixel in the multiple pixel, determines the corresponding window of each pixel Size, including:At least one detection pixel is determined from the multiple pixel, according at least one detection pixel In each detection pixel depth value, determine the corresponding window size of each detection pixel;The position and the corresponding window ruler of each pixel according to each pixel in the depth image It is very little, determine the corresponding detection window of each pixel, including:According to each detection pixel in the depth image In position and the corresponding window size of each detection pixel, determine the corresponding detection window of each detection pixel Mouthful;It is described to use the corresponding detection window of each pixel, Face datection is carried out in the depth image, including:Adopt With the corresponding detection window of each detection pixel, Face datection is carried out in the depth image.
- 8. method according to any one of claims 1 to 3, it is characterised in that the method further includes:Before the depth image of human body is captured by the stereo camera shooting component of terminal, whether the terminal is detected in being held State;When the terminal is held state described in, the stereo camera shooting component of the terminal is opened.
- 9. a kind of human face detection device, it is characterised in that described device includes:Trapping module, is configured as capturing the depth image of human body, the depth image bag by the stereo camera shooting component of terminal Include multiple pixels, the depth value of each pixel in the multiple pixel represents each pixel on human body The distance between correspondence position point and the stereo camera shooting component;First determining module, the depth value for each pixel being configured as in the multiple pixel, determines described every The corresponding window size of a pixel;Second determining module, is configured as according to each position and each picture of the pixel in the depth image The corresponding window size of vegetarian refreshments, determines the corresponding detection window of each pixel;First detection module, is configured as using each corresponding detection window of pixel, in the depth image into Row Face datection.
- 10. device according to claim 9, it is characterised in that the detection window is positive K sides shape window, and the K is big In or equal to 3 integer, the window size be the detection window the length of side,First determining module, is configured as the depth value according to each pixel, determines that formula determines using size The corresponding window size of each pixel;Wherein, the size determines that formula is f=ax2+ bx+c, the f represent the window size, and the x represents the depth Value, a, the b and the c are constant.
- 11. device according to claim 10, it is characterised in that described device further includes:Acquisition module, is configured as in the depth value according to each pixel, and it is described every to determine that formula determines using size Before the corresponding window size of a pixel, window size sample set is obtained, the window size sample set includes p window ruler It is very little and be integer more than or equal to 1 with the one-to-one p depth value of the p window size, the p;3rd determining module, is configured as according to the p window size and the p depth value, determine a, the b and The value of the c.
- 12. according to claim 9 to 11 any one of them device, it is characterised in that the first detection module, including:Determination sub-module, is configured as each correspondence area of the corresponding detection window of pixel in the depth image Domain is determined as the corresponding detection zone of each pixel;Extracting sub-module, is configured as from the corresponding detection zone of each pixel, extracts characteristics of image;Detection sub-module, is configured as whether detection described image feature is people's face feature.
- 13. device according to claim 12, it is characterised in thatThe detection sub-module, is configured as using character separation device whether to detect described image feature for people's face feature;Wherein, the feature classifiers be grader generating means using feature samples collection as training data, according to sorting algorithm What training obtained be used to determine characteristics of image whether be face characteristic grader, the feature samples, which are concentrated, includes n face Feature samples and m non-face feature samples, the n and the m are the integer more than 0.
- 14. device according to claim 13, it is characterised in thatThe detection sub-module, is configured as:Receive the feature classifiers that the grader generating means are sent;The described image feature extracted is inputted into the feature classifiers, to pass through the feature classifiers output characteristic mark Label, the feature tag are used for the classification for indicating described image feature, and the classification of described image feature includes face characteristic or non- Face characteristic;According to the feature tag, determine whether described image feature is people's face feature.
- 15. according to claim 9 to 11 any one of them device, it is characterised in thatFirst determining module, is configured as determining at least one detection pixel from the multiple pixel, according to institute The depth value of each detection pixel at least one detection pixel is stated, determines the corresponding window of each detection pixel Mouth size;Second determining module, is configured as position and institute in the depth image according to each detection pixel The corresponding window size of each detection pixel is stated, determines the corresponding detection window of each detection pixel;The first detection module, is configured as using the corresponding detection window of each detection pixel, in the depth Face datection is carried out in image.
- 16. according to claim 9 to 11 any one of them device, it is characterised in that described device further includes:Second detection module, is configured as before the depth image of human body is captured by the stereo camera shooting component of terminal, detection Whether the terminal, which is in, is held state;Opening module, is configured as the terminal and is in described when being held state, open the stereo camera shooting group of the terminal Part.
- A kind of 17. terminal, it is characterised in that including:Processor;Memory for the executable instruction for storing the processor;Wherein, the processor is configured as:The depth image of human body is captured by the stereo camera shooting component of terminal, the depth image includes multiple pixels, described The depth value of each pixel in multiple pixels represent each correspondence position point of the pixel on human body with it is described The distance between stereo camera shooting component;The depth value of each pixel in the multiple pixel, determines the corresponding window ruler of each pixel It is very little;According to position of each pixel in the depth image and the corresponding window size of each pixel, really The corresponding detection window of fixed each pixel;Using the corresponding detection window of each pixel, Face datection is carried out in the depth image.
- A kind of 18. computer-readable recording medium, it is characterised in that instruction is stored with the computer-readable recording medium, When described instruction is being run on the processing component of terminal so that the processing component perform claim is required described in 1 to 8 any one Method for detecting human face.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711305306.8A CN107977636B (en) | 2017-12-11 | 2017-12-11 | Face detection method and device, terminal and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711305306.8A CN107977636B (en) | 2017-12-11 | 2017-12-11 | Face detection method and device, terminal and storage medium |
Publications (2)
Publication Number | Publication Date |
---|---|
CN107977636A true CN107977636A (en) | 2018-05-01 |
CN107977636B CN107977636B (en) | 2021-11-30 |
Family
ID=62009839
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201711305306.8A Active CN107977636B (en) | 2017-12-11 | 2017-12-11 | Face detection method and device, terminal and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107977636B (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108960097A (en) * | 2018-06-22 | 2018-12-07 | 维沃移动通信有限公司 | A kind of method and device obtaining face depth information |
CN109710371A (en) * | 2019-02-20 | 2019-05-03 | 北京旷视科技有限公司 | Font adjusting method, apparatus and system |
CN110046552A (en) * | 2019-03-21 | 2019-07-23 | 南京华捷艾米软件科技有限公司 | Protect the method for user's eyesight and the device of protection user's eyesight |
CN113792718A (en) * | 2021-11-18 | 2021-12-14 | 北京的卢深视科技有限公司 | Method for positioning face area in depth map, electronic device and storage medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100158387A1 (en) * | 2008-12-22 | 2010-06-24 | Electronics And Telecommunications Research Institute | System and method for real-time face detection using stereo vision |
US20110058032A1 (en) * | 2009-09-07 | 2011-03-10 | Samsung Electronics Co., Ltd. | Apparatus and method for detecting face |
CN106250850A (en) * | 2016-07-29 | 2016-12-21 | 深圳市优必选科技有限公司 | Face detection tracking method and device, and robot head rotation control method and system |
CN106886216A (en) * | 2017-01-16 | 2017-06-23 | 深圳前海勇艺达机器人有限公司 | Robot automatic tracking method and system based on RGBD Face datections |
CN106919251A (en) * | 2017-01-09 | 2017-07-04 | 重庆邮电大学 | A kind of collaborative virtual learning environment natural interactive method based on multi-modal emotion recognition |
-
2017
- 2017-12-11 CN CN201711305306.8A patent/CN107977636B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100158387A1 (en) * | 2008-12-22 | 2010-06-24 | Electronics And Telecommunications Research Institute | System and method for real-time face detection using stereo vision |
US20110058032A1 (en) * | 2009-09-07 | 2011-03-10 | Samsung Electronics Co., Ltd. | Apparatus and method for detecting face |
CN106250850A (en) * | 2016-07-29 | 2016-12-21 | 深圳市优必选科技有限公司 | Face detection tracking method and device, and robot head rotation control method and system |
CN106919251A (en) * | 2017-01-09 | 2017-07-04 | 重庆邮电大学 | A kind of collaborative virtual learning environment natural interactive method based on multi-modal emotion recognition |
CN106886216A (en) * | 2017-01-16 | 2017-06-23 | 深圳前海勇艺达机器人有限公司 | Robot automatic tracking method and system based on RGBD Face datections |
Non-Patent Citations (1)
Title |
---|
周羽: ""基于深度相机的三维人脸识别系统"", 《万方数据》 * |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108960097A (en) * | 2018-06-22 | 2018-12-07 | 维沃移动通信有限公司 | A kind of method and device obtaining face depth information |
CN109710371A (en) * | 2019-02-20 | 2019-05-03 | 北京旷视科技有限公司 | Font adjusting method, apparatus and system |
CN110046552A (en) * | 2019-03-21 | 2019-07-23 | 南京华捷艾米软件科技有限公司 | Protect the method for user's eyesight and the device of protection user's eyesight |
CN113792718A (en) * | 2021-11-18 | 2021-12-14 | 北京的卢深视科技有限公司 | Method for positioning face area in depth map, electronic device and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN107977636B (en) | 2021-11-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109034102B (en) | Face living body detection method, device, equipment and storage medium | |
CN108960209B (en) | Identity recognition method, identity recognition device and computer readable storage medium | |
CN106295566B (en) | Facial expression recognizing method and device | |
CN108594997B (en) | Gesture skeleton construction method, device, equipment and storage medium | |
WO2017181769A1 (en) | Facial recognition method, apparatus and system, device, and storage medium | |
JP2020523665A (en) | Biological detection method and device, electronic device, and storage medium | |
CN105608425B (en) | The method and device of classification storage is carried out to photo | |
CN111242090B (en) | Human face recognition method, device, equipment and medium based on artificial intelligence | |
WO2019024717A1 (en) | Anti-counterfeiting processing method and related product | |
CN105654039B (en) | The method and apparatus of image procossing | |
CN112036331B (en) | Living body detection model training method, device, equipment and storage medium | |
CN110807361A (en) | Human body recognition method and device, computer equipment and storage medium | |
CN105224924A (en) | Living body faces recognition methods and device | |
US9536132B2 (en) | Facilitating image capture and image review by visually impaired users | |
CN105678242B (en) | Focusing method and device under hand-held certificate mode | |
EP3985553A1 (en) | Fingerprint anti-counterfeiting method, and electronic device | |
CN107766403B (en) | Photo album processing method, mobile terminal and computer readable storage medium | |
US9047504B1 (en) | Combined cues for face detection in computing devices | |
Vazquez-Fernandez et al. | Built-in face recognition for smart photo sharing in mobile devices | |
CN107977636A (en) | Method for detecting human face and device, terminal, storage medium | |
CN108198159A (en) | A kind of image processing method, mobile terminal and computer readable storage medium | |
CN106295515A (en) | Determine the method and device of human face region in image | |
CN105574512A (en) | Method and device for processing image | |
CN108197585A (en) | Recognition algorithms and device | |
WO2014088125A1 (en) | Image photographing device and method for same |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |