TWI701607B - Dynamic face recognition system and method thereof - Google Patents
Dynamic face recognition system and method thereof Download PDFInfo
- Publication number
- TWI701607B TWI701607B TW108119853A TW108119853A TWI701607B TW I701607 B TWI701607 B TW I701607B TW 108119853 A TW108119853 A TW 108119853A TW 108119853 A TW108119853 A TW 108119853A TW I701607 B TWI701607 B TW I701607B
- Authority
- TW
- Taiwan
- Prior art keywords
- facial
- motion
- user
- file
- images
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
- G06V40/176—Dynamic expression
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
- G06V40/167—Detection; Localisation; Normalisation using comparisons between temporally consecutive images
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Image Analysis (AREA)
Abstract
Description
本發明關於一種動態臉部識別,尤指一種用於動態臉部識別的使用者臉部動作檔案之設定。 The present invention relates to a dynamic face recognition, in particular to a setting of a user's facial motion file used for dynamic face recognition.
由於其廣泛的應用,臉部識別已成為身份識別的一種常規方案。業界提了多種新方法以提高臉部識別率。現有先前技術的臉部識別純粹是基於靜態圖像。一些系統可能聲稱它們是動態的,亦即使用者在移動中也可辨識,但其操作卻完全是基於即時拍攝的靜態圖像,本質上來說,這些所謂的動態是別還是屬於靜態的範疇。 Due to its wide range of applications, face recognition has become a conventional solution for identity recognition. The industry has proposed a variety of new methods to improve the facial recognition rate. The prior art facial recognition is purely based on static images. Some systems may claim that they are dynamic, that is, the user can recognize them while they are moving, but their operations are entirely based on real-time captured static images. In essence, these so-called dynamics are still in the static category.
根據一實施例,描述了一基於臉部動作的臉部識別系統。該系統包含一資料庫、一影像擷取裝置、及一處理單元。該資料庫用以儲存基於一使用者作出之一預設動作的一基準檔案且包含至少一臉部特徵點。該影像擷取裝置追蹤該使用者的臉部動作並拍攝一近似視頻片段,其中該使用者在一時幀內做出一近似臉部動作及至少一臉部表情。該處理單元與該資料庫及該影像擷取裝 置相連接。該處理單元用於設立該基準檔案及該近似臉部動作的一近似檔案。該處理單元進一步將該近似檔案與該基準檔案進行對比以驗證該近似臉部動作。 According to an embodiment, a facial recognition system based on facial motion is described. The system includes a database, an image capture device, and a processing unit. The database is used to store a reference file based on a predetermined action made by a user and includes at least one facial feature point. The image capturing device tracks the facial movements of the user and shoots an approximate video segment, wherein the user performs an approximate facial motion and at least one facial expression in a time frame. The processing unit and the database and the image capture device Set phase connection. The processing unit is used to set up the reference file and an approximate file of the approximate facial motion. The processing unit further compares the approximate file with the reference file to verify the approximate facial motion.
根據一實施例,描述了一種基於由具有至少一臉部特徵點的使用者通過作出臉部動作以設定臉部動作檔案的方法。此方法包含:通過一檔案設定進程設定包含一預設臉部動作的一基準檔案,在一時幀內拍攝一近似視頻片段,並在該時幀內,由使用者作出一近似臉部動作及至少一臉部表情;通過檔案設定進程設定近似臉部動作檔案;並將該近似臉部動作檔案與該基準檔案進行比較,以驗證該近似臉部動作檔案。 According to an embodiment, a method for setting a facial motion file based on a facial motion performed by a user with at least one facial feature point is described. This method includes: setting a reference file containing a preset facial motion through a file setting process, shooting an approximate video clip in a time frame, and in the time frame, the user makes an approximate facial motion and at least A facial expression; the approximate facial motion file is set through the file setting process; the approximate facial motion file is compared with the reference file to verify the approximate facial motion file.
根據一實施例,一動態臉部識別的方法如下。該方法包含設立一使用者的一臉部動作的一檔案以進行識別。該檔案設立流程進一步包括:提取該使用者的至少一臉部特徵點;在一時幀上追蹤並拍攝一視頻片段,其中該視頻片包含該使用者臉部動作的複數個主要臉部特徵,及使用者在該臉部動作之前,之後,或之前及之後作出至少一臉部表情的複數個附屬臉部特徵;在每個主要和附屬臉部特徵上標注臉部特徵點;並基於主要和附屬臉部特徵來設立該臉部動作的檔案。 According to an embodiment, a method of dynamic face recognition is as follows. The method includes creating a file of a facial movement of a user for recognition. The file establishment process further includes: extracting at least one facial feature point of the user; tracking and shooting a video clip on a time frame, wherein the video clip contains a plurality of main facial features of the user's facial movements, and Before, after, or before and after the facial action, the user makes at least one facial expression with a plurality of subsidiary facial features; mark the facial feature points on each main and subsidiary facial features; and based on the main and subsidiary facial features Facial features to set up a profile of the facial movement.
有關本發明之前述及其他技術內容、特點與功效,在以下配合參考圖式之實施例的詳細說明中,將可清楚的呈現。 The foregoing and other technical content, features, and effects of the present invention will be clearly presented in the following detailed description of the embodiments with reference to the drawings.
100:識別系統 100: identification system
110:影像擷取裝置 110: Image capture device
120:資料庫 120: database
130:處理單元 130: processing unit
300:視頻 300: Video
310:預設動態臉部圖像 310: Preset dynamic facial images
320:預設空白臉部圖像 320: preset blank face image
S801~S806、S901~S904:步驟 S801~S806, S901~S904: steps
第1圖為從一使用者臉部提取68個臉部特徵點的示例;第2圖示例了一種與本發明相關的基於臉部動作的臉部識別系統; 第3圖示例了在一時幀的一視頻片段包含多個動態臉部特徵及面無表情特徵;第4圖示例了20個臉部特徵點在時幀t1,t16,ta,tb,tN-15及tN的變化;第5圖示例了臉部特徵點f49在時幀t1,t16,ta,tb,tN-15及tN的相對位置;第6圖示例了臉部特徵點f49的一柱狀圖;第7圖示例了一似微笑之臉部特徵及其他臉部表情;第8圖示例了一基於本發明的動態臉部識別的流程圖;以及第9圖示例了一基於本發明的用於臉部識別的臉部動作檔案設立流程圖。 Figure 1 is an example of extracting 68 facial feature points from a user’s face; Figure 2 illustrates a facial recognition system based on facial motions related to the present invention; Figure 3 illustrates a moment A video segment of the frame contains multiple dynamic facial features and facial expressionless features; Figure 4 illustrates 20 facial feature points in time frames t 1 , t 16 , t a , t b , t N-15 and The change of t N ; Figure 5 illustrates the relative positions of the facial feature point f 49 in time frames t 1 , t 16 , t a , t b , t N-15 and t N ; Figure 6 illustrates the face A histogram of the facial feature point f 49 ; Figure 7 illustrates a smile-like facial feature and other facial expressions; Figure 8 illustrates a flow chart of dynamic facial recognition based on the present invention; and Figure 9 illustrates a flow chart of setting up a facial motion file for facial recognition based on the present invention.
以下實施例中所提到的方向用語,例如:上、下、左、右、前或後等,是參考附加圖式的方向。因此,使用的方向用語是用來說明並非用來限制本發明。 The directional terms mentioned in the following embodiments, for example: up, down, left, right, front or back, etc., refer to the directions of the attached drawings. Therefore, the directional terms used are used to illustrate but not to limit the present invention.
本申請於2018年5月10日提交台灣專利申請案號107115835(“835申請案”)相關,本申請案將835申請案的內容合併於此。根據835申請案的揭露,假設一使用者預先錄製一臉部動作(例如:微笑)綁定動作(例如,開啟車輛的雨刮器)的金鑰。根據835申請案,車內的一影像擷取裝置連續追蹤該使用者的臉部動作以偵測此使用者是否微笑以執行該特定動作。但假設在一情景中,該使用者在開車過程中與他人交談,該影像擷取裝置可能拍攝到該使用者的臉部圖像(facial patterns)正好與微笑相同。在此情況下,由於識別成功,雨刮器可能在未得到進一步確認的情況下便開始運行。這種突然的運行可分散司機的注意力且造成危險。 This application is related to Taiwan Patent Application No. 107115835 ("835 application") filed on May 10, 2018, and this application incorporates the content of the 835 application here. According to the disclosure of the 835 application, it is assumed that a user pre-records a key to a facial motion (for example, a smile) binding motion (for example, turning on a wiper of a vehicle). According to the 835 application, an image capture device in the car continuously tracks the user's facial motion to detect whether the user is smiling to perform the specific action. However, assuming that in a scenario, the user is talking with others while driving, the image capture device may capture the user's facial patterns which are exactly the same as the smile. In this case, due to the successful recognition, the wiper may start to operate without further confirmation. This sudden movement can distract the driver and cause danger.
儘管傳統的臉部識別系統已經進行了許多創新以減少誤識別,然而當涉及到動態臉部識別領域時,它們都不適用。主要的原因是在傳統靜態的臉 部識別中,所有因素和參數都是恒定的,所以相對容易除錯。相反地,在處理動態臉部識別時,所有臉部特徵的變化都需要被考慮,這些都複雜了動態臉部識別的計算和處理。 Although traditional facial recognition systems have made many innovations to reduce misrecognition, they are not applicable when it comes to the field of dynamic facial recognition. The main reason is in the traditional static face In part recognition, all factors and parameters are constant, so it is relatively easy to debug. On the contrary, when dealing with dynamic face recognition, all facial feature changes need to be considered, which complicates the calculation and processing of dynamic face recognition.
本發明主要是針對835申請案的再發明。本發明披露了一種通過設立臉部動作檔案(Profile)以進行動態臉部識別的系統和方法。採用本發明不僅可以提高動態臉部識別的準確性,並消除動作誤判的可能。 The present invention is mainly a re-invention of the 835 application. The invention discloses a system and method for dynamic facial recognition by setting up a facial action profile (Profile). The invention can not only improve the accuracy of dynamic face recognition, but also eliminate the possibility of misjudgment of actions.
臉部識別的運作始於收集一使用者的臉部特徵(facial features)。對於專業人士而言有多種已知方法可以用來選擇臉部特徵、特徵點(landmarks)和/或關鍵點(key points)。在一實施例中,一臉部識別系統認出了一主要使用者(primary user)臉部的多個獨特的臉部特徵/臉部特徵點。如第1圖所示,在一實施例中,一使用者的臉部影像(facial image)過濾出了68個獨特的特徵點(統稱“標注資料(label data)”),標注資料經過如類神經網路等習知技術的訓練可獲得一訓練模型(trained model)。標注資料及訓練模型儲存於資料庫供之後使用。後續識別中,系統將嘗試提取待驗證使用者的68個臉部特徵點並輸入該訓練模型中。如待驗證使用者的臉部特徵點與資料庫儲存的資料匹配,則系統認證驗證使用者確實為主要使用者。對系統而言,多個使用者的標注資料都可以儲存在資料庫中作為驗證。此外,上述例子擷取68個特徵點僅是一個例子,系統可能擷取更多或較少的特徵點,但亦可以達到與本申請案相同的目地,因此數目68不應作為對本發明的限制。
The operation of facial recognition begins with the collection of facial features of a user. There are many known methods for professionals to select facial features, landmarks and/or key points. In one embodiment, a facial recognition system recognizes a plurality of unique facial features/facial feature points of a primary user's face. As shown in Figure 1, in one embodiment, a user’s facial image (facial image) filters out 68 unique feature points (collectively referred to as "label data"). Training of conventional technologies such as neural networks can obtain a trained model. The annotation data and training model are stored in the database for later use. In the subsequent recognition, the system will try to extract 68 facial feature points of the user to be verified and input them into the training model. If the facial feature points of the user to be verified match the data stored in the database, the system verifies that the user is indeed the primary user. For the system, the annotation data of multiple users can be stored in the database for verification. In addition, the above example extracting 68 feature points is only an example. The system may extract more or less feature points, but it can also achieve the same purpose as the application. Therefore, the
另外值得注意的是,在臉部識別的過程中並不一定會考慮全部68個臉部特徵點。在某些情況下,系統可能只考慮它認為相關的臉部特徵點並忽略那些不相關的臉部特徵點。儘管精確度可能會受影響,但這種方法可以縮短計算處理時間。 It is also worth noting that all 68 facial feature points are not necessarily considered in the process of face recognition. In some cases, the system may only consider facial feature points that it considers relevant and ignore those that are not relevant. Although accuracy may be affected, this method can shorten the calculation processing time.
基於動態臉部動作驗證使用者身份的過程已於835申請案中詳細的討論,本發明基於835申請案近一步揭露減少動態臉部動作誤驗證的方法及系統。 The process of verifying the user's identity based on dynamic facial movements has been discussed in detail in the 835 application. Based on the 835 application, the present invention further discloses a method and system for reducing false verification of dynamic facial movements.
本發明的動態臉部識別系統100的運作是通過建立臉部動作檔案(facial motion profiles)開始。識別系統100的結構如第2圖所示。本發明的識別系統100包括一資料庫120,存儲一使用者的至少一個臉部特徵點以及使用者的基準檔案(predetermined profile)。其中,基準檔案是由使用者作出一個預設的臉部動作(default facial motion)而建立。本發明的識別系統100還包括一影像擷取裝置110,用以追蹤使用者的臉部變化並拍攝一時幀的視頻,以及處理單元130。其中處理單元130的作用包含建立臉部動作檔案,以及比較近似檔案(resembling profile)與基準檔案,最後並判斷該近似檔案是否與基準檔案相似。識別系統100還可以配備各種類型的記憶體(圖中未示出)以存儲由影像擷取裝置110拍攝的圖像及視頻,以及處理單元130所處理的資料。第3圖為本發明其中一實施例。如第3圖所示,影像擷取裝置110在一段時間T內擷取到一段使用者微笑及面無表情的視頻。以下說明請一併參考第2圖和第3圖以了解臉部識別系統100的操作。
The operation of the dynamic
在一實施例中,系統100首先就一主要使用者作出一預設臉部動作建立一基準檔案。這個基準檔案作為之後動態臉部識別的參照。如835申請案已經提及的,預設臉部動作可與一功能綁定,當使用者作出預設臉部動作時,該綁並功能將被啟動。綁定功能可能與汽車功能相關,例如打開雨刷等等這種以往需要由駕駛手動控制的功能。建立基準檔案的步驟如下。首先,影像擷取裝置100擷取主要使用者的臉部動作並記錄成一個視頻300,視頻的時間總長度為T。其中視頻300至少包含一預設動態臉部圖像310(例如:當主要使用者微笑的臉部圖像)以及一預設空白臉部圖像320(例如:主要使用者面無表情的臉部圖像)。
其中視頻時間總長度T可被分割為多個單位時間Tx={t1,t2,...,tN},每一個單位時間點Tx皆對應一個臉部圖像(facial pattern)。需要注意的是,臉部圖像的數目不一定等於幀數或與視頻取樣速度正相關。基於一實施例,一臉部圖像可為多幀的組合(例如:10幀),而一單位時間便等同於多幀組合的時間總長(例如:10幀的時間總長)。
In one embodiment, the
假設視頻300共包含N個臉部圖像Mx={M1,M2,...,MN},其中部分為動態臉部圖像310,部分為空白面度圖像320。此外,各臉部圖像表現了相關單位時間點時瞬間的臉部快照。例如,一第一臉部圖像M1為使用者於第一單位時間點t1的快照,而一第二臉部圖像M2為使用者於第二時間點t2的快照等。在一實施例中,假設一預設臉部動作為微笑,如第3圖所示,從視頻300的角度上來看,主要使用者于第16個單位時間點t16(亦即對應第16個臉部圖像M16)開始微笑;而空白臉部圖像則出現於視頻300的前後各15個單位時間點中(亦即對應第1個臉部圖像M1至第15個臉部圖像M15,以及第N-14個臉部圖像MN-14到第N個臉部圖像MN)。在此實施例中假定空白臉部表情320出現於動態臉部圖像310的兩側,但本發明不局限於此。主要使用者可決定在微笑之前或之後作出面無表情的動作。換句話說,空白臉部圖像320可出現在動態臉部圖像310之前,之後,或之前與之後;以上各種方式皆可達到本發明的目的。 It is assumed that the video 300 contains a total of N facial images Mx={M 1 , M 2 ,..., M N }, part of which is a dynamic facial image 310 and a part is a blank face image 320. In addition, each facial image expresses a facial snapshot at an instant in the relevant unit time point. For example, a first facial image M 1 is a snapshot of the user at a first unit time point t 1 , and a second facial image M 2 is a snapshot of the user at a second time point t 2 . In one embodiment, it is assumed that a preset facial action is a smile. As shown in Figure 3, from the perspective of the video 300, the main user at the 16th unit time point t 16 (that is, corresponding to the 16th The facial image M 16 ) starts to smile; and the blank facial image appears in each of the 15 unit time points before and after the video 300 (that is, corresponding to the first facial image M 1 to the 15th facial image like M 15, N-14 and a second face images M N-14 to N-th face image M N). In this embodiment, it is assumed that blank facial expressions 320 appear on both sides of the dynamic facial image 310, but the invention is not limited to this. The primary user can decide to make a blank gesture before or after smiling. In other words, the blank facial image 320 can appear before, after, or before and after the dynamic facial image 310; all the above methods can achieve the purpose of the present invention.
當視頻300的記錄完成後,68個臉部特徵點(即“標注資料”)可被標注於每一個動態及空白臉部圖像Mx={M1,M2,...,MN}上,且每個面部特徵點隨時間T的變化量(variation)亦可得到。值得一提的是,在實施例使用微笑作為預設的臉部動作,因此系統100可以只提取使主要使用者嘴部周邊的臉部特徵點而不必提取所有的68個臉部各徵點。例如,本發明的系統100在此情境中可以只提取嘴部周邊的20個臉部特徵點作為計算參考。如第1圖所示,假設這20個相關的臉
部特徵點為f49,f50,...,及f68,表示為Fx={f49,f50,...,f68}。
When the recording of the video 300 is completed, 68 facial feature points (ie "labeled data") can be labeled on each dynamic and blank facial image Mx={M 1 ,M 2 ,...,M N } Above, and the variation of each facial feature point over time T can also be obtained. It is worth mentioning that in the embodiment, a smile is used as the preset facial action, so the
第4圖示範說明了本系統100參考這20個臉部特徵點在相關時間點的臉部圖像變化,尤其是這20個臉部特徵分別在單位時間點t1,t16,ta,tb,tN-15及tN上對應臉部圖像M1,M16,Ma,Mb,MN-15,及MN的變化。如之前提及的,微笑開始於第16個單位時間點且持續至第(N-15)個單位時間點(即t16-tN-15),其中空白臉部表情出現於前15個和最後15個單位時間點(t1-t15及tN-14-tN),其中M1及MN為空白臉部圖像,而M16,Ma,Mb,MN-15為動態臉部圖像。
Figure 4 demonstrates that the
本發明的臉部識別系統100自每一單位時間點針對每一個特徵點fx(fx為f49,f50,...,或f68)計算一量值-Res_fx(t)來表示在單位時間點上的該特徵點的量化值。該量值可以是在一單位時間時特徵點的角度,距離或任一角度和距離所積累的位置變化。量值Res_fx(t)的計算可以通過數學變換(conversion)計算得到。例如,量值可以透過轉換一特徵點的座標、角度、距離、位置或相對位置等等計算得到。在一實施例中,量值Res_fx(t)被定義為一特徵點在一單位時間基於一預設錨點(如鼻尖)的相對位置。另外,量值亦可以是一特徵點在某一單位時間點前所有位置的累計積分值。例如,通過積分一特徵點自t1至t3的位置軌跡可以得到Res_fx(t3)。第5圖說明了特徵點f49在單位時間點t1,t16,ta,tb,tN-15及tN與一預設錨點(例如:鼻尖)的相對位置變化。在另一實施例中,量值的計算會將使用者臉部的寬度考慮進去用以消除比例的影響。利用上述方式,任何一個特徵值在任何一單位時間點對應的量值Res_f49(t)(t為t1,...,或tN)皆可得到。除了上述方法外,習知技術的人亦可用其他的方式得到單位時間點上特徵值的量化值,本發明對如何獲取這些量化值沒有限制,以上例子僅用於示範本發明的相關操作。
The
當取得一特徵點fx在每個單位時間點的量值之後,可將這些量值標
示成隨時間T改變的柱狀圖。藉由量值,系統100可得知任一特徵點在時間T內的變化情況。這種變化情形可用一柱狀圖表示。第6圖為示範例,示範一特徵點f49在時間T的變化情況。其中X軸代表單位時間點,而Y軸代表在每一單位時間點的量值。如圖所示,假設量值在t16,t17及t18,時分別為V16,V17及V18。整體來說,在時間T中特徵點f49的所有量值可以用一餘函數(residue function)Res_f49(t)集合表示。
After obtaining the magnitude of a feature point fx at each unit time point, these magnitudes can be marked as a histogram that changes with time T. Based on the value, the
隨後系統100的處理單元1302對餘函數作積分運算。積分的結果定義為特徵點f49的餘值(residue value),以RES(f49)表示。須注意的是,本發明的計算除了考慮微笑所發生的時間段(如t16至tN-15)外,亦考慮空白表情所發生的時間段(亦即t1至t15及tN-14至tN)。總結來說,積分計算的範圍是整個視頻的總時間T(亦即從t1至tN)。本發明任何一特徵點的餘值可以由下列積分公式表示:,其中fx為f49,f50,...及f68 Then the processing unit 1302 of the
在本發明中,空白臉部圖像320可視為動態臉部圖像310的保護間隔(guard-band),實際而言,空白臉部圖像320是做為預設臉部動作的同位檢查(parity check)。雖然第6圖將這些空白臉部圖像的量值以無數值(null)表示,但即使是空白表情亦可以透過上述任何方法量化以得到各單位時間點的量值。此外,本發明亦可另外定義一閾值(threshold),當量值低於該閾值時亦即表示為對應的為空白臉部圖像。 In the present invention, the blank facial image 320 can be regarded as the guard-band of the dynamic facial image 310. Actually, the blank facial image 320 is used as a parity check for preset facial movements ( parity check). Although Figure 6 shows the magnitude of these blank facial images as null, even the blank expression can be quantified by any of the above methods to obtain the magnitude of each unit time point. In addition, the present invention may additionally define a threshold (threshold), and when the value is lower than the threshold, it means that the corresponding blank face image is indicated.
所有20個考慮的特徵點的餘值都可經由上述積分公式計算得到,它們可以分別表示成RES(f49),RES(f50),...RES(f68);且可以集合表示為RES(X)={RES(f49),RES(f50),...RES(f68)},定義為預設臉部動作(亦即微笑)的變化向量
(variation vector)。此變化向量即是主要使用者做出預設臉部動作-微笑的基準檔案。此後,當系統100便參照基準檔案驗證使用者是否意圖要執行一綁定動作。
The residual values of all 20 considered feature points can be calculated by the above integral formula, they can be expressed as RES(f49), RES(f50),...RES(f68); and can be collectively expressed as RES(X )={RES(f49),RES(f50),...RES(f68)}, defined as the change vector of the preset facial action (ie smile)
(variation vector). This change vector is the reference file for the main user to make the default facial action-smile. Thereafter, the
如上所述,空白臉部圖像320出現於動態臉部圖像310的前後。在本發明中,建立預設臉部動作(如微笑)的基準檔案時不僅考慮預設臉部動作本身特徵點的變化,亦同時考慮在使用者做出預設臉部動作前後面無表情情況下臉部特徵點的不變化。在驗證臉部動作的同時考慮空白表情圖像是本發明的關鍵。如上面已經提到,即使當使用者作出一個與預設臉部動作極為相似的臉部動作時,因為相似臉部動作的前後不會是空白表情,因此針對近似臉部動作所建立的檔案在任何情況下不會與基準檔案符合,所以系統100不會誤判進而誤執行綁定動作。
As described above, the blank facial image 320 appears before and after the dynamic facial image 310. In the present invention, the creation of a reference file for a preset facial action (such as a smile) not only considers the changes in the feature points of the preset facial action itself, but also considers the absence of expression before and after the user makes the preset facial action The feature points of the lower face remain unchanged. It is the key of the present invention to consider blank expression images while verifying facial movements. As mentioned above, even when the user makes a facial motion that is very similar to the default facial motion, because the similar facial motion will not be blank expressions before and after the similar facial motion, the file created for the approximate facial motion is in In any case, it will not match the reference file, so the
第7圖為使用本發明系統以及方法驗證使用者臉部動作的示範例。延續上面提到的例子,假設預設臉部動作是一個微笑。即使使用者可能在交談過程中作出了相似的微笑動作,但由於在近似微笑的臉部圖像前後出現了的臉部圖像並非空白臉部圖像,在總體判斷下,該變化向量,即提取的該似微笑臉部動作與一純微笑並不相同。因此通過在檔案建制過程中增加至少一保護間隔,本發明中的該臉部識別系統可被進一步增強識別準確率同時消除誤識別。 Figure 7 is an example of using the system and method of the present invention to verify the user's facial movements. Continuing the example mentioned above, suppose the default facial action is a smile. Even if the user may have made a similar smiling action during the conversation, since the facial images that appear before and after the similarly smiling facial image are not blank facial images, under the overall judgment, the change vector is The extracted facial action that looks like a smile is not the same as a pure smile. Therefore, by adding at least one guard interval during the file creation process, the face recognition system of the present invention can further enhance the recognition accuracy and eliminate misrecognition.
在以上實施例中,量值Res_fx(t)的計算是基於單一臉部特徵點的變化軌跡。然而,如本領域的一般技術人員所理解的,此計算僅作為示範例。本發明亦可以考慮兩個或更多臉部特徵點來計算量值Res_fx(t)。例如,一量值Res_fx(t)可以是任何3個臉部特徵點的角度軌跡,或是任何兩個臉部特徵點之間的距離等等。 In the above embodiment, the calculation of the magnitude Res_fx(t) is based on the change trajectory of a single facial feature point. However, as understood by those of ordinary skill in the art, this calculation is only an example. The present invention can also consider two or more facial feature points to calculate the magnitude Res_fx(t). For example, a value of Res_fx(t) can be the angle trajectory of any three facial feature points, or the distance between any two facial feature points, and so on.
此外,上述實施例以面無表情作為保護間隔用來驗證預設臉部動作是否真的被做出,但習知技術的人應該知道本發明不限於此。在其他實施例中, 只要某一臉部表情的量值低於一預設閾值即可作為保護間隔,用以分辨預設臉部動作及其他臉部動作。 In addition, the above embodiment uses a blank expression as a guard interval to verify whether the preset facial motion is actually performed, but those skilled in the art should know that the present invention is not limited to this. In other embodiments, As long as the magnitude of a facial expression is lower than a preset threshold, it can be used as a guard interval to distinguish the preset facial motion from other facial motions.
一旦基準檔案建立完畢,本發明的臉部識別系統100便可基於基準檔案進行臉部動作的辨別。相關步驟可見第8圖並同時參照第2圖,並於下述篇幅中詳細說明。
Once the reference file is created, the
首先假設一使用者的身份(identity)已確定匹配。首先,在步驟801,影像擷取裝置110可連續追蹤並記錄該使用者的臉部動作。接著,於步驟802中,處理單元130判斷使用者的連續臉部動作是否與之前記錄的預設臉部動作(例如:一微笑)相似。如果兩個臉部動作相似,則影像擷取裝置110將近似臉部動作計錄成時間長度T的一對比視頻(步驟803)。其中對比視頻包含複數個關鍵臉部圖像(critical facial patterns),而這些關鍵臉部圖像的組合即為近似臉部動作;此外,對比視頻亦包含複數個次要臉部圖像(secondary facial patters),而這些次要臉部圖像可能出現於對比臉部動作的前面、後面、或前面以及後面。需注意的是,這些次要臉部圖像可以是除了對比臉部動作以外的任何臉部表情或動作。在本實施例中,對比視頻的時間長度T也同樣被切分成複數個單位時間。
First, assume that the identity of a user has been determined to match. First, in step 801, the
接著,於步驟804中,本發明系統100採取與設立基準檔案同樣的方法根據對比視頻針對近似臉部動作建立一對比檔案。延續上面微笑的例子,簡單來說,處理單元130在每一個關鍵臉部圖像以及次要臉部圖像中標注相同的20個臉部特徵點。接著,計算每一個臉部特徵點在單位時間的量值。再接著計算出每一個臉部特徵點的柱狀圖及餘函數。隨後處理單元130積分每一臉部特徵點的餘函數以求得一餘值。所有餘值的集合即是近似臉部動作的檔案(亦稱為向量變化量(vector variation))。值得注意的是,在本發明中前述的積分計算既考慮對比臉部動作亦考慮其前後出現的其他臉部表情。如上所述,利用相同方法,對
比面部動作的對比檔案即可建立完成。
Next, in step 804, the
接著於步驟805,在對比檔案建立完成後,處理單元130將其與儲存在資料庫120的基準檔案進行比較。如果兩個檔案符合,則系統100驗證近似臉部動作即為預設臉部動作(步驟806)。換句話說,使用者實際上做出了和預設臉部動作一樣的臉部動作。
Then in step 805, after the comparison file is created, the
由於保護間隔(亦即:空白表情)的存在,本發明系統100不會因為動態臉部圖像310(即預設臉部動作:微笑)與關鍵臉部圖像(即對比臉部動作:似微笑)完全相同而發生誤判的情況。本發明的系統100會考慮在對比臉部動作之前,之後,或之前及之後所出現的臉部表情以確定該使用者是否實際上作出了一微笑動作(即預設臉部動作)。
Due to the existence of the guard interval (ie: blank expression), the
若兩個檔案匹配,與本發明的臉部動作驗證步驟即結束。然而,如同835申請案已經提到的,本發明系統100可以近一步的執行一動態時間規整(dynamic time warping)去更精確驗證近似臉部動作。
If the two files match, the facial motion verification step of the present invention ends. However, as mentioned in the 835 application, the
第9圖簡述了針對一臉部動作建立一檔案的步驟。該步驟可適用於上述近似臉部動作檔案以及基準檔案的建立。 Figure 9 briefly describes the steps to create a profile for a facial action. This step can be applied to the creation of the aforementioned approximate facial motion files and reference files.
首先於步驟901中,系統100提取使用者的至少一臉部特徵點。臉部特徵點可儲存於資料庫120中。其次,於步驟902中,影像擷取裝置110錄攝一視頻,其中該視頻的時間長度為T,且使用者於視頻中作出一關鍵臉部動作及至少一次要臉部動作。此外,視頻包含複數個主要臉部圖像(primary facial patterns)以及附屬臉部圖像(collateral facial pattern);其中主要臉部圖像是關鍵臉部動作的組合,而附屬臉部圖像則為次要臉部動作的集合。此外,臉部圖像可透過以單位時間取樣視頻取得。如上所述,次要臉部動作可能出現在關鍵臉部動作的前、後、或者前以及後。本發明的處理單元130具備有分辨關鍵臉部動作以及次要臉
部動作的能力。舉例來說,處理單元130可藉由一預設閾值辨別一臉部圖像為主要臉部圖像或附屬臉部圖像。接著於步驟903,處理單元130分別於每一張主要臉部圖像以及附屬臉部圖像中標注臉部特徵點。最後步驟904,處理單元130針對這些標注的臉部特徵點建立一臉部動作檔案。
First, in step 901, the
在一實施例中,若第9圖的方法應用於建立預設臉部動作的基準檔案,則主要臉部圖像即為使用者做出預設臉部動作(例如:微笑)的動態臉部圖像;而附屬臉部圖像則為使用者面無表情時的空白臉部圖像。另外,本發明系統100可以儲存多個基準檔案以對應多個綁定功能。相關的說明以於835申請案中揭露,在此不贅述。
In one embodiment, if the method in Figure 9 is applied to create a reference file of preset facial motions, the main facial image is the dynamic face of the user that makes the default facial motions (for example, smiling) Image; while the attached face image is a blank face image when the user has no expression on his face. In addition, the
在另一實施例中,若第9圖的方法適用於建立對比臉部動作的檔案,則主要臉部圖像即為使用者做出對比臉部動作(例如:近似微笑)的臉部圖像;而附屬臉部圖像則為出現在對比臉部動作前、後或者前跟後的臉部動作。 In another embodiment, if the method in Figure 9 is suitable for creating a file for comparing facial motions, the main facial image is the facial image of the user who made the facial motions (for example, a similar smile). ; The subsidiary facial image is the facial motion that appears before, after, or before and after the contrast facial motion.
除上述外,一檔案的建立還可包含以下額外步驟(未見於第9圖)。首先,處理單元130基於主要臉部圖像及附屬臉部圖像計算每一個臉部特徵點的複數個量值。該量值代表了臉部特徵點在時間T上相對於一預設錨點的變化。該變化可以是位置,距離,角度,或以上任一的組合的變化。當取得於餘函數表示的量值後,處理單元130積分時間範圍T的餘函數以取得某一臉部特徵點的餘值。在此情況下,由於只有一個臉部特徵點被考慮,該餘值(即變化向量)即為該臉部動作的檔案。如果考慮多個臉部特徵點,該檔案即時所有臉部特徵點餘值的集合。
In addition to the above, the creation of a file can also include the following additional steps (not shown in Figure 9). First, the
以上所有實施例中皆假設使用者身份已經驗證完畢。驗證的方法可以是任何習知方法或是835申請案中所揭露的方法。此外,即便使用者的身份並未事先驗證,由於本發明中利用使用者的臉部特徵點進行臉部動作的核實,因 此在核實的同時也可以一併完成使用者身分驗證。換句話說,在本發明其中一個實施例中,使用者驗證的步驟可以忽略。 In all the above embodiments, it is assumed that the user identity has been verified. The verification method can be any conventional method or the method disclosed in the 835 application. In addition, even if the user’s identity is not verified in advance, because the present invention uses the user’s facial feature points to verify facial movements, This verification can also complete the user identity verification at the same time. In other words, in one of the embodiments of the present invention, the user authentication step can be omitted.
雖然以上實施例通過一單一臉部動作來說明本發明的工作原理,在本發明其它實施例中亦可以採用複數個臉部動作。舉例來說,假設系統預設兩個連續的臉部動作(例如:微笑並眨眼)為進行一綁定操作的預設臉部動作,則微笑(第一臉部動作)的保護間隔可以是微笑發生前的面無表情動作和/或是在微笑之後發生的前幾個的眨眼(第二臉部動作)臉部圖像。另外,眨眼(第二臉部動作)的保護間隔可以前面微笑(第一臉部動作)的最後幾個臉部圖像和/或一在眨眼後的做出的臉部表情動作。亦或者,本發明系統與方法可以將微笑並眨眼當作同一個臉部動作並使用上述相同方法進行辨識。 Although the above embodiments illustrate the working principle of the present invention through a single facial motion, in other embodiments of the present invention, multiple facial motions can also be used. For example, suppose that the system presets two consecutive facial actions (for example, smiling and blinking) as the preset facial actions for a binding operation, and the guard interval of the smile (first facial action) may be smile The expressionless face action before the occurrence and/or the first few blinks (second facial action) face images that occurred after the smile. In addition, the guard interval for blinking (second facial motion) may be the last few facial images of a front smile (first facial motion) and/or a facial expression motion made after blinking. Or alternatively, the system and method of the present invention can treat smiling and blinking as the same facial action and use the same method for recognition.
以上所述僅為本發明之較佳實施例,凡依本發明申請專利範圍所做之均等變化與修飾,皆應屬本發明之涵蓋範圍。 The foregoing descriptions are only preferred embodiments of the present invention, and all equivalent changes and modifications made in accordance with the scope of the patent application of the present invention shall fall within the scope of the present invention.
100:識別系統 100: identification system
110:影像擷取裝置 110: Image capture device
120:資料庫 120: database
130:處理單元 130: processing unit
Claims (18)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/285,267 | 2019-02-26 | ||
US16/285,267 US10460454B2 (en) | 2017-11-09 | 2019-02-26 | Motion-based facial recognition |
Publications (2)
Publication Number | Publication Date |
---|---|
TWI701607B true TWI701607B (en) | 2020-08-11 |
TW202032421A TW202032421A (en) | 2020-09-01 |
Family
ID=72199287
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
TW108119853A TWI701607B (en) | 2019-02-26 | 2019-06-10 | Dynamic face recognition system and method thereof |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN111611833A (en) |
TW (1) | TWI701607B (en) |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9202105B1 (en) * | 2012-01-13 | 2015-12-01 | Amazon Technologies, Inc. | Image analysis for user authentication |
TW201701192A (en) * | 2015-06-30 | 2017-01-01 | 芋頭科技(杭州)有限公司 | Facial recognition system and facial recognition method |
US20170061970A1 (en) * | 2015-08-25 | 2017-03-02 | Malaspina Labs (Barbados), Inc. | Speaker Dependent Voiced Sound Pattern Detection Thresholds |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2014116033A (en) * | 2008-06-16 | 2014-06-26 | Canon Inc | Personal authentication device, personal authentication method, camera, program, and storage medium |
CN105993022B (en) * | 2016-02-17 | 2019-12-27 | 香港应用科技研究院有限公司 | Method and system for recognition and authentication using facial expressions |
US10783352B2 (en) * | 2017-11-09 | 2020-09-22 | Mindtronic Ai Co., Ltd. | Face recognition system and method thereof |
-
2019
- 2019-06-10 TW TW108119853A patent/TWI701607B/en active
- 2019-06-27 CN CN201910569612.5A patent/CN111611833A/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9202105B1 (en) * | 2012-01-13 | 2015-12-01 | Amazon Technologies, Inc. | Image analysis for user authentication |
TW201701192A (en) * | 2015-06-30 | 2017-01-01 | 芋頭科技(杭州)有限公司 | Facial recognition system and facial recognition method |
US20170061970A1 (en) * | 2015-08-25 | 2017-03-02 | Malaspina Labs (Barbados), Inc. | Speaker Dependent Voiced Sound Pattern Detection Thresholds |
Also Published As
Publication number | Publication date |
---|---|
CN111611833A (en) | 2020-09-01 |
TW202032421A (en) | 2020-09-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9985963B2 (en) | Method and system for authenticating liveness face, and computer program product thereof | |
US11386900B2 (en) | Visual speech recognition by phoneme prediction | |
CN106557723B (en) | Face identity authentication system with interactive living body detection and method thereof | |
WO2019091012A1 (en) | Security check method based on facial recognition, application server, and computer readable storage medium | |
US9208375B2 (en) | Face recognition mechanism | |
JP3335538B2 (en) | Method and apparatus for collating a handwritten character string | |
WO2015131468A1 (en) | Method and system for estimating fingerprint pose | |
CN109766785A (en) | A kind of biopsy method and device of face | |
KR102062708B1 (en) | Apparatus and method for detection presentaion attack for face recognition system | |
Omoyiola | Overview of biometric and facial recognition techniques | |
Siirtola et al. | Effect of context in swipe gesture-based continuous authentication on smartphones | |
TWI526953B (en) | Face recognition method and system | |
KR102366777B1 (en) | Apparatus and method for domain adaptation-based object recognition | |
TWI701607B (en) | Dynamic face recognition system and method thereof | |
JP2011221699A (en) | Operation instruction recognition device and robot | |
Huang et al. | A new dataset for smartphone gesture-based authentication | |
US11017198B2 (en) | Slap segmentation of contactless fingerprint images | |
US9613252B1 (en) | Fingerprint matching method and device | |
KR20120132337A (en) | Apparatus and Method for Controlling User Interface Using Sound Recognition | |
CN111401167A (en) | Method for realizing biological recognition processing based on face recognition and palm vein recognition and corresponding system | |
US20230154236A1 (en) | Landmark-based ensemble network creation method for facial expression classification and facial expression classification method using created ensemble network | |
KR101972231B1 (en) | Apparatus and method of signature authentication using deep learning | |
US10460454B2 (en) | Motion-based facial recognition | |
CN112149559A (en) | Face recognition method and device, readable storage medium and computer equipment | |
Putz-Leszczynska et al. | Gait biometrics with a Microsoft Kinect sensor |