[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

TWI701607B - Dynamic face recognition system and method thereof - Google Patents

Dynamic face recognition system and method thereof Download PDF

Info

Publication number
TWI701607B
TWI701607B TW108119853A TW108119853A TWI701607B TW I701607 B TWI701607 B TW I701607B TW 108119853 A TW108119853 A TW 108119853A TW 108119853 A TW108119853 A TW 108119853A TW I701607 B TWI701607 B TW I701607B
Authority
TW
Taiwan
Prior art keywords
facial
motion
user
file
images
Prior art date
Application number
TW108119853A
Other languages
Chinese (zh)
Other versions
TW202032421A (en
Inventor
黃慕真
江昱嫺
瑞福 施
Original Assignee
英屬開曼群島商麥迪創科技股份有限公司
大陸商上海蔚蘭動力科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US16/285,267 external-priority patent/US10460454B2/en
Application filed by 英屬開曼群島商麥迪創科技股份有限公司, 大陸商上海蔚蘭動力科技有限公司 filed Critical 英屬開曼群島商麥迪創科技股份有限公司
Application granted granted Critical
Publication of TWI701607B publication Critical patent/TWI701607B/en
Publication of TW202032421A publication Critical patent/TW202032421A/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/174Facial expression recognition
    • G06V40/176Dynamic expression
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • G06V40/167Detection; Localisation; Normalisation using comparisons between temporally consecutive images

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

A dynamic face recognition system and method is provided based on facial motion data. The system includes a database, an image capture device, and a processing unit. The database is configured to store a preset facial motion of a user, which including at least one facial feature point. The image capture device tracks the facial motion of the user and captures an approximate video segment, wherein the user makes an approximate facial motion and at least an approximate facial expression in a time frame. The processing unit is coupled to the database and the image capturing device. The processing unit establishes a reference data and an approximate data of the approximate facial motion in the same file establishment process. The processing unit further compares the data with the reference data to verify the approximate facial motion.

Description

動態臉部識別系統與動態臉部識別方法 Dynamic face recognition system and dynamic face recognition method

本發明關於一種動態臉部識別,尤指一種用於動態臉部識別的使用者臉部動作檔案之設定。 The present invention relates to a dynamic face recognition, in particular to a setting of a user's facial motion file used for dynamic face recognition.

由於其廣泛的應用,臉部識別已成為身份識別的一種常規方案。業界提了多種新方法以提高臉部識別率。現有先前技術的臉部識別純粹是基於靜態圖像。一些系統可能聲稱它們是動態的,亦即使用者在移動中也可辨識,但其操作卻完全是基於即時拍攝的靜態圖像,本質上來說,這些所謂的動態是別還是屬於靜態的範疇。 Due to its wide range of applications, face recognition has become a conventional solution for identity recognition. The industry has proposed a variety of new methods to improve the facial recognition rate. The prior art facial recognition is purely based on static images. Some systems may claim that they are dynamic, that is, the user can recognize them while they are moving, but their operations are entirely based on real-time captured static images. In essence, these so-called dynamics are still in the static category.

根據一實施例,描述了一基於臉部動作的臉部識別系統。該系統包含一資料庫、一影像擷取裝置、及一處理單元。該資料庫用以儲存基於一使用者作出之一預設動作的一基準檔案且包含至少一臉部特徵點。該影像擷取裝置追蹤該使用者的臉部動作並拍攝一近似視頻片段,其中該使用者在一時幀內做出一近似臉部動作及至少一臉部表情。該處理單元與該資料庫及該影像擷取裝 置相連接。該處理單元用於設立該基準檔案及該近似臉部動作的一近似檔案。該處理單元進一步將該近似檔案與該基準檔案進行對比以驗證該近似臉部動作。 According to an embodiment, a facial recognition system based on facial motion is described. The system includes a database, an image capture device, and a processing unit. The database is used to store a reference file based on a predetermined action made by a user and includes at least one facial feature point. The image capturing device tracks the facial movements of the user and shoots an approximate video segment, wherein the user performs an approximate facial motion and at least one facial expression in a time frame. The processing unit and the database and the image capture device Set phase connection. The processing unit is used to set up the reference file and an approximate file of the approximate facial motion. The processing unit further compares the approximate file with the reference file to verify the approximate facial motion.

根據一實施例,描述了一種基於由具有至少一臉部特徵點的使用者通過作出臉部動作以設定臉部動作檔案的方法。此方法包含:通過一檔案設定進程設定包含一預設臉部動作的一基準檔案,在一時幀內拍攝一近似視頻片段,並在該時幀內,由使用者作出一近似臉部動作及至少一臉部表情;通過檔案設定進程設定近似臉部動作檔案;並將該近似臉部動作檔案與該基準檔案進行比較,以驗證該近似臉部動作檔案。 According to an embodiment, a method for setting a facial motion file based on a facial motion performed by a user with at least one facial feature point is described. This method includes: setting a reference file containing a preset facial motion through a file setting process, shooting an approximate video clip in a time frame, and in the time frame, the user makes an approximate facial motion and at least A facial expression; the approximate facial motion file is set through the file setting process; the approximate facial motion file is compared with the reference file to verify the approximate facial motion file.

根據一實施例,一動態臉部識別的方法如下。該方法包含設立一使用者的一臉部動作的一檔案以進行識別。該檔案設立流程進一步包括:提取該使用者的至少一臉部特徵點;在一時幀上追蹤並拍攝一視頻片段,其中該視頻片包含該使用者臉部動作的複數個主要臉部特徵,及使用者在該臉部動作之前,之後,或之前及之後作出至少一臉部表情的複數個附屬臉部特徵;在每個主要和附屬臉部特徵上標注臉部特徵點;並基於主要和附屬臉部特徵來設立該臉部動作的檔案。 According to an embodiment, a method of dynamic face recognition is as follows. The method includes creating a file of a facial movement of a user for recognition. The file establishment process further includes: extracting at least one facial feature point of the user; tracking and shooting a video clip on a time frame, wherein the video clip contains a plurality of main facial features of the user's facial movements, and Before, after, or before and after the facial action, the user makes at least one facial expression with a plurality of subsidiary facial features; mark the facial feature points on each main and subsidiary facial features; and based on the main and subsidiary facial features Facial features to set up a profile of the facial movement.

有關本發明之前述及其他技術內容、特點與功效,在以下配合參考圖式之實施例的詳細說明中,將可清楚的呈現。 The foregoing and other technical content, features, and effects of the present invention will be clearly presented in the following detailed description of the embodiments with reference to the drawings.

100:識別系統 100: identification system

110:影像擷取裝置 110: Image capture device

120:資料庫 120: database

130:處理單元 130: processing unit

300:視頻 300: Video

310:預設動態臉部圖像 310: Preset dynamic facial images

320:預設空白臉部圖像 320: preset blank face image

S801~S806、S901~S904:步驟 S801~S806, S901~S904: steps

第1圖為從一使用者臉部提取68個臉部特徵點的示例;第2圖示例了一種與本發明相關的基於臉部動作的臉部識別系統; 第3圖示例了在一時幀的一視頻片段包含多個動態臉部特徵及面無表情特徵;第4圖示例了20個臉部特徵點在時幀t1,t16,ta,tb,tN-15及tN的變化;第5圖示例了臉部特徵點f49在時幀t1,t16,ta,tb,tN-15及tN的相對位置;第6圖示例了臉部特徵點f49的一柱狀圖;第7圖示例了一似微笑之臉部特徵及其他臉部表情;第8圖示例了一基於本發明的動態臉部識別的流程圖;以及第9圖示例了一基於本發明的用於臉部識別的臉部動作檔案設立流程圖。 Figure 1 is an example of extracting 68 facial feature points from a user’s face; Figure 2 illustrates a facial recognition system based on facial motions related to the present invention; Figure 3 illustrates a moment A video segment of the frame contains multiple dynamic facial features and facial expressionless features; Figure 4 illustrates 20 facial feature points in time frames t 1 , t 16 , t a , t b , t N-15 and The change of t N ; Figure 5 illustrates the relative positions of the facial feature point f 49 in time frames t 1 , t 16 , t a , t b , t N-15 and t N ; Figure 6 illustrates the face A histogram of the facial feature point f 49 ; Figure 7 illustrates a smile-like facial feature and other facial expressions; Figure 8 illustrates a flow chart of dynamic facial recognition based on the present invention; and Figure 9 illustrates a flow chart of setting up a facial motion file for facial recognition based on the present invention.

以下實施例中所提到的方向用語,例如:上、下、左、右、前或後等,是參考附加圖式的方向。因此,使用的方向用語是用來說明並非用來限制本發明。 The directional terms mentioned in the following embodiments, for example: up, down, left, right, front or back, etc., refer to the directions of the attached drawings. Therefore, the directional terms used are used to illustrate but not to limit the present invention.

本申請於2018年5月10日提交台灣專利申請案號107115835(“835申請案”)相關,本申請案將835申請案的內容合併於此。根據835申請案的揭露,假設一使用者預先錄製一臉部動作(例如:微笑)綁定動作(例如,開啟車輛的雨刮器)的金鑰。根據835申請案,車內的一影像擷取裝置連續追蹤該使用者的臉部動作以偵測此使用者是否微笑以執行該特定動作。但假設在一情景中,該使用者在開車過程中與他人交談,該影像擷取裝置可能拍攝到該使用者的臉部圖像(facial patterns)正好與微笑相同。在此情況下,由於識別成功,雨刮器可能在未得到進一步確認的情況下便開始運行。這種突然的運行可分散司機的注意力且造成危險。 This application is related to Taiwan Patent Application No. 107115835 ("835 application") filed on May 10, 2018, and this application incorporates the content of the 835 application here. According to the disclosure of the 835 application, it is assumed that a user pre-records a key to a facial motion (for example, a smile) binding motion (for example, turning on a wiper of a vehicle). According to the 835 application, an image capture device in the car continuously tracks the user's facial motion to detect whether the user is smiling to perform the specific action. However, assuming that in a scenario, the user is talking with others while driving, the image capture device may capture the user's facial patterns which are exactly the same as the smile. In this case, due to the successful recognition, the wiper may start to operate without further confirmation. This sudden movement can distract the driver and cause danger.

儘管傳統的臉部識別系統已經進行了許多創新以減少誤識別,然而當涉及到動態臉部識別領域時,它們都不適用。主要的原因是在傳統靜態的臉 部識別中,所有因素和參數都是恒定的,所以相對容易除錯。相反地,在處理動態臉部識別時,所有臉部特徵的變化都需要被考慮,這些都複雜了動態臉部識別的計算和處理。 Although traditional facial recognition systems have made many innovations to reduce misrecognition, they are not applicable when it comes to the field of dynamic facial recognition. The main reason is in the traditional static face In part recognition, all factors and parameters are constant, so it is relatively easy to debug. On the contrary, when dealing with dynamic face recognition, all facial feature changes need to be considered, which complicates the calculation and processing of dynamic face recognition.

本發明主要是針對835申請案的再發明。本發明披露了一種通過設立臉部動作檔案(Profile)以進行動態臉部識別的系統和方法。採用本發明不僅可以提高動態臉部識別的準確性,並消除動作誤判的可能。 The present invention is mainly a re-invention of the 835 application. The invention discloses a system and method for dynamic facial recognition by setting up a facial action profile (Profile). The invention can not only improve the accuracy of dynamic face recognition, but also eliminate the possibility of misjudgment of actions.

臉部識別的運作始於收集一使用者的臉部特徵(facial features)。對於專業人士而言有多種已知方法可以用來選擇臉部特徵、特徵點(landmarks)和/或關鍵點(key points)。在一實施例中,一臉部識別系統認出了一主要使用者(primary user)臉部的多個獨特的臉部特徵/臉部特徵點。如第1圖所示,在一實施例中,一使用者的臉部影像(facial image)過濾出了68個獨特的特徵點(統稱“標注資料(label data)”),標注資料經過如類神經網路等習知技術的訓練可獲得一訓練模型(trained model)。標注資料及訓練模型儲存於資料庫供之後使用。後續識別中,系統將嘗試提取待驗證使用者的68個臉部特徵點並輸入該訓練模型中。如待驗證使用者的臉部特徵點與資料庫儲存的資料匹配,則系統認證驗證使用者確實為主要使用者。對系統而言,多個使用者的標注資料都可以儲存在資料庫中作為驗證。此外,上述例子擷取68個特徵點僅是一個例子,系統可能擷取更多或較少的特徵點,但亦可以達到與本申請案相同的目地,因此數目68不應作為對本發明的限制。 The operation of facial recognition begins with the collection of facial features of a user. There are many known methods for professionals to select facial features, landmarks and/or key points. In one embodiment, a facial recognition system recognizes a plurality of unique facial features/facial feature points of a primary user's face. As shown in Figure 1, in one embodiment, a user’s facial image (facial image) filters out 68 unique feature points (collectively referred to as "label data"). Training of conventional technologies such as neural networks can obtain a trained model. The annotation data and training model are stored in the database for later use. In the subsequent recognition, the system will try to extract 68 facial feature points of the user to be verified and input them into the training model. If the facial feature points of the user to be verified match the data stored in the database, the system verifies that the user is indeed the primary user. For the system, the annotation data of multiple users can be stored in the database for verification. In addition, the above example extracting 68 feature points is only an example. The system may extract more or less feature points, but it can also achieve the same purpose as the application. Therefore, the number 68 should not be used as a limitation to the present invention. .

另外值得注意的是,在臉部識別的過程中並不一定會考慮全部68個臉部特徵點。在某些情況下,系統可能只考慮它認為相關的臉部特徵點並忽略那些不相關的臉部特徵點。儘管精確度可能會受影響,但這種方法可以縮短計算處理時間。 It is also worth noting that all 68 facial feature points are not necessarily considered in the process of face recognition. In some cases, the system may only consider facial feature points that it considers relevant and ignore those that are not relevant. Although accuracy may be affected, this method can shorten the calculation processing time.

基於動態臉部動作驗證使用者身份的過程已於835申請案中詳細的討論,本發明基於835申請案近一步揭露減少動態臉部動作誤驗證的方法及系統。 The process of verifying the user's identity based on dynamic facial movements has been discussed in detail in the 835 application. Based on the 835 application, the present invention further discloses a method and system for reducing false verification of dynamic facial movements.

本發明的動態臉部識別系統100的運作是通過建立臉部動作檔案(facial motion profiles)開始。識別系統100的結構如第2圖所示。本發明的識別系統100包括一資料庫120,存儲一使用者的至少一個臉部特徵點以及使用者的基準檔案(predetermined profile)。其中,基準檔案是由使用者作出一個預設的臉部動作(default facial motion)而建立。本發明的識別系統100還包括一影像擷取裝置110,用以追蹤使用者的臉部變化並拍攝一時幀的視頻,以及處理單元130。其中處理單元130的作用包含建立臉部動作檔案,以及比較近似檔案(resembling profile)與基準檔案,最後並判斷該近似檔案是否與基準檔案相似。識別系統100還可以配備各種類型的記憶體(圖中未示出)以存儲由影像擷取裝置110拍攝的圖像及視頻,以及處理單元130所處理的資料。第3圖為本發明其中一實施例。如第3圖所示,影像擷取裝置110在一段時間T內擷取到一段使用者微笑及面無表情的視頻。以下說明請一併參考第2圖和第3圖以了解臉部識別系統100的操作。 The operation of the dynamic facial recognition system 100 of the present invention is started by creating facial motion profiles. The structure of the recognition system 100 is shown in FIG. 2. The recognition system 100 of the present invention includes a database 120 that stores at least one facial feature point of a user and the user's reference file (predetermined profile). Among them, the reference file is created by the user making a default facial motion. The recognition system 100 of the present invention further includes an image capturing device 110 for tracking the user's face changes and shooting a time frame of video, and a processing unit 130. The function of the processing unit 130 includes creating a facial motion file, comparing a resembling profile with a reference file, and finally determining whether the resembling profile is similar to the reference file. The recognition system 100 can also be equipped with various types of memory (not shown in the figure) to store the images and videos taken by the image capture device 110 and the data processed by the processing unit 130. Figure 3 shows one embodiment of the present invention. As shown in Figure 3, the image capturing device 110 captures a video of the user smiling and expressionless within a period of time T. In the following description, please refer to FIG. 2 and FIG. 3 together to understand the operation of the face recognition system 100.

在一實施例中,系統100首先就一主要使用者作出一預設臉部動作建立一基準檔案。這個基準檔案作為之後動態臉部識別的參照。如835申請案已經提及的,預設臉部動作可與一功能綁定,當使用者作出預設臉部動作時,該綁並功能將被啟動。綁定功能可能與汽車功能相關,例如打開雨刷等等這種以往需要由駕駛手動控制的功能。建立基準檔案的步驟如下。首先,影像擷取裝置100擷取主要使用者的臉部動作並記錄成一個視頻300,視頻的時間總長度為T。其中視頻300至少包含一預設動態臉部圖像310(例如:當主要使用者微笑的臉部圖像)以及一預設空白臉部圖像320(例如:主要使用者面無表情的臉部圖像)。 其中視頻時間總長度T可被分割為多個單位時間Tx={t1,t2,...,tN},每一個單位時間點Tx皆對應一個臉部圖像(facial pattern)。需要注意的是,臉部圖像的數目不一定等於幀數或與視頻取樣速度正相關。基於一實施例,一臉部圖像可為多幀的組合(例如:10幀),而一單位時間便等同於多幀組合的時間總長(例如:10幀的時間總長)。 In one embodiment, the system 100 first creates a reference file for a predetermined facial movement made by a main user. This benchmark file serves as a reference for dynamic facial recognition later. As mentioned in the 835 application, the default facial action can be bound to a function, and the binding function will be activated when the user performs the preset facial action. The binding function may be related to the car function, such as turning on the wiper, etc., which previously required manual control by the driver. The steps to create a benchmark file are as follows. First, the image capturing device 100 captures the facial movements of the main user and records it into a video 300, the total time length of the video is T. The video 300 includes at least a preset dynamic face image 310 (for example, a face image of the main user smiling) and a preset blank face image 320 (for example: a face of the main user without expression image). The total length of video time T can be divided into multiple unit times Tx={t 1 , t 2 ,..., t N }, and each unit time point Tx corresponds to a facial image (facial pattern). It should be noted that the number of facial images is not necessarily equal to the number of frames or positively correlated with the video sampling speed. According to an embodiment, a face image can be a combination of multiple frames (for example, 10 frames), and a unit time is equivalent to the total time length of the multiple frame combination (for example, the total time length of 10 frames).

假設視頻300共包含N個臉部圖像Mx={M1,M2,...,MN},其中部分為動態臉部圖像310,部分為空白面度圖像320。此外,各臉部圖像表現了相關單位時間點時瞬間的臉部快照。例如,一第一臉部圖像M1為使用者於第一單位時間點t1的快照,而一第二臉部圖像M2為使用者於第二時間點t2的快照等。在一實施例中,假設一預設臉部動作為微笑,如第3圖所示,從視頻300的角度上來看,主要使用者于第16個單位時間點t16(亦即對應第16個臉部圖像M16)開始微笑;而空白臉部圖像則出現於視頻300的前後各15個單位時間點中(亦即對應第1個臉部圖像M1至第15個臉部圖像M15,以及第N-14個臉部圖像MN-14到第N個臉部圖像MN)。在此實施例中假定空白臉部表情320出現於動態臉部圖像310的兩側,但本發明不局限於此。主要使用者可決定在微笑之前或之後作出面無表情的動作。換句話說,空白臉部圖像320可出現在動態臉部圖像310之前,之後,或之前與之後;以上各種方式皆可達到本發明的目的。 It is assumed that the video 300 contains a total of N facial images Mx={M 1 , M 2 ,..., M N }, part of which is a dynamic facial image 310 and a part is a blank face image 320. In addition, each facial image expresses a facial snapshot at an instant in the relevant unit time point. For example, a first facial image M 1 is a snapshot of the user at a first unit time point t 1 , and a second facial image M 2 is a snapshot of the user at a second time point t 2 . In one embodiment, it is assumed that a preset facial action is a smile. As shown in Figure 3, from the perspective of the video 300, the main user at the 16th unit time point t 16 (that is, corresponding to the 16th The facial image M 16 ) starts to smile; and the blank facial image appears in each of the 15 unit time points before and after the video 300 (that is, corresponding to the first facial image M 1 to the 15th facial image like M 15, N-14 and a second face images M N-14 to N-th face image M N). In this embodiment, it is assumed that blank facial expressions 320 appear on both sides of the dynamic facial image 310, but the invention is not limited to this. The primary user can decide to make a blank gesture before or after smiling. In other words, the blank facial image 320 can appear before, after, or before and after the dynamic facial image 310; all the above methods can achieve the purpose of the present invention.

當視頻300的記錄完成後,68個臉部特徵點(即“標注資料”)可被標注於每一個動態及空白臉部圖像Mx={M1,M2,...,MN}上,且每個面部特徵點隨時間T的變化量(variation)亦可得到。值得一提的是,在實施例使用微笑作為預設的臉部動作,因此系統100可以只提取使主要使用者嘴部周邊的臉部特徵點而不必提取所有的68個臉部各徵點。例如,本發明的系統100在此情境中可以只提取嘴部周邊的20個臉部特徵點作為計算參考。如第1圖所示,假設這20個相關的臉 部特徵點為f49,f50,...,及f68,表示為Fx={f49,f50,...,f68}。 When the recording of the video 300 is completed, 68 facial feature points (ie "labeled data") can be labeled on each dynamic and blank facial image Mx={M 1 ,M 2 ,...,M N } Above, and the variation of each facial feature point over time T can also be obtained. It is worth mentioning that in the embodiment, a smile is used as the preset facial action, so the system 100 can only extract facial feature points around the main user's mouth instead of extracting all the 68 facial feature points. For example, the system 100 of the present invention can extract only 20 facial feature points around the mouth as a calculation reference in this situation. As shown in Figure 1, suppose the 20 relevant facial feature points are f 49 , f 50 ,..., and f 68 , expressed as Fx={f 49 ,f 50 ,...,f 68 } .

第4圖示範說明了本系統100參考這20個臉部特徵點在相關時間點的臉部圖像變化,尤其是這20個臉部特徵分別在單位時間點t1,t16,ta,tb,tN-15及tN上對應臉部圖像M1,M16,Ma,Mb,MN-15,及MN的變化。如之前提及的,微笑開始於第16個單位時間點且持續至第(N-15)個單位時間點(即t16-tN-15),其中空白臉部表情出現於前15個和最後15個單位時間點(t1-t15及tN-14-tN),其中M1及MN為空白臉部圖像,而M16,Ma,Mb,MN-15為動態臉部圖像。 Figure 4 demonstrates that the system 100 refers to the changes in the facial image at the relevant time points of the 20 facial feature points, especially the 20 facial features at the unit time points t 1 , t 16 , t a , t b, t N-15 and, M 16, M a, M b, M N-15, M N 1 and change the facial image corresponding to the M t N. As mentioned earlier, the smile starts at the 16th unit time point and continues to the (N-15)th unit time point (ie t 16 -t N-15 ), where blank facial expressions appear in the first 15 and Finally, 15 units of time (t 1 -t 15 and t N-14 -t N), wherein M 1 and M N is a blank facial image, and M 16, M a, M b , M N-15 is Dynamic facial images.

本發明的臉部識別系統100自每一單位時間點針對每一個特徵點fx(fx為f49,f50,...,或f68)計算一量值-Res_fx(t)來表示在單位時間點上的該特徵點的量化值。該量值可以是在一單位時間時特徵點的角度,距離或任一角度和距離所積累的位置變化。量值Res_fx(t)的計算可以通過數學變換(conversion)計算得到。例如,量值可以透過轉換一特徵點的座標、角度、距離、位置或相對位置等等計算得到。在一實施例中,量值Res_fx(t)被定義為一特徵點在一單位時間基於一預設錨點(如鼻尖)的相對位置。另外,量值亦可以是一特徵點在某一單位時間點前所有位置的累計積分值。例如,通過積分一特徵點自t1至t3的位置軌跡可以得到Res_fx(t3)。第5圖說明了特徵點f49在單位時間點t1,t16,ta,tb,tN-15及tN與一預設錨點(例如:鼻尖)的相對位置變化。在另一實施例中,量值的計算會將使用者臉部的寬度考慮進去用以消除比例的影響。利用上述方式,任何一個特徵值在任何一單位時間點對應的量值Res_f49(t)(t為t1,...,或tN)皆可得到。除了上述方法外,習知技術的人亦可用其他的方式得到單位時間點上特徵值的量化值,本發明對如何獲取這些量化值沒有限制,以上例子僅用於示範本發明的相關操作。 The face recognition system 100 of the present invention calculates a value-Res_fx(t) for each feature point fx (fx is f 49 , f 50 ,..., or f 68 ) from each unit time point to express in the unit The quantized value of the feature point at the time point. The value can be the angle of the feature point in a unit time, the distance, or the position change accumulated by any angle and distance. The calculation of the magnitude Res_fx(t) can be calculated by mathematical conversion (conversion). For example, the magnitude can be calculated by converting the coordinates, angle, distance, position, or relative position of a feature point. In one embodiment, the magnitude Res_fx(t) is defined as the relative position of a characteristic point based on a predetermined anchor point (such as the tip of the nose) in a unit time. In addition, the magnitude can also be the cumulative integral value of all positions of a characteristic point before a certain unit time point. For example, Res_fx(t 3 ) can be obtained by integrating the position trajectory of a feature point from t 1 to t 3 . Figure 5 illustrates the relative position changes of the feature point f 49 at unit time points t 1 , t 16 , t a , t b , t N-15 and t N and a preset anchor point (for example, the tip of the nose). In another embodiment, the calculation of the magnitude takes into account the width of the user's face to eliminate the influence of proportion. Using the above method, the magnitude Res_f 49 (t) (t is t 1 ,..., or t N ) corresponding to any eigenvalue at any unit time point can be obtained. In addition to the above methods, people in the prior art can also use other methods to obtain the quantified value of the characteristic value per unit time point. The present invention has no limitation on how to obtain these quantized values. The above examples are only used to demonstrate the relevant operations of the present invention.

當取得一特徵點fx在每個單位時間點的量值之後,可將這些量值標 示成隨時間T改變的柱狀圖。藉由量值,系統100可得知任一特徵點在時間T內的變化情況。這種變化情形可用一柱狀圖表示。第6圖為示範例,示範一特徵點f49在時間T的變化情況。其中X軸代表單位時間點,而Y軸代表在每一單位時間點的量值。如圖所示,假設量值在t16,t17及t18,時分別為V16,V17及V18。整體來說,在時間T中特徵點f49的所有量值可以用一餘函數(residue function)Res_f49(t)集合表示。 After obtaining the magnitude of a feature point fx at each unit time point, these magnitudes can be marked as a histogram that changes with time T. Based on the value, the system 100 can know the change of any characteristic point in time T. This change can be represented by a histogram. Figure 6 is an example of how a feature point f 49 changes at time T. The X axis represents the unit time point, and the Y axis represents the value at each unit time point. As shown in the figure, assuming that the values are at t 16 , t 17 and t 18 , the time is V 16 , V 17 and V 18 respectively . On the whole, all the values of the feature point f 49 at time T can be represented by a set of residue functions (residue function) Res_f 49 (t).

隨後系統100的處理單元1302對餘函數作積分運算。積分的結果定義為特徵點f49的餘值(residue value),以RES(f49)表示。須注意的是,本發明的計算除了考慮微笑所發生的時間段(如t16至tN-15)外,亦考慮空白表情所發生的時間段(亦即t1至t15及tN-14至tN)。總結來說,積分計算的範圍是整個視頻的總時間T(亦即從t1至tN)。本發明任何一特徵點的餘值可以由下列積分公式表示:

Figure 108119853-A0305-02-0011-1
,其中fx為f49,f50,...及f68 Then the processing unit 1302 of the system 100 performs an integral operation on the remainder function. The result of the integration is defined as the residual value of the feature point f 49 , which is represented by RES(f 49 ). It should be noted that the calculation of the present invention not only considers the time period when a smile occurs (such as t 16 to t N-15 ), but also considers the time period when a blank expression occurs (that is, t 1 to t 15 and t N- 14 to t N ). In summary, the range of integral calculation is the total time T of the entire video (that is, from t 1 to t N ). The residual value of any characteristic point of the present invention can be expressed by the following integral formula:
Figure 108119853-A0305-02-0011-1
, Where fx is f 49 , f 50 ,... and f 68

在本發明中,空白臉部圖像320可視為動態臉部圖像310的保護間隔(guard-band),實際而言,空白臉部圖像320是做為預設臉部動作的同位檢查(parity check)。雖然第6圖將這些空白臉部圖像的量值以無數值(null)表示,但即使是空白表情亦可以透過上述任何方法量化以得到各單位時間點的量值。此外,本發明亦可另外定義一閾值(threshold),當量值低於該閾值時亦即表示為對應的為空白臉部圖像。 In the present invention, the blank facial image 320 can be regarded as the guard-band of the dynamic facial image 310. Actually, the blank facial image 320 is used as a parity check for preset facial movements ( parity check). Although Figure 6 shows the magnitude of these blank facial images as null, even the blank expression can be quantified by any of the above methods to obtain the magnitude of each unit time point. In addition, the present invention may additionally define a threshold (threshold), and when the value is lower than the threshold, it means that the corresponding blank face image is indicated.

所有20個考慮的特徵點的餘值都可經由上述積分公式計算得到,它們可以分別表示成RES(f49),RES(f50),...RES(f68);且可以集合表示為RES(X)={RES(f49),RES(f50),...RES(f68)},定義為預設臉部動作(亦即微笑)的變化向量 (variation vector)。此變化向量即是主要使用者做出預設臉部動作-微笑的基準檔案。此後,當系統100便參照基準檔案驗證使用者是否意圖要執行一綁定動作。 The residual values of all 20 considered feature points can be calculated by the above integral formula, they can be expressed as RES(f49), RES(f50),...RES(f68); and can be collectively expressed as RES(X )={RES(f49),RES(f50),...RES(f68)}, defined as the change vector of the preset facial action (ie smile) (variation vector). This change vector is the reference file for the main user to make the default facial action-smile. Thereafter, the system 100 refers to the reference file to verify whether the user intends to perform a binding action.

如上所述,空白臉部圖像320出現於動態臉部圖像310的前後。在本發明中,建立預設臉部動作(如微笑)的基準檔案時不僅考慮預設臉部動作本身特徵點的變化,亦同時考慮在使用者做出預設臉部動作前後面無表情情況下臉部特徵點的不變化。在驗證臉部動作的同時考慮空白表情圖像是本發明的關鍵。如上面已經提到,即使當使用者作出一個與預設臉部動作極為相似的臉部動作時,因為相似臉部動作的前後不會是空白表情,因此針對近似臉部動作所建立的檔案在任何情況下不會與基準檔案符合,所以系統100不會誤判進而誤執行綁定動作。 As described above, the blank facial image 320 appears before and after the dynamic facial image 310. In the present invention, the creation of a reference file for a preset facial action (such as a smile) not only considers the changes in the feature points of the preset facial action itself, but also considers the absence of expression before and after the user makes the preset facial action The feature points of the lower face remain unchanged. It is the key of the present invention to consider blank expression images while verifying facial movements. As mentioned above, even when the user makes a facial motion that is very similar to the default facial motion, because the similar facial motion will not be blank expressions before and after the similar facial motion, the file created for the approximate facial motion is in In any case, it will not match the reference file, so the system 100 will not misjudge and perform the binding action by mistake.

第7圖為使用本發明系統以及方法驗證使用者臉部動作的示範例。延續上面提到的例子,假設預設臉部動作是一個微笑。即使使用者可能在交談過程中作出了相似的微笑動作,但由於在近似微笑的臉部圖像前後出現了的臉部圖像並非空白臉部圖像,在總體判斷下,該變化向量,即提取的該似微笑臉部動作與一純微笑並不相同。因此通過在檔案建制過程中增加至少一保護間隔,本發明中的該臉部識別系統可被進一步增強識別準確率同時消除誤識別。 Figure 7 is an example of using the system and method of the present invention to verify the user's facial movements. Continuing the example mentioned above, suppose the default facial action is a smile. Even if the user may have made a similar smiling action during the conversation, since the facial images that appear before and after the similarly smiling facial image are not blank facial images, under the overall judgment, the change vector is The extracted facial action that looks like a smile is not the same as a pure smile. Therefore, by adding at least one guard interval during the file creation process, the face recognition system of the present invention can further enhance the recognition accuracy and eliminate misrecognition.

在以上實施例中,量值Res_fx(t)的計算是基於單一臉部特徵點的變化軌跡。然而,如本領域的一般技術人員所理解的,此計算僅作為示範例。本發明亦可以考慮兩個或更多臉部特徵點來計算量值Res_fx(t)。例如,一量值Res_fx(t)可以是任何3個臉部特徵點的角度軌跡,或是任何兩個臉部特徵點之間的距離等等。 In the above embodiment, the calculation of the magnitude Res_fx(t) is based on the change trajectory of a single facial feature point. However, as understood by those of ordinary skill in the art, this calculation is only an example. The present invention can also consider two or more facial feature points to calculate the magnitude Res_fx(t). For example, a value of Res_fx(t) can be the angle trajectory of any three facial feature points, or the distance between any two facial feature points, and so on.

此外,上述實施例以面無表情作為保護間隔用來驗證預設臉部動作是否真的被做出,但習知技術的人應該知道本發明不限於此。在其他實施例中, 只要某一臉部表情的量值低於一預設閾值即可作為保護間隔,用以分辨預設臉部動作及其他臉部動作。 In addition, the above embodiment uses a blank expression as a guard interval to verify whether the preset facial motion is actually performed, but those skilled in the art should know that the present invention is not limited to this. In other embodiments, As long as the magnitude of a facial expression is lower than a preset threshold, it can be used as a guard interval to distinguish the preset facial motion from other facial motions.

一旦基準檔案建立完畢,本發明的臉部識別系統100便可基於基準檔案進行臉部動作的辨別。相關步驟可見第8圖並同時參照第2圖,並於下述篇幅中詳細說明。 Once the reference file is created, the facial recognition system 100 of the present invention can recognize facial movements based on the reference file. The relevant steps can be seen in Figure 8 and refer to Figure 2 at the same time, and are described in detail in the following pages.

首先假設一使用者的身份(identity)已確定匹配。首先,在步驟801,影像擷取裝置110可連續追蹤並記錄該使用者的臉部動作。接著,於步驟802中,處理單元130判斷使用者的連續臉部動作是否與之前記錄的預設臉部動作(例如:一微笑)相似。如果兩個臉部動作相似,則影像擷取裝置110將近似臉部動作計錄成時間長度T的一對比視頻(步驟803)。其中對比視頻包含複數個關鍵臉部圖像(critical facial patterns),而這些關鍵臉部圖像的組合即為近似臉部動作;此外,對比視頻亦包含複數個次要臉部圖像(secondary facial patters),而這些次要臉部圖像可能出現於對比臉部動作的前面、後面、或前面以及後面。需注意的是,這些次要臉部圖像可以是除了對比臉部動作以外的任何臉部表情或動作。在本實施例中,對比視頻的時間長度T也同樣被切分成複數個單位時間。 First, assume that the identity of a user has been determined to match. First, in step 801, the image capturing device 110 can continuously track and record the facial movements of the user. Next, in step 802, the processing unit 130 determines whether the user's continuous facial motion is similar to a previously recorded preset facial motion (for example, a smile). If the two facial movements are similar, the image capturing device 110 records the approximate facial movements into a comparison video of the time length T (step 803). The comparison video contains a plurality of critical facial images (critical facial patterns), and the combination of these key facial images is an approximate facial action; in addition, the comparison video also contains a plurality of secondary facial images (secondary facial patterns). patters), and these secondary facial images may appear in front of, behind, or in front of and behind contrasting facial movements. It should be noted that these secondary facial images can be any facial expressions or actions other than contrasting facial actions. In this embodiment, the time length T of the comparison video is also divided into a plurality of unit times.

接著,於步驟804中,本發明系統100採取與設立基準檔案同樣的方法根據對比視頻針對近似臉部動作建立一對比檔案。延續上面微笑的例子,簡單來說,處理單元130在每一個關鍵臉部圖像以及次要臉部圖像中標注相同的20個臉部特徵點。接著,計算每一個臉部特徵點在單位時間的量值。再接著計算出每一個臉部特徵點的柱狀圖及餘函數。隨後處理單元130積分每一臉部特徵點的餘函數以求得一餘值。所有餘值的集合即是近似臉部動作的檔案(亦稱為向量變化量(vector variation))。值得注意的是,在本發明中前述的積分計算既考慮對比臉部動作亦考慮其前後出現的其他臉部表情。如上所述,利用相同方法,對 比面部動作的對比檔案即可建立完成。 Next, in step 804, the system 100 of the present invention adopts the same method as setting up a reference file to create a comparison file for approximate facial actions based on the comparison video. Continuing the example of smiling above, in simple terms, the processing unit 130 labels the same 20 facial feature points in each key facial image and secondary facial image. Next, calculate the magnitude of each facial feature point per unit time. Then calculate the histogram and co-function of each facial feature point. The processing unit 130 then integrates the remainder function of each facial feature point to obtain a remainder value. The set of all residual values is a file that approximates facial motions (also called vector variation). It is worth noting that the aforementioned integral calculation in the present invention considers both the comparison of facial actions and other facial expressions before and after it. As mentioned above, using the same method, The comparison file of facial movements can be created.

接著於步驟805,在對比檔案建立完成後,處理單元130將其與儲存在資料庫120的基準檔案進行比較。如果兩個檔案符合,則系統100驗證近似臉部動作即為預設臉部動作(步驟806)。換句話說,使用者實際上做出了和預設臉部動作一樣的臉部動作。 Then in step 805, after the comparison file is created, the processing unit 130 compares it with the reference file stored in the database 120. If the two files match, the system 100 verifies that the approximate facial motion is the default facial motion (step 806). In other words, the user actually made the same facial motion as the preset facial motion.

由於保護間隔(亦即:空白表情)的存在,本發明系統100不會因為動態臉部圖像310(即預設臉部動作:微笑)與關鍵臉部圖像(即對比臉部動作:似微笑)完全相同而發生誤判的情況。本發明的系統100會考慮在對比臉部動作之前,之後,或之前及之後所出現的臉部表情以確定該使用者是否實際上作出了一微笑動作(即預設臉部動作)。 Due to the existence of the guard interval (ie: blank expression), the system 100 of the present invention will not cause the dynamic facial image 310 (ie, the preset facial motion: smile) to be compared with the key facial image (ie, comparing the facial motion: similar). Smile) is exactly the same and misjudgment occurs. The system 100 of the present invention considers facial expressions before, after, or before and after the facial motions are compared to determine whether the user actually made a smiling motion (ie, a preset facial motion).

若兩個檔案匹配,與本發明的臉部動作驗證步驟即結束。然而,如同835申請案已經提到的,本發明系統100可以近一步的執行一動態時間規整(dynamic time warping)去更精確驗證近似臉部動作。 If the two files match, the facial motion verification step of the present invention ends. However, as mentioned in the 835 application, the system 100 of the present invention can further perform a dynamic time warping to more accurately verify the approximate facial motion.

第9圖簡述了針對一臉部動作建立一檔案的步驟。該步驟可適用於上述近似臉部動作檔案以及基準檔案的建立。 Figure 9 briefly describes the steps to create a profile for a facial action. This step can be applied to the creation of the aforementioned approximate facial motion files and reference files.

首先於步驟901中,系統100提取使用者的至少一臉部特徵點。臉部特徵點可儲存於資料庫120中。其次,於步驟902中,影像擷取裝置110錄攝一視頻,其中該視頻的時間長度為T,且使用者於視頻中作出一關鍵臉部動作及至少一次要臉部動作。此外,視頻包含複數個主要臉部圖像(primary facial patterns)以及附屬臉部圖像(collateral facial pattern);其中主要臉部圖像是關鍵臉部動作的組合,而附屬臉部圖像則為次要臉部動作的集合。此外,臉部圖像可透過以單位時間取樣視頻取得。如上所述,次要臉部動作可能出現在關鍵臉部動作的前、後、或者前以及後。本發明的處理單元130具備有分辨關鍵臉部動作以及次要臉 部動作的能力。舉例來說,處理單元130可藉由一預設閾值辨別一臉部圖像為主要臉部圖像或附屬臉部圖像。接著於步驟903,處理單元130分別於每一張主要臉部圖像以及附屬臉部圖像中標注臉部特徵點。最後步驟904,處理單元130針對這些標注的臉部特徵點建立一臉部動作檔案。 First, in step 901, the system 100 extracts at least one facial feature point of the user. The facial feature points can be stored in the database 120. Next, in step 902, the image capturing device 110 records a video, where the duration of the video is T, and the user makes a key facial motion and at least one facial motion in the video. In addition, the video contains multiple primary facial patterns and collateral facial patterns; the primary facial image is a combination of key facial actions, and the secondary facial image is A collection of secondary facial movements. In addition, facial images can be obtained by sampling the video per unit time. As mentioned above, secondary facial movements may appear before, after, or before and after key facial movements. The processing unit 130 of the present invention is equipped to distinguish between key facial movements and secondary facial movements. Ability to move. For example, the processing unit 130 can distinguish a facial image as a main facial image or a subsidiary facial image by a predetermined threshold. Then in step 903, the processing unit 130 respectively annotates facial feature points in each of the main facial image and the subsidiary facial image. Finally, in step 904, the processing unit 130 creates a facial motion file for these labeled facial feature points.

在一實施例中,若第9圖的方法應用於建立預設臉部動作的基準檔案,則主要臉部圖像即為使用者做出預設臉部動作(例如:微笑)的動態臉部圖像;而附屬臉部圖像則為使用者面無表情時的空白臉部圖像。另外,本發明系統100可以儲存多個基準檔案以對應多個綁定功能。相關的說明以於835申請案中揭露,在此不贅述。 In one embodiment, if the method in Figure 9 is applied to create a reference file of preset facial motions, the main facial image is the dynamic face of the user that makes the default facial motions (for example, smiling) Image; while the attached face image is a blank face image when the user has no expression on his face. In addition, the system 100 of the present invention can store multiple reference files to correspond to multiple binding functions. The relevant explanation was disclosed in the 835 application, so I won’t repeat it here.

在另一實施例中,若第9圖的方法適用於建立對比臉部動作的檔案,則主要臉部圖像即為使用者做出對比臉部動作(例如:近似微笑)的臉部圖像;而附屬臉部圖像則為出現在對比臉部動作前、後或者前跟後的臉部動作。 In another embodiment, if the method in Figure 9 is suitable for creating a file for comparing facial motions, the main facial image is the facial image of the user who made the facial motions (for example, a similar smile). ; The subsidiary facial image is the facial motion that appears before, after, or before and after the contrast facial motion.

除上述外,一檔案的建立還可包含以下額外步驟(未見於第9圖)。首先,處理單元130基於主要臉部圖像及附屬臉部圖像計算每一個臉部特徵點的複數個量值。該量值代表了臉部特徵點在時間T上相對於一預設錨點的變化。該變化可以是位置,距離,角度,或以上任一的組合的變化。當取得於餘函數表示的量值後,處理單元130積分時間範圍T的餘函數以取得某一臉部特徵點的餘值。在此情況下,由於只有一個臉部特徵點被考慮,該餘值(即變化向量)即為該臉部動作的檔案。如果考慮多個臉部特徵點,該檔案即時所有臉部特徵點餘值的集合。 In addition to the above, the creation of a file can also include the following additional steps (not shown in Figure 9). First, the processing unit 130 calculates a plurality of magnitudes of each facial feature point based on the main facial image and the subsidiary facial image. The value represents the change of the facial feature point relative to a preset anchor point at time T. The change can be a change in position, distance, angle, or any combination of the above. After obtaining the value represented by the co-function, the processing unit 130 integrates the co-function in the time range T to obtain the residual value of a certain facial feature point. In this case, since only one facial feature point is considered, the residual value (that is, the change vector) is the file of the facial motion. If multiple facial feature points are considered, the file is a collection of residual values of all facial feature points in real time.

以上所有實施例中皆假設使用者身份已經驗證完畢。驗證的方法可以是任何習知方法或是835申請案中所揭露的方法。此外,即便使用者的身份並未事先驗證,由於本發明中利用使用者的臉部特徵點進行臉部動作的核實,因 此在核實的同時也可以一併完成使用者身分驗證。換句話說,在本發明其中一個實施例中,使用者驗證的步驟可以忽略。 In all the above embodiments, it is assumed that the user identity has been verified. The verification method can be any conventional method or the method disclosed in the 835 application. In addition, even if the user’s identity is not verified in advance, because the present invention uses the user’s facial feature points to verify facial movements, This verification can also complete the user identity verification at the same time. In other words, in one of the embodiments of the present invention, the user authentication step can be omitted.

雖然以上實施例通過一單一臉部動作來說明本發明的工作原理,在本發明其它實施例中亦可以採用複數個臉部動作。舉例來說,假設系統預設兩個連續的臉部動作(例如:微笑並眨眼)為進行一綁定操作的預設臉部動作,則微笑(第一臉部動作)的保護間隔可以是微笑發生前的面無表情動作和/或是在微笑之後發生的前幾個的眨眼(第二臉部動作)臉部圖像。另外,眨眼(第二臉部動作)的保護間隔可以前面微笑(第一臉部動作)的最後幾個臉部圖像和/或一在眨眼後的做出的臉部表情動作。亦或者,本發明系統與方法可以將微笑並眨眼當作同一個臉部動作並使用上述相同方法進行辨識。 Although the above embodiments illustrate the working principle of the present invention through a single facial motion, in other embodiments of the present invention, multiple facial motions can also be used. For example, suppose that the system presets two consecutive facial actions (for example, smiling and blinking) as the preset facial actions for a binding operation, and the guard interval of the smile (first facial action) may be smile The expressionless face action before the occurrence and/or the first few blinks (second facial action) face images that occurred after the smile. In addition, the guard interval for blinking (second facial motion) may be the last few facial images of a front smile (first facial motion) and/or a facial expression motion made after blinking. Or alternatively, the system and method of the present invention can treat smiling and blinking as the same facial action and use the same method for recognition.

以上所述僅為本發明之較佳實施例,凡依本發明申請專利範圍所做之均等變化與修飾,皆應屬本發明之涵蓋範圍。 The foregoing descriptions are only preferred embodiments of the present invention, and all equivalent changes and modifications made in accordance with the scope of the patent application of the present invention shall fall within the scope of the present invention.

100:識別系統 100: identification system

110:影像擷取裝置 110: Image capture device

120:資料庫 120: database

130:處理單元 130: processing unit

Claims (18)

一動態臉部識別裝置,包含;一資料庫,其用來儲存一使用者作出的一預設臉部動作的一基準檔案,其中該預設臉部動作被錄製成一第一視頻片段,該第一視頻片段包含複數個臉部圖像,而每張該臉部圖像至少包含該使用者至少一臉部特徵點以及該使用者臉上的一預設錨點;一影像擷取裝置,其用來追蹤該使用者的臉部動作並錄製成一第二視頻片段,其中該第二視頻片段包含該使用者在一時間內作出的一對比臉部動作;及一處理單元,連接至該資料庫以及該影像擷取裝置,該處理單元通過一臉部動作檔案建立步驟針對該預設臉部動作建立該基準檔案,以及針對該對比臉部動作建立一對比檔案,其中該處理單元將該對比檔案與儲存於該資料庫的基準檔案比較用以判斷該對比臉部動作是否即為該預設臉部動作。 A dynamic facial recognition device, including; a database for storing a reference file of a predetermined facial motion made by a user, wherein the predetermined facial motion is recorded as a first video segment, the first video segment A video segment includes a plurality of facial images, and each facial image includes at least one facial feature point of the user and a preset anchor point on the user's face; an image capturing device, which Used to track the user's facial movement and record it into a second video segment, where the second video segment includes a comparative facial movement made by the user in a time; and a processing unit connected to the database And the image capturing device, the processing unit creates the reference file for the preset facial motion through a facial motion file creation step, and creates a comparison file for the contrast facial motion, wherein the processing unit creates the comparison file It is compared with a reference file stored in the database to determine whether the contrast facial motion is the default facial motion. 如請求項1所述的動態臉部識別裝置,其中該臉部動作檔案建立步驟包含:該處理單元將該影像擷取裝置在一時間內拍攝的一視頻片段分類為複數個主要臉部圖像以及複數個附屬臉部圖像,其中該些主要臉部圖像由一關鍵臉部動作組成,而其中該些附屬臉部圖像則由一次要臉部動作組成。 The dynamic facial recognition device according to claim 1, wherein the step of creating a facial motion file includes: the processing unit classifies a video segment captured by the image capturing device within a time into a plurality of main facial images And a plurality of subsidiary facial images, where the main facial images are composed of a key facial action, and the subsidiary facial images are composed of a primary facial action. 如請求項2所述的動態臉部識別裝置,其中該次要臉部動作出現於在該關鍵臉部動作之前、之後、或者之前以及之後。 The dynamic face recognition device according to claim 2, wherein the secondary facial action occurs before, after, or before and after the key facial action. 如請求項2所述的動態臉部識別裝置,其中該關鍵臉部動作為該預設 臉部動作或者該對比臉部動作,而該次要臉部動作為一面無表情動作或者其他臉部動作。 The dynamic face recognition device according to claim 2, wherein the key facial action is the preset The facial motion or the contrast facial motion, and the secondary facial motion is an expressionless motion or other facial motions. 如請求項2所述的動態臉部識別裝置,其中該處理單元在該些主要臉部圖像以及該些附屬臉部圖像中找出該臉部特徵點。 The dynamic facial recognition device according to claim 2, wherein the processing unit finds the facial feature points in the main facial images and the subsidiary facial images. 如請求項5所述的動態臉部識別裝置,其中該處理單元針對該些臉部特徵點計算得到複數個量值,此外該處理單元更近一步將該些量值對該時間做數學積分以得到一餘值。 The dynamic face recognition device according to claim 5, wherein the processing unit calculates a plurality of values for the facial feature points, and the processing unit further performs mathematical integration of the values for the time to Get a residual value. 如請求項6所述的動態臉部識別裝置,其中該量值為該臉部特徵點在一時間點相對該預設錨點的相對位置。 The dynamic face recognition device according to claim 6, wherein the value is the relative position of the facial feature point with respect to the preset anchor point at a point in time. 如請求項1所述的動態臉部識別裝置,其中該預設臉部動作與一功能綁定,若該處理單元判斷該對比檔案與該基準檔案符合,則使得該功能被執行。 The dynamic face recognition device according to claim 1, wherein the preset facial action is bound to a function, and if the processing unit determines that the comparison file matches the reference file, the function is executed. 一種動態臉部識別方法,包含:透針對一使用者的一預設臉部動作建立一基準檔案;在一時間上擷取一視頻片段,其中在該視頻片段包含該使用者作出的至少一個近似臉部動作;針對該視頻片段建立一對比檔案;以及比較該對比檔案與該基準檔案以判斷該近似臉部動作是否符合該預設臉部動作, 其中將該預設臉部動作與一功能綁定,當該對比檔案以及該基準檔案相符時使得該功能被執行。 A dynamic face recognition method, including: creating a reference file based on a predetermined facial action of a user; capturing a video segment in time, wherein the video segment includes at least one approximation made by the user Facial action; create a comparison file for the video clip; and compare the comparison file with the reference file to determine whether the approximate facial motion matches the preset facial motion, The preset facial action is bound to a function, and the function is executed when the comparison file matches the reference file. 如請求項9所述的動態臉部識別方法,其中該基準檔案的建立是基於複數個主要臉部圖像以及複數個附屬臉部圖像,其中該些複數個主要臉部圖像為該使用者做出該預設臉部動作的連續動作圖像,而該些附屬臉部圖像為出現在該預設臉部動作之前、之後、或者之前和之後的面無表情圖像。 The dynamic face recognition method according to claim 9, wherein the establishment of the reference file is based on a plurality of main facial images and a plurality of subsidiary facial images, and the plurality of main facial images are used The person makes a continuous action image of the preset facial action, and the subsidiary facial images are expressionless images that appear before, after, or before and after the preset facial action. 如請求項9所述的動態臉部識別方法,其中該對比檔案的建立包含下列步驟:將該視頻片段分類成複數個關鍵臉部圖像以及複數個次要臉部圖像,且其中該些複數個關鍵臉部圖像為該使用者做出該近似臉部動作的圖像組合,而該些次要臉部圖像為該使用者於該近似臉部動作之前、之後、或者之前和之後做出的其他臉部動作組合;在每一個該些關鍵臉部圖像以及該些次要臉部圖像中標注出至少一個臉部特徵點;根據每個該些關鍵臉部圖像以及每個該些次要臉部圖像針對該臉部特徵點分別計算出一量值;以及將該些量值對該時間做數學積分得到該臉部特徵點的一餘值。 The dynamic face recognition method according to claim 9, wherein the establishment of the comparison file includes the following steps: classifying the video segment into a plurality of key facial images and a plurality of secondary facial images, and wherein The plurality of key facial images are the image combination of the user making the approximate facial motion, and the secondary facial images are the user before, after, or before and after the approximate facial motion Make other facial action combinations; mark at least one facial feature point in each of the key facial images and the secondary facial images; according to each of the key facial images and each Each of the secondary facial images calculates a magnitude for the facial feature point; and mathematically integrates the magnitudes over the time to obtain a residual value of the facial feature point. 如請求項11所述的動態臉部識別方法,其中該量值為該臉部特徵點在一時間點相對於一預設錨點的位置。 The dynamic face recognition method according to claim 11, wherein the value is the position of the facial feature point relative to a preset anchor point at a time. 一種電腦程式用來執行臉部動作識別功能,該電腦程式係儲存於一 非暫態的電腦可讀取儲存媒體中,而該電腦程式包含下列可執行下列指令的程式碼:透過一臉部檔案建立步驟針對一使用者的一預設臉部動作建立一基準檔案;取得該使用者做出的一近似臉部動作;透過該臉部檔案建立步驟針對該對比臉部動作建立一對比檔案;以及比較該對比檔案與該基準檔案以判斷該對比臉部動作是否符合該預設臉部動作。 A computer program used to perform facial motion recognition functions, the computer program is stored in a A non-transitory computer can read the storage medium, and the computer program contains the following code that can execute the following commands: create a reference file for a user's preset facial motion through a face file creation step; obtain An approximate facial action made by the user; a comparison file is created for the contrast facial motion through the facial file creation step; and the comparison file is compared with the reference file to determine whether the contrast facial motion meets the prediction Set facial movements. 如請求項13所述的電腦程式,其中該臉部檔案建立步驟包含下列指令步驟:在一時間內錄製一視頻片段,其中該視頻片段包含一關鍵臉部動作以及一次要臉部動作;以一單位時間取樣該視頻片段以取得複數個關鍵臉部圖像以及複數個次要臉部圖像;從每個該些關鍵臉部圖像以及該些次要臉部圖像中標注該使用者的至少一個臉部特徵點;針對該臉部特徵點在該單位時間內計算得到一量值;以及在該時間內數學積分該些量值以得到一餘值。 The computer program according to claim 13, wherein the face file creation step includes the following instruction steps: recording a video clip within a time, wherein the video clip includes a key facial movement and a primary facial movement; The video clip is sampled per unit time to obtain a plurality of key facial images and a plurality of secondary facial images; each of the key facial images and the secondary facial images is marked with the user’s At least one facial feature point; calculating a value for the facial feature point within the unit time; and mathematically integrating the values within the time to obtain a residual value. 如請求項14所述的電腦程式,其中該關鍵臉部動作為該預設臉部動作,而該次要臉部動作為發生在該預設臉部動作之前、之後、或者之前和之後的一面無表情動作;而該餘值為該基準檔案。 The computer program according to claim 14, wherein the key facial motion is the default facial motion, and the secondary facial motion is a side that occurs before, after, or before and after the default facial motion Expressionless action; and the residual value is the reference file. 如請求項14所述的電腦程式,其中該關鍵臉部動作為該近似臉部動 作,而該次要臉部動作為發生在該近似臉部動作之前、之後、或者之前和之後的一面無表情動作或者任何其他臉部動作;而該餘值為該對比檔案。 The computer program according to claim 14, wherein the key facial motion is the approximate facial motion The secondary facial motion is an expressionless motion or any other facial motion that occurs before, after, or before and after the approximate facial motion; and the residual value is the comparison file. 如請求項14所述的電腦程式,其中該量值為該臉部特徵點在每一個該單位時間相對於一預設錨點的位置。 The computer program according to claim 14, wherein the measurement value is the position of the facial feature point relative to a predetermined anchor point in each unit time. 如請求項13所述的電腦程式,其中使得該預設臉部動作與一功能綁定,若該對比檔案以及該基準檔案相符,則使得該功能被執行。 The computer program according to claim 13, wherein the default facial action is bound to a function, and if the comparison file matches the reference file, the function is executed.
TW108119853A 2019-02-26 2019-06-10 Dynamic face recognition system and method thereof TWI701607B (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US16/285,267 2019-02-26
US16/285,267 US10460454B2 (en) 2017-11-09 2019-02-26 Motion-based facial recognition

Publications (2)

Publication Number Publication Date
TWI701607B true TWI701607B (en) 2020-08-11
TW202032421A TW202032421A (en) 2020-09-01

Family

ID=72199287

Family Applications (1)

Application Number Title Priority Date Filing Date
TW108119853A TWI701607B (en) 2019-02-26 2019-06-10 Dynamic face recognition system and method thereof

Country Status (2)

Country Link
CN (1) CN111611833A (en)
TW (1) TWI701607B (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9202105B1 (en) * 2012-01-13 2015-12-01 Amazon Technologies, Inc. Image analysis for user authentication
TW201701192A (en) * 2015-06-30 2017-01-01 芋頭科技(杭州)有限公司 Facial recognition system and facial recognition method
US20170061970A1 (en) * 2015-08-25 2017-03-02 Malaspina Labs (Barbados), Inc. Speaker Dependent Voiced Sound Pattern Detection Thresholds

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2014116033A (en) * 2008-06-16 2014-06-26 Canon Inc Personal authentication device, personal authentication method, camera, program, and storage medium
CN105993022B (en) * 2016-02-17 2019-12-27 香港应用科技研究院有限公司 Method and system for recognition and authentication using facial expressions
US10783352B2 (en) * 2017-11-09 2020-09-22 Mindtronic Ai Co., Ltd. Face recognition system and method thereof

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9202105B1 (en) * 2012-01-13 2015-12-01 Amazon Technologies, Inc. Image analysis for user authentication
TW201701192A (en) * 2015-06-30 2017-01-01 芋頭科技(杭州)有限公司 Facial recognition system and facial recognition method
US20170061970A1 (en) * 2015-08-25 2017-03-02 Malaspina Labs (Barbados), Inc. Speaker Dependent Voiced Sound Pattern Detection Thresholds

Also Published As

Publication number Publication date
CN111611833A (en) 2020-09-01
TW202032421A (en) 2020-09-01

Similar Documents

Publication Publication Date Title
US9985963B2 (en) Method and system for authenticating liveness face, and computer program product thereof
US11386900B2 (en) Visual speech recognition by phoneme prediction
CN106557723B (en) Face identity authentication system with interactive living body detection and method thereof
WO2019091012A1 (en) Security check method based on facial recognition, application server, and computer readable storage medium
US9208375B2 (en) Face recognition mechanism
JP3335538B2 (en) Method and apparatus for collating a handwritten character string
WO2015131468A1 (en) Method and system for estimating fingerprint pose
CN109766785A (en) A kind of biopsy method and device of face
KR102062708B1 (en) Apparatus and method for detection presentaion attack for face recognition system
Omoyiola Overview of biometric and facial recognition techniques
Siirtola et al. Effect of context in swipe gesture-based continuous authentication on smartphones
TWI526953B (en) Face recognition method and system
KR102366777B1 (en) Apparatus and method for domain adaptation-based object recognition
TWI701607B (en) Dynamic face recognition system and method thereof
JP2011221699A (en) Operation instruction recognition device and robot
Huang et al. A new dataset for smartphone gesture-based authentication
US11017198B2 (en) Slap segmentation of contactless fingerprint images
US9613252B1 (en) Fingerprint matching method and device
KR20120132337A (en) Apparatus and Method for Controlling User Interface Using Sound Recognition
CN111401167A (en) Method for realizing biological recognition processing based on face recognition and palm vein recognition and corresponding system
US20230154236A1 (en) Landmark-based ensemble network creation method for facial expression classification and facial expression classification method using created ensemble network
KR101972231B1 (en) Apparatus and method of signature authentication using deep learning
US10460454B2 (en) Motion-based facial recognition
CN112149559A (en) Face recognition method and device, readable storage medium and computer equipment
Putz-Leszczynska et al. Gait biometrics with a Microsoft Kinect sensor