[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN103729462B - A kind of pedestrian retrieval method blocked based on rarefaction representation process - Google Patents

A kind of pedestrian retrieval method blocked based on rarefaction representation process Download PDF

Info

Publication number
CN103729462B
CN103729462B CN201410014852.6A CN201410014852A CN103729462B CN 103729462 B CN103729462 B CN 103729462B CN 201410014852 A CN201410014852 A CN 201410014852A CN 103729462 B CN103729462 B CN 103729462B
Authority
CN
China
Prior art keywords
image
sparse representation
pedestrian
image block
image blocks
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201410014852.6A
Other languages
Chinese (zh)
Other versions
CN103729462A (en
Inventor
胡瑞敏
王亦民
梁超
黄冰月
郑淇
陈军
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan University WHU
Original Assignee
Wuhan University WHU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan University WHU filed Critical Wuhan University WHU
Priority to CN201410014852.6A priority Critical patent/CN103729462B/en
Publication of CN103729462A publication Critical patent/CN103729462A/en
Application granted granted Critical
Publication of CN103729462B publication Critical patent/CN103729462B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/7837Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using objects detected or recognised in the video content
    • G06F16/784Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using objects detected or recognised in the video content the detected or recognised objects being people
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Library & Information Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Image Analysis (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention discloses a kind of pedestrian retrieval method processing based on rarefaction representation and blocking, the present invention obtains the distance metric between inquiry pedestrian and pedestrian to be measured initially with block-based similarity coupling, then the method using rarefaction representation obtains the coverage extent between inquiry pedestrian and pedestrian to be measured, last comprehensive distance tolerance and coverage extent calculate pedestrian between similarity, improve the accuracy with a group traveling together's coupling under multi-cam.

Description

Pedestrian retrieval method based on sparse representation processing shielding
Technical Field
The invention relates to the technical field of surveillance video retrieval, in particular to a pedestrian retrieval method based on sparse representation processing occlusion.
Background
Surveillance video pedestrian retrieval is a technique to match specific pedestrian objects under multiple cameras with no overlap of illumination areas. In the actual video investigation, an investigator mainly locks, inspects and tracks a suspected target quickly according to a moving picture and a track of a same pedestrian object. The traditional video detection mode of manual browsing needs to consume a large amount of manpower and time, and the time for solving a case is easy to be delayed. The pedestrian re-identification technology is convenient for a video inspector to quickly and accurately find moving pictures and tracks of suspected targets, and has important significance for improving the case rate of a public security department and maintaining the life and property safety of people.
The existing pedestrian retrieval (also called pedestrian re-identification) method can be divided into two categories:
the first category mainly constructs robust visual features and then performs similarity measurement using standard distance functions (e.g., euclidean distance, etc.). For example, a pedestrian re-identification method based on multiple local feature matching of symmetric segmentation, firstly, a body is segmented horizontally and vertically by using color feature clues; secondly, extracting various color and texture characteristics of each region, and weighting the visual characteristics based on a horizontal central axis; finally, the characteristics are comprehensively used for representing and matching the object;
the second type has no strict requirement on feature construction, more accurate distance measurement is carried out mainly by learning a proper scale, difference vectors of samples of the same type and difference vectors of different samples are respectively expressed into different Gaussian distributions, then the distance between the samples is measured by using the ratio of probability, and finally the ratio of the Gaussian distributions is converted into a form of Mahalanobis distance, so that a proper Mahalanobis distance function is learned.
The method is to sort the to-be-detected set according to the distances of the appearance features of the pedestrian object to be inquired and all the to-be-detected pedestrian objects, and does not consider the self-shielding and other pedestrian or object shielding conditions caused by the visual angle transformation under different cameras. However, in an actual video monitoring environment, the same pedestrian often has a shielding situation under multiple cameras, so that the appearance features have significant differences, and the retrieval result is inaccurate.
Disclosure of Invention
Aiming at the defects of the existing method, the invention provides a pedestrian retrieval method based on sparse representation processing shielding, and the accuracy of matching of the same pedestrian under multiple cameras is improved.
In order to achieve the purpose, the invention adopts the following technical scheme: a pedestrian retrieval method for processing occlusion based on sparse representation is characterized by comprising the following steps:
step 1: dividing an inquiry pedestrian image P and a pedestrian image Q to be detected into m rows and n columns of small image blocks, and representing the P and the Q as a set of image blocks, namely P = { P = { P =ij|i=1,…,m;j=1,…,n},Q={QijI =1, …, m, j =1, …, n }, wherein m is greater than or equal to 1, and n is greater than or equal to 1;
step 2: extracting the characteristics of each image block of the P and the Q, and expressing the P and the Q by the characteristics based on the image blocks;
and step 3: carrying out similarity matching based on the image blocks on the P and the Q to obtain a distance measurement result of the P and the Q based on the similarity matching of the image blocks;
and 4, step 4: calculating sparse representation of the P image blocks relative to the Q, and obtaining the shielding degree of the P relative to the Q; calculating sparse representation of the image block of Q relative to the P, and obtaining the shielding degree of Q relative to the P;
and 5: and calculating the similarity between P and Q according to the distance measurement result of matching P and Q based on the similarity of the image blocks in the step 3 and the shielding degree between P and Q in the step 4.
Preferably, the features of the image block described in step 2 are grayscale, color, and SIFT features.
Preferably, the obtaining of the result of the distance metric based on similarity matching of P and Q in step 3 includes the following steps:
step 3.1: determining a search area, determining a search area for each image block in the P image block set, and setting the P image block as PabThen P isabThe search area consists of a plurality of image blocks in the image block set of Q, and the image block PabThe search area of (a) is represented as:
step 3.2: calculating PabTo the search areaThe Euclidean distance of each image block, the block with the minimum distance is the distance PabThe most similar block, which is denoted as:then the distance measurement result of P and Q based on similarity matching of image blocks is: d ( P , Q ) = Σ m , n d ( P ab , Q ab * ) .
preferably, said computing a sparse representation of said P image block relative to said Q in step 4 is implemented by: let the image block of P be PabThen P isabSparse representation with respect to said Q:
min β Pab | | V p ab - D P ab Q β P ab | | 2 2 + λ | | β P ab | | 1
the presentation error is:whereinIs PabIs determined by the feature vector of (a),is PabThe sparse coefficient vector of (a) is,is PabA dictionary of, andλ is a parameter of the sparse representation model;
the step 4 of calculating the sparse representation of the image block of Q relative to P specifically implements the following process: let the image block of Q be QabThen Q isabSparse representation with respect to said P:
min β Q ab * | | V Q ab * - D Q ab * P β Q ab * | | 2 2 + λ | | β Q ab * | | 1
the presentation error is:whereinIs composed ofIs determined by the feature vector of (a),is composed ofThe sparse coefficient vector of (a) is,is composed ofA dictionary of, andλ is a parameter of the sparse representation model.
Preferably, the similarity between P and Q is calculated by the following formula:
sim ( P , Q ) = Σ m , n ρ ( P ab , Q ) · sim ( P ab , Q ab * ) · ρ ( Q ab * , P )
wherein, sim ( P ab , Q ab * ) = exp ( - d ( P ab , Q ab * ) 2 2 σ 2 ) , ρ ( P ab , Q ) = exp ( - ( E P ab Q ) 2 ) , sigma is the bandwidth of the gaussian function,based on image blocks P for said P and QabAndthe greater sim (P, Q), the more similar said P and said Q.
The invention has the following advantages and positive effects:
(1) compared with the prior art, the method introduces the idea of processing the shielding based on sparse representation, calculates the similarity between the inquired pedestrian and the pedestrian to be detected through the distance and the shielding degree based on the similarity matching of the blocks, and can obtain a more accurate pedestrian retrieval result;
(2) the method not only considers the mismatching caused by the posture change, but also considers the influence of the shielding on similarity calculation, and is more robust to the shielding caused by the visual angle change.
Drawings
FIG. 1: is a method flow diagram of an embodiment of the invention.
FIG. 2: is a technical scheme schematic diagram of the embodiment of the invention.
Detailed Description
The technical solution of the present invention is further explained by the following embodiments with reference to the accompanying drawings.
Referring to fig. 2, the embodiment provided by the present invention uses MATLAB7 as a simulation experiment platform to perform tests on a commonly used pedestrian retrieval dataset VIPeR. The VIPeR data set has 632 pedestrian image pairs under two cameras, and obvious differences of visual angles, illumination and the like exist between the cameras.
Referring to fig. 1, the technical scheme adopted by the invention is as follows: a pedestrian retrieval method based on sparse representation processing occlusion comprises the following steps:
step 1: dividing the inquiry pedestrian image P and the pedestrian image Q to be detected into m rows and n columns of small image blocks, and representing P and Q as a set of image blocks, namely P = { P = { (P)ij|i=1,…,m;j=1,…,n},Q={QijI =1, …, m, j =1, …, n }, where m > 1, n > 1, the image size is 128 × 48 pixels, the image is partitioned by a window of 10 × 10 pixels with a step size of 4, each image is divided into 30 × 10 image blocks.
Step 2: extracting the characteristics of each image block of P and Q, wherein the characteristics of the image blocks are the gray scale, color and SIFT characteristics, and the P and Q are expressed by the characteristics based on the image blocks; the gray scale feature is 100 dimensions, the color feature is represented as 672 dimensions by using LAB and SIFT feature, and the feature of all blocks in each image is added to represent the appearance feature of a pedestrian object in one image.
And step 3: similarity matching based on the image blocks is carried out on the P and the Q, and a distance measurement result of the P and the Q based on the similarity matching of the image blocks is obtained; the specific implementation comprises the following substeps:
step 3.1: determining a search area, determining a search area for each image block in the P image block set, and setting the P image block as PabThen P isabThe search area of (A) is composed of a plurality of image blocks in the image block set of (Q), the image block (P)abThe search area of (a) is represented as: S P ab Q = { Q ij | i = a - 1 , . . . , a . . . , a + 1 ; j = 1 , . . . , n } , take l =1 as an example;
step 3.2: calculating PabTo the search areaThe Euclidean distance of each image block, the block with the minimum distance is the distance PabThe most similar block, which is denoted as:the distance metric result of P and Q matching based on similarity of image blocks is: d ( P , Q ) = Σ m , n d ( P ab , Q ab * ) .
and 4, step 4: calculating sparse representation of an image block of P relative to Q, and obtaining the shielding degree of P relative to Q; calculating sparse representation of the image block of Q relative to P, and obtaining the shielding degree of Q relative to P;
sparse representation of image blocks in which P is computed relative to Q, which is concreteThe current process is as follows: let P's image block be PabThen P isabSparse representation with respect to Q:
min β Pab | | V p ab - D P ab Q β P ab | | 2 2 + λ | | β P ab | | 1
the presentation error is:whereinIs PabIs determined by the feature vector of (a),is PabThe sparse coefficient vector of (a) is,is PabA dictionary of, andλ is a parameter of the sparse representation model;
the sparse representation of the image block of Q relative to P is calculated, and the specific implementation process is as follows: let the image block of Q be QabThen Q isabSparse representation with respect to P:
min β Q ab * | | V Q ab * - D Q ab * P β Q ab * | | 2 2 + λ | | β Q ab * | | 1
the presentation error is:whereinIs composed ofIs determined by the feature vector of (a),is composed ofThe sparse coefficient vector of (a) is,is composed ofA dictionary of, andλ is a parameter of the sparse representation model.
And 5: calculating the similarity between P and Q according to the distance measurement result of the similarity matching of the image blocks in the step 3P and the step Q and the shielding degree between the step 4P and the step Q; calculating the similarity between P and Q, and obtaining by adopting the following formula:
sim ( P , Q ) = Σ m , n ρ ( P ab , Q ) · sim ( P ab , Q ab * ) · ρ ( Q ab * , P )
wherein, sim ( P ab , Q ab * ) = exp ( - d ( P ab , Q ab * ) 2 2 σ 2 ) , ρ ( P ab , Q ) = exp ( - ( E P ab Q ) 2 ) , sigma is the bandwidth of the gaussian function,based on image blocks P for P and QabAndthe greater sim (P, Q), the more similar P and Q.
The pedestrian retrieval evaluation index uses a CMC value, and the CMC value is the probability that the correct pedestrian object exists in the previous r results returned in N times of queries. When the former r results are returned, the higher the CMC value, the better the pedestrian retrieval performance is indicated. The test procedure of this example was repeated 10 times and the average CMC value was calculated. The method for identifying pedestrians based on characteristics by the ELF is compared with the method for processing the shielded result by sparse representation. The specific results are shown in Table 1, which are CMC values when the previous 1, 10, 25 and 50 results are returned on the VIPeR;
table 1: CMC values at return of the first 1, 10, 25, 50 results on VIPeR
Method of producing a composite material Rank@1 10 25 50
ELF 12 31 41 58
SDALF 20 39 49 66
KISS 20 49 62 78
Sparse representation handling occlusion 31 50 65 79
From comparison of CMC values in Table 1, it can be found that the performance of the pedestrian re-identification method based on sparse representation processing occlusion provided by the invention is obviously superior to that of a comparison algorithm.
According to the method, firstly, the distance measurement between the query pedestrian and the pedestrian to be detected is obtained through block-based similarity matching, then the shielding degree between the query pedestrian and the pedestrian to be detected is obtained through a sparse representation method, and finally the similarity between pedestrian pairs is calculated through the distance measurement and the shielding degree, so that the matching accuracy of the same pedestrian under multiple cameras is improved.
The present invention is not limited to the above embodiments, and any modifications, equivalent replacements, improvements, etc. within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (4)

1. A pedestrian retrieval method for processing occlusion based on sparse representation is characterized by comprising the following steps:
step 1: dividing the inquiry pedestrian image P and the pedestrian image Q to be detected into m rows and n columns of small image blocks, and expressing the P and the Q as a set of image blocks, namely P ═ { P ═ P }ij|i=1,…,m;j=1,…,n},Q={Qij1, | i ═ …, m; j is 1, …, n, wherein m is more than or equal to 1, n is more than or equal to 1;
step 2: extracting the characteristics of each image block of the P and the Q, and expressing the P and the Q by the characteristics based on the image blocks;
and step 3: carrying out similarity matching based on the image blocks on the P and the Q to obtain a distance measurement result of the P and the Q based on the similarity matching of the image blocks; the specific implementation comprises the following substeps:
step 3.1: determining a search area, determining a search area for each image block in the P image block set, and setting the P image block as PabThen P isabThe search area consists of a plurality of image blocks in the image block set of Q, and the image block PabThe search area of (a) is represented as:
step 3.2: calculating PabTo the search areaThe Euclidean distance of each image block, the block with the minimum distance is the distance PabThe most similar block, which is denoted as:then the distance measurement result of P and Q based on similarity matching of image blocks is:
and 4, step 4: calculating sparse representation of the P image blocks relative to the Q, and obtaining the shielding degree of the P relative to the Q; calculating sparse representation of the image block of Q relative to the P, and obtaining the shielding degree of Q relative to the P;
and 5: and calculating the similarity between P and Q according to the distance measurement result of matching P and Q based on the similarity of the image blocks in the step 3 and the shielding degree between P and Q in the step 4.
2. The pedestrian retrieval method based on sparse representation processing occlusion of claim 1, wherein: the characteristics of the image block in the step 2 are gray scale, color and SIFT characteristics.
3. The pedestrian retrieval method based on sparse representation processing occlusion of claim 1, wherein: the step 4 of calculating the sparse representation of the image block of P relative to Q specifically implements the following process: let the image block of P be PabThen P isabSparse representation with respect to said Q:
m i n β P a b | | V p a b - D P a b Q β P a b | | 2 2 + λ | | β P a b | | 1
the presentation error is:whereinIs PabIs determined by the feature vector of (a),is PabThe sparse coefficient vector of (a) is,is PabA dictionary of, andλ is a parameter of the sparse representation model;
the step 4 of calculating the sparse representation of the image block of Q relative to P specifically implements the following process: let the image block of Q be QabThen Q isabSparse representation with respect to said P:
m i n β Q a b * | | V Q a b * - D Q a b * P β Q a b * | | 2 2 + λ | | β Q a b * | | 1
the presentation error is:whereinIs composed ofIs determined by the feature vector of (a),is composed ofThe sparse coefficient vector of (a) is,is composed ofA dictionary of, andλ is a parameter of the sparse representation model.
4. The pedestrian retrieval method based on sparse representation processing occlusion of claim 3, wherein: the similarity between P and Q is calculated by adopting the following formula:
s i m ( P , Q ) = Σ m , n ρ ( P a b , Q ) · s i m ( P a b , Q a b * ) · ρ ( Q a b * , P )
wherein,sigma is the bandwidth of the gaussian function,based on image blocks P for said P and QabAnd Q* abThe greater sim (P, Q), the more similar said P and said Q.
CN201410014852.6A 2014-01-13 2014-01-13 A kind of pedestrian retrieval method blocked based on rarefaction representation process Expired - Fee Related CN103729462B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410014852.6A CN103729462B (en) 2014-01-13 2014-01-13 A kind of pedestrian retrieval method blocked based on rarefaction representation process

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410014852.6A CN103729462B (en) 2014-01-13 2014-01-13 A kind of pedestrian retrieval method blocked based on rarefaction representation process

Publications (2)

Publication Number Publication Date
CN103729462A CN103729462A (en) 2014-04-16
CN103729462B true CN103729462B (en) 2016-09-14

Family

ID=50453536

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410014852.6A Expired - Fee Related CN103729462B (en) 2014-01-13 2014-01-13 A kind of pedestrian retrieval method blocked based on rarefaction representation process

Country Status (1)

Country Link
CN (1) CN103729462B (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104200493B (en) * 2014-09-05 2017-02-01 武汉大学 Similarity measurement based real-time target tracking algorithm
CN104200206B (en) * 2014-09-09 2017-04-26 武汉大学 Double-angle sequencing optimization based pedestrian re-identification method
CN104298992B (en) * 2014-10-14 2017-07-11 武汉大学 A kind of adaptive scale pedestrian recognition methods again based on data-driven
CN104715071B (en) * 2015-04-02 2017-10-03 武汉大学 A kind of specific pedestrian retrieval method described based on imperfect text
CN107844752A (en) * 2017-10-20 2018-03-27 常州大学 A kind of recognition methods again of the pedestrian based on block rarefaction representation
CN110826417B (en) * 2019-10-12 2022-08-16 昆明理工大学 Cross-view pedestrian re-identification method based on discriminant dictionary learning

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6947933B2 (en) * 2003-01-23 2005-09-20 Verdasys, Inc. Identifying similarities within large collections of unstructured data
CN101141633A (en) * 2007-08-28 2008-03-12 湖南大学 Moving object detecting and tracing method in complex scene
CN102521616A (en) * 2011-12-28 2012-06-27 江苏大学 Pedestrian detection method on basis of sparse representation
CN102667815A (en) * 2009-10-02 2012-09-12 高通股份有限公司 Methods and systems for occlusion tolerant face recognition
CN102855496A (en) * 2012-08-24 2013-01-02 苏州大学 Method and system for authenticating shielded face
CN103049749A (en) * 2012-12-30 2013-04-17 信帧电子技术(北京)有限公司 Method for re-recognizing human body under grid shielding

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6947933B2 (en) * 2003-01-23 2005-09-20 Verdasys, Inc. Identifying similarities within large collections of unstructured data
CN101141633A (en) * 2007-08-28 2008-03-12 湖南大学 Moving object detecting and tracing method in complex scene
CN102667815A (en) * 2009-10-02 2012-09-12 高通股份有限公司 Methods and systems for occlusion tolerant face recognition
CN102521616A (en) * 2011-12-28 2012-06-27 江苏大学 Pedestrian detection method on basis of sparse representation
CN102855496A (en) * 2012-08-24 2013-01-02 苏州大学 Method and system for authenticating shielded face
CN103049749A (en) * 2012-12-30 2013-04-17 信帧电子技术(北京)有限公司 Method for re-recognizing human body under grid shielding

Also Published As

Publication number Publication date
CN103729462A (en) 2014-04-16

Similar Documents

Publication Publication Date Title
Li et al. Adaptively constrained dynamic time warping for time series classification and clustering
CN103729462B (en) A kind of pedestrian retrieval method blocked based on rarefaction representation process
CN104298992B (en) A kind of adaptive scale pedestrian recognition methods again based on data-driven
CN108388896B (en) License plate identification method based on dynamic time sequence convolution neural network
CN102722892B (en) SAR (synthetic aperture radar) image change detection method based on low-rank matrix factorization
CN102819740B (en) A kind of Single Infrared Image Frame Dim targets detection and localization method
CN109544592B (en) Moving object detection algorithm for camera movement
CN103793702A (en) Pedestrian re-identifying method based on coordination scale learning
CN104200471B (en) SAR image change detection based on adaptive weight image co-registration
CN108765470A (en) One kind being directed to the improved KCF track algorithms of target occlusion
CN104732546B (en) The non-rigid SAR image registration method of region similitude and local space constraint
CN105787943B (en) SAR image registration method based on multi-scale image block feature and rarefaction representation
CN108447057A (en) SAR image change detection based on conspicuousness and depth convolutional network
CN104182985A (en) Remote sensing image change detection method
CN105469111A (en) Small sample set object classification method on basis of improved MFA and transfer learning
CN103886337A (en) Nearest neighbor subspace SAR target identification method based on multiple sparse descriptions
CN103500345A (en) Method for learning person re-identification based on distance measure
CN108171119B (en) SAR image change detection method based on residual error network
CN106557740A (en) The recognition methods of oil depot target in a kind of remote sensing images
CN104732552B (en) SAR image segmentation method based on nonstationary condition
CN103824302A (en) SAR (synthetic aperture radar) image change detecting method based on direction wave domain image fusion
CN106682278A (en) Supersonic flow field predicting accuracy determination device and method based on image processing
CN103500453A (en) SAR(synthetic aperture radar) image significance region detection method based on Gamma distribution and neighborhood information
CN104182768B (en) The quality classification method of ISAR image
CN102254185B (en) Background clutter quantizing method based on contrast ratio function

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20160914

Termination date: 20220113

CF01 Termination of patent right due to non-payment of annual fee