CN100525395C - Pedestrian tracting method based on principal axis marriage under multiple vedio cameras - Google Patents
Pedestrian tracting method based on principal axis marriage under multiple vedio cameras Download PDFInfo
- Publication number
- CN100525395C CN100525395C CNB200510108137XA CN200510108137A CN100525395C CN 100525395 C CN100525395 C CN 100525395C CN B200510108137X A CNB200510108137X A CN B200510108137XA CN 200510108137 A CN200510108137 A CN 200510108137A CN 100525395 C CN100525395 C CN 100525395C
- Authority
- CN
- China
- Prior art keywords
- main shaft
- people
- under
- value
- camera
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
- 238000000034 method Methods 0.000 title claims abstract description 47
- 238000001514 detection method Methods 0.000 claims abstract description 20
- 230000033001 locomotion Effects 0.000 claims abstract description 12
- 230000000007 visual effect Effects 0.000 claims description 45
- 230000008878 coupling Effects 0.000 claims description 31
- 238000010168 coupling process Methods 0.000 claims description 31
- 238000005859 coupling reaction Methods 0.000 claims description 31
- 230000000903 blocking effect Effects 0.000 claims description 12
- 239000011159 matrix material Substances 0.000 claims description 12
- 230000011218 segmentation Effects 0.000 claims description 7
- 238000001914 filtration Methods 0.000 claims description 5
- 238000007476 Maximum Likelihood Methods 0.000 claims description 4
- 239000000284 extract Substances 0.000 claims description 4
- 230000015572 biosynthetic process Effects 0.000 claims description 2
- 238000006243 chemical reaction Methods 0.000 claims description 2
- 230000007797 corrosion Effects 0.000 claims description 2
- 238000005260 corrosion Methods 0.000 claims description 2
- 230000000877 morphologic effect Effects 0.000 claims description 2
- 238000012545 processing Methods 0.000 claims description 2
- 230000004927 fusion Effects 0.000 claims 1
- 238000012360 testing method Methods 0.000 description 8
- 238000012544 monitoring process Methods 0.000 description 7
- 230000014509 gene expression Effects 0.000 description 4
- 210000003141 lower extremity Anatomy 0.000 description 3
- 238000011160 research Methods 0.000 description 3
- 238000011161 development Methods 0.000 description 2
- 230000018109 developmental process Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000002474 experimental method Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 238000005286 illumination Methods 0.000 description 2
- 238000005303 weighing Methods 0.000 description 2
- 206010000117 Abnormal behaviour Diseases 0.000 description 1
- 230000002159 abnormal effect Effects 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 239000004744 fabric Substances 0.000 description 1
- 230000013011 mating Effects 0.000 description 1
- 230000002265 prevention Effects 0.000 description 1
- 230000009017 pursuit movement Effects 0.000 description 1
Images
Landscapes
- Image Analysis (AREA)
- Closed-Circuit Television Systems (AREA)
Abstract
The method comprises: making motion detection for sequence image; extracting the feature of main axis; making trace under single video camera; according to the main axis matching function, building main axis matching pair; combining multi view angles information to optimize and renew the trace result.
Description
Technical field
The present invention relates to technical field of computer vision, particularly the multiple-camera matching process.
Background technology
Along with the development of society, it is important that safety problem becomes day by day.Vision monitoring becomes the requirement in a lot of large-scale places.These places comprise security department, the military affairs that country is important.Meanwhile in a lot of public places, it is most important that vision monitoring also becomes, and wherein typical application is a field of traffic.Because traditional visual monitor system all is a kind of monitoring afterwards, all accidents and abnormal conditions all are that the security personnel finds at that time or afterwards, and the behavior of various breach securitys is not had the prevention effect; Therefore traditional supervisory control system can not adapt to demands of social development.At this moment, intelligentized visual monitor system becomes the requirement in epoch and researcher's target.Intelligentized vision monitoring final purpose is to realize the behavior of Real time identification moving target, thereby abnormal behaviour is in time made judgement and warning.And depend on bottom layer treatment, i.e. the tracking of moving target with the solution of analyzing the contour level problem as the behavior identification of moving target.In monitoring scene, people's motion is the focus of our concern often.In addition, people's motion is a kind of non-movement of Rigid Body, and it is complicated more that other motion tracking (such as the motion tracking of vehicle) is compared in its tracking.Therefore people's tracking becomes vital part of Intellectualized monitoring.
Nowadays, people are more and more higher to the intellectuality and the security requirement of visual monitor system, use single camera that the people who moves is followed the tracks of far away and can not satisfy needs of society.In the last few years, utilized multiple-camera to come the people of pursuit movement to become the research focus.
Multiple-camera servant's tracking is the problem of a multiple-camera coupling in essence, promptly sets up the corresponding relation between the moving object under the different visual angles (video camera) at synchronization.The multiple-camera coupling is newer problem of computer vision.In this respect, abroad some universities and research institution (Maryland, Oxford MIT) have carried out some research work.Type according to selected feature can roughly be divided into related work two big classes: based on the method in zone with based on the method for characteristic point.
Matching process based on the zone is that the people is regarded as a moving region, utilizes the feature of moving region to set up different visual angles servant's corresponding relation.In these class methods, color is a most frequently used provincial characteristics.Orwell etc. and Krumm etc. utilize color histogram to estimate that people's field color distributes, and set up the coupling of multiple-camera by comparing color histogram; Mittal etc. have set up Gauss's color model for people's moving region, utilize these modelling couplings then; Another method is to utilize the distribution of cuclear density Function Estimation field color, sets up coupling on this basis.Though color characteristic is more directly perceived, however its robust not very aspect coupling.This be because, depend on the color of people's clothes based on the coupling of color, when two people's clothes color is identical, may produce wrong coupling; Secondly, owing to be subjected to the influence at illumination and visual angle, observed same individual's clothes color may change.
Method based on characteristic point is to regard the people as a series of characteristic point, and different visual angles servant's coupling just is converted into the coupling of characteristic point.The coupling of characteristic point is based on certain geometrical constraint.According to the difference of selected geometrical constraint, these class methods can be divided into two subclasses: three-dimensional method and two-dimension method.A.Utsumi chooses the barycenter of moving object as characteristic point, compares their tripleplane's point then; Q.Cai chooses some characteristic points of people's upper part of the body center line, and limit of utilization retrains and seeks coupling then.Three-dimensional method all needs video camera is demarcated.When employed video camera number is very big in the monitoring scene, camera calibration will be a huger task.In order to overcome the shortcoming of three-dimensional method, some researcher has proposed to utilize two-dimensional signal to set up coupling between the multiple-camera.Khan etc. utilize the constraint (homography constraint) of singly reflecting of ground level to come the characteristic point on the match people pin.But these characteristic points are subjected to The noise easily.Block or to detect effect bad when existence, when the people had only part to be observed, the characteristic point of extraction is robust not, and therefore the performance of coupling will variation.
Worldwide obtained paying close attention to widely and studying though it is emphasized that multiple-camera servant's tracking, still had many difficult points aspect the multiple-camera matching technique.Therefore how to choose more robust, feature is mated and is still a challenge accurately.
Summary of the invention
The objective of the invention is to avoid conventional method to need calibration, adopt feature to be subject to shortcomings such as noise and visual angle factor affecting, provide a kind of multiple-camera matching process simple, robust to be used for multiple-camera servant's tracking.
For achieving the above object, the pedestrian tracting method based on the main shaft coupling comprises step under the multiple-camera:
(1) moving Object Segmentation;
(2) extraction people's main shaft feature;
(3) tracking under the single camera;
(4) it is right to seek all optimum Match according to main shaft match likelihood function;
(5) merge various visual angles information updating tracking results.
Multiple-camera is meant that using two or more single cameras, the precondition of this method is to have a public plane in the viewing area of hypothesis different cameras, and generally speaking, this public plane is meant ground level.
The present invention extracts people's main shaft as feature, is a kind of novel multiple-camera matching process based on main shaft, has good application prospects.
Description of drawings
Fig. 1 is single people's main shaft detection example figure.
Fig. 2 is group's main shaft detection example figure.
Fig. 3 is the main shaft detection example figure of the people under the situation of blocking.
Fig. 4 is single camera servant's a tracking block diagram.
Fig. 5 is multiple-camera servant's main shaft projection relation figure.
Fig. 6 (a) is many people tracking test figure as a result under two video cameras of NLPR database.
Fig. 6 (b) is many people tracking test figure as a result under three video cameras of NLPR database.
Fig. 6 (c) is to the tracking test of single people under the situation of blocking figure as a result under two video cameras of PETS2001 database.
Fig. 6 (d) is to the tracking test of group under the situation of blocking figure as a result under two video cameras of PETS2001 database.
Fig. 7 is based on the general introduction figure of the pedestrian tracting method of main shaft coupling under the multiple-camera.
Embodiment
Main feature of the present invention is: 1) extract the feature of people's main shaft as coupling.Because people's main shaft is the symmetry axis of human region, according to symmetry, the point that is symmetrically distributed in the main shaft both sides their error of can cancelling each other, thus make main shaft robust more.And people's main shaft is subjected to motion detection and the result of cutting apart to influence less; 2) detection method of three kinds of different situations servants' main shaft has been proposed: the detection that promptly single people's main shaft detects, group's main shaft detects and blocks the main shaft of the people under the situation; 3) based on the geometrical relationship constraint of different visual angles lower main axis, defined main shaft match likelihood function and weighed the right similarity of different visual angles lower main axis; 4) merge various visual angles information according to matching result, tracking results is optimized renewal.
The general frame of scheme is seen accompanying drawing 7.At first the image sequence under the single camera is carried out motion detection, extract people's main shaft feature, carry out the tracking under the single camera; Then according to singly reflect relation constraint to the main shaft under the different visual angles to mating; Merge various visual angles information updating tracking results according to matching result at last.
Provide the explanation of each related in this invention technical scheme detailed problem below in detail.
(1) moving Object Segmentation
Moving Object Segmentation is the first step of motion tracking, and algorithm adopts is that the background method of wiping out of single Gauss model detects the moving region.In order to reduce the influence of illumination variation and shade, adopt normalized color rgs model, r=R/ (R+G+B) wherein, g=G/ (R+G+B), s=R+G+B.Earlier one section background image sequence is carried out filtering, obtain single Gauss's background model.Gauss's parameter of each point is
U wherein
iAnd σ
i(i=r, g s) are the average and the variance of this background model respectively.Then present image and background image are carried out difference processing, can obtain binary image M in that difference image is carried out threshold process
Xy, the foregrounding pixel value is 1 in bianry image, background pixel value is 0.Its process is:
I wherein
i(x, y), (i=r, g s) are pixel (x, current measured value y).α
iBe threshold parameter, can in experiment, determine by experience.After obtaining bianry image, it is being carried out corrosion and the next further filtering noise of expansive working in the morphological operator.Connected component analysis by binaryzation at last is used to extract simply connected foreground area, as the moving region after cutting apart.
(2) detection of people's main shaft
For simplicity, suppose that the people is upright walking, and people's main shaft exists.Suppose that based on this we discuss the detection method of three kinds of people's under the different situations main shaft, promptly single people's main shaft detects; Group's main shaft detects; Block the detection of situation lower main axis.These three kinds of situations can be judged according to corresponding relation in following the tracks of.
Single people's main shaft detects.For single people, adopt minimum intermediate value quadratic method (LeastMedianSequence) to detect its main shaft.Suppose i foreground pixel X
iTo the vertical range for the treatment of boning out l is D (X
i, l), according to minimum intermediate value quadratic method, then all foreground pixels are to square intermediate value minimum of the vertical range of main shaft.
Accompanying drawing 1 has provided single people's main shaft detection example.
Group's main shaft detects.Group's main shaft detection comprises that individuality is cut apart and main shaft detects the two large divisions.
For individuality is cut apart, introduce the upright projection histogram, the individuality among the group is corresponding to the histogrammic peak region of upright projection.Have only the peak region that satisfies certain condition just corresponding to single individuality, we call the obvious peak value zone to these peak regions.Two conditions must be satisfied in the obvious peak value zone:
A) maximum of peak region must be greater than a certain specific threshold value, peak threshold P
T
B) minimum value of peak region must be less than a certain specific threshold value, trough threshold value C
T
Suppose in a peak region P
1, P
2...., P
nIt is its local extremum.C
l, C
rBe respectively this regional left and right sides trough value, then above two conditions can be expressed as with mathematic(al) representation:
max(P
1,P
2,.....,P
n)>P
T (3)
C
l<C
T,C
r<C
T (4)
Trough threshold value C
TBe chosen for whole histogrammic average, crest threshold value P
TBe chosen for people in image coordinate height 80 percent.The method that in the second step main shaft detects, can adopt above-mentioned single people's main shaft to detect.
Fig. 2 has provided the example that a plurality of people's main shafts detect.(b) being detected foreground area, (c) is its upright projection histogram, (d) is the result after cutting apart.In upright projection histogram (c), three tangible peak regions are arranged, can be that three parts correspond respectively to three individualities thus with this Region Segmentation.At last, their main shaft provides in figure (e).
The detection of the main shaft of the people under the situation of blocking.Earlier the people is split, find out the pixel of its foreground area, detect people's main shaft in method the minimum intermediate value square error of foreground pixel utilization that splits.We adopt and realize cutting apart under the situation of blocking based on the color template method.This model comprises a color model and an additional probability mask.When occurring blocking, cutting apart of moving object can be expressed as classification problem, judges promptly which model which foreground pixel belongs to.This problem can solve with bayes rule.If satisfy following equation:
Then foreground pixel X belongs to k moving object (model).
The main shaft that blocks under the situation detects as shown in Figure 3.
(3) tracking under the single camera
Adopt Kalman filter to realize following the tracks of
Choose state vector X comprise the position of people in image (x, y) and movement velocity (v
x, v
y), observation vector Z be the people the position (x, y).The measured value of people's movement position (x y) is estimated as people pin point (" land-point in image "), the i.e. intersection point of the lower limb of people's main shaft and rectangle frame.Then:
Z
t=[x
t,y
t]
T X
t=[x
t,v
x,t,y
t,v
y,t]
T (6)
The speed of people's walking can be thought constant approx, and state-transition matrix and observing matrix are:
By the position of the position prediction present frame object of former frame object, the measured value (intersection point of people's main shaft and rectangle frame lower limb) with present frame compares then.If the distance between measured value and the predicted value is very little, show that then measured value is with a high credibility, measured value can be represented the position of present frame object; If both distances surpass certain threshold value (such as people's the Lower Half this situation that is blocked), then measured value is insincere, and the intersection point that we have defined main shaft and another line upgrades measured value, and this line is exactly the vertical line of future position to main shaft.Tracking framework under the whole single camera can be referring to accompanying drawing 4.
(4) it is right to seek all best main shaft couplings
At first need to calculate and singly reflect matrix between the different images plane.Singly reflected matrix description between the different images about the one-to-one relationship of the point on the same plane.Can find the solution the parameter of matrix by simultaneous equations by given corresponding points.The corresponding points of this method are obtained the employing manual mode: promptly in advance index point is set in scene or utilizes some special corresponding points in the scene.
Define main shaft match likelihood function then.
Main shaft match likelihood function is used for weighing the matching degree of different visual angles servant's main shaft.Before providing concrete definition, the geometrical relationship of different visual angles lower main axis can be referring to accompanying drawing 5.As shown in the figure, suppose to have two video camera i and j.Suppose that cameras view to the main shaft of people s is
,
Be
Projection on ground level.Observing people k for video camera j has accordingly
With
The video camera i plane of delineation to the matrix that singly reflects of the video camera j plane of delineation is
We will by singly reflecting matrix
Project in the plane of delineation coordinate system of video camera j and can obtain
With
Will intersect at a point
According to the character of singly reflecting matrix, if the people k that people s that video camera i observes and video camera j observe is corresponding to the same individual in the three dimensions, then
Corresponding to this people's " land-point ", the i.e. intersection point of people's main shaft and ground level.Therefore, the measured value of " land-point " and intersection point
Between distance can be used for weighing matching degree between the main shaft.Distance is mated between the bright main shaft of novel more more.
According to the geometrical relationship of different visual angles lower main axis, the match likelihood function between definition people s and the k main shaft is
Wherein
Be " land-point " that video camera i observes people s,
Be " land-point " that video camera j observes people k
It is main shaft s is transformed into j visual angle and main shaft k from the i visual angle intersection point.
In order to be without loss of generality, suppose top two probability density functions clothes cloth, then:
At last find all optimum Match right according to matching algorithm
The coupling of multiple-camera in fact can modeling becomes the problem of maximum likelihood function, and the promptly corresponding mutually right main shaft match likelihood function of main shaft is maximum in the right main shaft match likelihood function of numerous main shafts.For the simplification problem, we have defined the main shaft matching distance, and the maximum likelihood function problem is converted into the minimal matching span problem.
Wherein
Be the main shaft matching distance, definition main shaft matching distance as shown in the formula:
The main shaft matching algorithm is right in order to seek global best matches, makes theirs
The sum minimum.Main shaft matching algorithm between two video cameras is described below:
Step 1: detected main shaft under two visual angles is made up in twos, form the main shaft that might mate right.Be without loss of generality, suppose M≤N, then M main shaft selected M main shaft successively according to priority and matched formation altogether from N main shaft
Plant combination.Each combination has following form:
Step 2: for each the combination in each main shaft to { m, n} calculate its matching distance
And guarantee
D wherein
TBe the threshold value that draws by experience, be used for judging that main shaft is to { m, whether n} mates.If do not satisfy above-mentioned constraint, then main shaft to { m, n} is from θ
kIn delete.Step 3: choose the maximum combination Θ of coupling logarithm l
k, all Θ
kForm set Θ:
Step 4: in set Θ, seek global best matches combination λ, feasible wherein all matching distance
Sum reaches minimum value, promptly satisfies following formula:
Step 5: the Θ that finally obtains
λThen be the global best matches combination, wherein each is to then right for the main shaft of coupling.
Above-mentioned algorithm is easy to expand to the situation more than two video cameras.At first video camera makes up in twos, for a pair of video camera that public ground level zone is arranged, sets up matching relationship and can adopt the algorithm of introducing above; When if coupling produces contradiction between the video camera in twos, it is right that then matching relationship only considers to have the main shaft of minimal matching span.
(5) merge various visual angles information updating tracking results
When find all coupling main shafts to after, these match information just can be used for upgrading the tracking results under the single camera.At the situation of two video cameras, have only when the people who follows the tracks of is in public ground level zone, two visual angles, it is just effective to upgrade this step.
Suppose to find corresponding under same individual two visual angles by matching algorithm above-mentioned, the main shaft that is visual angle i and visual angle j respectively is right, main shaft under the j of visual angle is transformed in the plane of delineation of visual angle i by the relation of singly reflecting between two planes of delineation, main shaft under the then original visual angle i is exactly the position of final this person in plane of delineation i with the intersection point of the straight line that conversion is come, and is used for upgrading the tracking results under the original single-view i.For visual angle j in like manner as can be known.
As accompanying drawing 5, if the main shaft under two visual angles
With
Corresponding to same people, then will
Be transformed into visual angle j from visual angle i and obtain straight line
, then this straight line with
Meet at a little
This intersection point just corresponding " land-point " of this people under the j of visual angle, i.e. the intersection point of people's main shaft and ground level.
For the situation more than two video cameras, a people has two or more such intersection points, then chooses the final position " land-point " of the mean value of these intersection points as the people.
Situation about being blocked for people's Lower Half, based on the position of prediction and the main shaft of detection, the coupling that this algorithm still can robust estimates the position (" land-point ") of people in image accurately.
In order to implement concretism of the present invention, we have done a large amount of experiments on two databases, have realized that the pedestrian based on the main shaft coupling follows the tracks of under the multiple-camera.Experimental result has further been verified the validity and the robustness of this method.
Experimental result as shown in Figure 6, tracked human rectangle frame is represented, the sequence number that digital watch below the rectangle frame is leted others have a look at, the center line of rectangle frame is represented detected main shaft, the intersection point of the lower limb line of main shaft and rectangle frame is represented the position of people in image estimated.
Numeral 1,2,3,4 expressions in (a) are many people tracking test results under two video cameras of NLPR database.
(b) numeral 1,2,3,4 expressions in are many people tracking test results under three video cameras of NLPR database.
Numeral 1,2 expressions among Fig. 6 (c) are to the tracking test result of single people under the situation of blocking under two video cameras of PETS2001 database.
Numeral 1,2,3,4 expressions among Fig. 6 (d) are to the tracking test result of group under the situation of blocking under two video cameras of PETS2001 database.
Claims (8)
- One kind under multiple-camera based on the pedestrian tracting method of main shaft coupling, comprise step:Multiple-camera is meant that using two or more single cameras, the precondition of this method is to have a public plane in the viewing area of hypothesis different cameras, and this public plane is meant ground level;(1) to image sequence filtering, obtain single Gauss's background model, carry out moving Object Segmentation by the background method of wiping out, obtain the moving region;(2) extract the feature of the main shaft feature of people in the moving region as coupling, the main shaft detection that three kinds of situations are lowered to is: single people, group and the main shaft detection of blocking the people under the situation;(3) main spindle's that obtains predicting by Kalman filter in conjunction with detected main shaft is realized the tracking under the single camera;(4) calculate and singly to reflect matrix between the different images plane, adopt optimum Match to algorithm, it is right to seek all optimum Match according to the main spindle's of prediction;(5) when find all optimum Match to after, merge match information and upgrade tracking results under the single-view.
- 2. by the described method of claim 1, it is characterized in that, moving Object Segmentation, moving Object Segmentation is the first step of motion tracking, algorithm adopts is that the background method of wiping out of single Gauss model detects the moving region, adopt normalized color rgs model, r=R/ (R+G+B) wherein, g=G/ (R+G+B), s=R+G+B carries out filtering to one section background image sequence earlier, obtains single Gauss's background model, then present image and background image are carried out difference processing, obtain binary image M in that difference image is carried out threshold process Xy, obtain binary image after, come further filtering noise with expansive working it being carried out in the morphological operator corrosion, the connected component analysis by binaryzation at last is used to extract simply connected foreground area, as the moving region after cutting apart.
- 3. by the described method of claim 1, it is characterized in that, tracking employing Kalman filter under the single camera realizes following the tracks of the position by the position prediction present frame object of former frame object, measured value with present frame compares then, if the distance between measured value and the predicted value is very little, show that then measured value is with a high credibility, measured value is represented the position of present frame object; If both distances surpass certain threshold value, then measured value is insincere, and the intersection point of definition main shaft and another line upgrades measured value, and this line is exactly the vertical line of future position to main shaft.
- 4. by the described method of claim 1, it is characterized in that all optimum Match of described searching is to comprising step:Calculate and singly to reflect matrix between the different images plane;Singly reflected matrix description between the different images about the one-to-one relationship of the point on the same plane, find the solution the parameter of matrix by given corresponding points simultaneous equations, corresponding points are obtained the employing manual mode: promptly in advance index point is set in scene or utilizes some special corresponding points in the scene;Calculate the right match likelihood functional value of all main shafts;According to the geometrical relationship of different visual angles lower main axis, the match likelihood function between definition main shaft s and the k is:Wherein Be " the pin point " that video camera i observes people s, Be " the pin point " that video camera j observes people k It is main shaft s is transformed into j visual angle and main shaft k from the i visual angle intersection point;Find all optimum Match right according to matching algorithm;The coupling modeling of multiple-camera becomes the problem of maximum likelihood function, the promptly corresponding mutually right main shaft match likelihood function of main shaft is maximum in the right main shaft match likelihood function of numerous main shafts, for the simplification problem changes into the minimal matching span problem to the maximum likelihood function problem;It is as follows to algorithm to seek all optimum Match:Step 1: detected main shaft under two visual angles is made up in twos, and it is right to form the main shaft that institute might mate, and is without loss of generality, and supposes M≤N, and then M main shaft selected M main shaft successively according to priority and matched common formation from N main shaft Plant combination, each combination has following form:Step 2: for each the combination in each main shaft to { m, n} calculate its matching distance And guaranteeStep 3: choose the maximum combination Θ of coupling logarithm l k, all Θ kForm set Θ:Step 4: in set Θ, seek global best matches combination λ, feasible wherein all matching distance Sum reaches minimum value, promptly satisfies following formula:Step 5: the Θ that finally obtains λThen be the global best matches combination, wherein each is to then right for the main shaft of coupling.
- 5. by the described method of claim 1, it is characterized in that the tracking results that described fusion match information is upgraded under the single-view comprises step:When find all coupling main shafts to after, these match information just are used for upgrading the tracking results under the single camera, at the situation of two video cameras, have only when the people who follows the tracks of is in public ground levels zone, two visual angles, upgrade this step just effectively,Suppose to find corresponding under same individual two visual angles by matching algorithm above-mentioned, the main shaft that is visual angle i and visual angle j respectively is right, main shaft under the j of visual angle is transformed in the plane of delineation of visual angle i by the relation of singly reflecting between two planes of delineation, main shaft under the then original visual angle i is exactly the position of final this person in plane of delineation i with the intersection point of the straight line that conversion is come, be used for upgrading the tracking results under the original single-view i, for visual angle j in like manner as can be knownFor the situation more than two video cameras, a people has two or more such intersection points, then chooses the final position of the mean value of these intersection points as the people.
- 6. by the described method of claim 3, it is characterized in that, the detection of single people's main shaft,Single people's main shaft detects, and for single people, adopts minimum intermediate value quadratic method to detect its main shaft, supposes i foreground pixel X iTo the vertical range for the treatment of boning out l is D (X i, l), according to minimum intermediate value quadratic method, then all foreground pixels are to square intermediate value minimum of the vertical range of main shaft,
- 7. by the described method of claim 3, it is characterized in that, group's main shaft detection comprises that individuality is cut apart and main shaft detects the two large divisions, for being cut apart, individuality introduces the upright projection histogram, individuality among the group is corresponding to the histogrammic peak region of upright projection, the peak region that satisfies condition is called the obvious peak value zone corresponding to single individuality, and described obvious peak value zone must be satisfied two conditions of described peak region and is:A) maximum of peak region must be greater than a certain specific threshold value, peak threshold P TB) minimum value of peak region must be less than a certain specific threshold value, trough threshold value C TSuppose in a peak region P 1, P 2...., P nBe its local extremum, C l, C rBe respectively this regional left and right sides trough value, then above two conditions are expressed as with mathematic(al) representation:max(P 1,P 2,......,P n)>P TC l<C T,C r<C TTrough threshold value C TBe chosen for whole histogrammic average, crest threshold value P TBe chosen for people in image coordinate height 80 percent, in main shaft detects, adopt single people's main shaft to detect.
- 8. by the described method of claim 3, it is characterized in that, the detection of the main shaft of the people under the situation of blocking, earlier the people is split, find out the pixel of its foreground area, detect people's main shaft in method to the minimum intermediate value square error of the foreground pixel utilization that splits, employing realizes cutting apart under the situation of blocking based on the color template method, this model comprises a color model and an additional probability mask, when occurring blocking, the contingency table of moving object is shown classification problem, judges promptly which model which foreground pixel belongs to, this classification problem solves with bayes rule, if satisfy following equation:Then foreground pixel X belongs to k moving object.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CNB200510108137XA CN100525395C (en) | 2005-09-29 | 2005-09-29 | Pedestrian tracting method based on principal axis marriage under multiple vedio cameras |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CNB200510108137XA CN100525395C (en) | 2005-09-29 | 2005-09-29 | Pedestrian tracting method based on principal axis marriage under multiple vedio cameras |
Publications (2)
Publication Number | Publication Date |
---|---|
CN1941850A CN1941850A (en) | 2007-04-04 |
CN100525395C true CN100525395C (en) | 2009-08-05 |
Family
ID=37959590
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CNB200510108137XA Expired - Fee Related CN100525395C (en) | 2005-09-29 | 2005-09-29 | Pedestrian tracting method based on principal axis marriage under multiple vedio cameras |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN100525395C (en) |
Families Citing this family (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5205337B2 (en) * | 2009-06-18 | 2013-06-05 | 富士フイルム株式会社 | Target tracking device, image tracking device, operation control method thereof, and digital camera |
CN101930603B (en) * | 2010-08-06 | 2012-08-22 | 华南理工大学 | Method for fusing image data of medium-high speed sensor network |
CN102385750B (en) * | 2011-06-22 | 2013-07-10 | 清华大学 | Line matching method and line matching system on basis of geometrical relationship |
CN103099623B (en) * | 2013-01-25 | 2014-11-05 | 中国科学院自动化研究所 | Extraction method of kinesiology parameters |
CN103164855B (en) * | 2013-02-26 | 2016-04-27 | 清华大学深圳研究生院 | A kind of Bayesian decision foreground extracting method in conjunction with reflected light photograph |
CN103729620B (en) * | 2013-12-12 | 2017-11-03 | 北京大学 | A kind of multi-view pedestrian detection method based on multi-view Bayesian network |
CN103700106A (en) * | 2013-12-26 | 2014-04-02 | 南京理工大学 | Distributed-camera-based multi-view moving object counting and positioning method |
CN104978735B (en) * | 2014-04-14 | 2018-02-13 | 航天信息股份有限公司 | It is suitable for the background modeling method of random noise and illumination variation |
JP6800628B2 (en) * | 2016-06-22 | 2020-12-16 | キヤノン株式会社 | Tracking device, tracking method, and program |
CN109521450B (en) * | 2017-09-20 | 2020-12-29 | 阿里巴巴(中国)有限公司 | Positioning drift detection method and device |
CN110276233A (en) * | 2018-03-15 | 2019-09-24 | 南京大学 | A kind of polyphaser collaboration tracking system based on deep learning |
CN109002825A (en) * | 2018-08-07 | 2018-12-14 | 成都睿码科技有限责任公司 | Hand bandage for dressing detection method based on video analysis |
CN109165600B (en) * | 2018-08-27 | 2021-11-26 | 浙江大丰实业股份有限公司 | Intelligent search platform for stage performance personnel |
WO2020061792A1 (en) * | 2018-09-26 | 2020-04-02 | Intel Corporation | Real-time multi-view detection of objects in multi-camera environments |
CN110503097A (en) * | 2019-08-27 | 2019-11-26 | 腾讯科技(深圳)有限公司 | Training method, device and the storage medium of image processing model |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1508756A (en) * | 2002-12-17 | 2004-06-30 | 中国科学院自动化研究所 | Sensitive image identifying method based on body local and shape information |
CN1508755A (en) * | 2002-12-17 | 2004-06-30 | 中国科学院自动化研究所 | Sensitive video-frequency detecting method |
WO2004081895A1 (en) * | 2003-03-10 | 2004-09-23 | Mobotix Ag | Monitoring device |
-
2005
- 2005-09-29 CN CNB200510108137XA patent/CN100525395C/en not_active Expired - Fee Related
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1508756A (en) * | 2002-12-17 | 2004-06-30 | 中国科学院自动化研究所 | Sensitive image identifying method based on body local and shape information |
CN1508755A (en) * | 2002-12-17 | 2004-06-30 | 中国科学院自动化研究所 | Sensitive video-frequency detecting method |
WO2004081895A1 (en) * | 2003-03-10 | 2004-09-23 | Mobotix Ag | Monitoring device |
Non-Patent Citations (1)
Title |
---|
一种能处理部分遮挡的鲁棒性实时跟踪方法. 田原,谭铁牛,孙洪赞,胡卫明.工程图学学报,第2001年增刊期. 2001 * |
Also Published As
Publication number | Publication date |
---|---|
CN1941850A (en) | 2007-04-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN100525395C (en) | Pedestrian tracting method based on principal axis marriage under multiple vedio cameras | |
CN106203274B (en) | Real-time pedestrian detection system and method in video monitoring | |
CN108447078B (en) | Interference perception tracking algorithm based on visual saliency | |
CN102663743B (en) | Personage's method for tracing that in a kind of complex scene, many Kameras are collaborative | |
CN106204640A (en) | A kind of moving object detection system and method | |
CN103164858B (en) | Adhesion crowd based on super-pixel and graph model is split and tracking | |
CN115311241B (en) | Underground coal mine pedestrian detection method based on image fusion and feature enhancement | |
CN101701818B (en) | Method for detecting long-distance barrier | |
CN101315701B (en) | Movement destination image partition method | |
CN106600625A (en) | Image processing method and device for detecting small-sized living thing | |
CN104978567B (en) | Vehicle checking method based on scene classification | |
CN102629326A (en) | Lane line detection method based on monocular vision | |
CN103077539A (en) | Moving object tracking method under complicated background and sheltering condition | |
CN105160649A (en) | Multi-target tracking method and system based on kernel function unsupervised clustering | |
CN104077605A (en) | Pedestrian search and recognition method based on color topological structure | |
Börcs et al. | Extraction of vehicle groups in airborne LiDAR point clouds with two-level point processes | |
CN101777188A (en) | Real-time bus passenger flow volume statistical method | |
CN116129302A (en) | Method for counting people flow in deep building site based on Yolov5 | |
Hautière et al. | Road scene analysis by stereovision: a robust and quasi-dense approach | |
CN104200483A (en) | Human body central line based target detection method under multi-camera environment | |
Wang et al. | Vision‐Based Lane Departure Detection Using a Stacked Sparse Autoencoder | |
Wu et al. | Enhanced roadway geometry data collection using an effective video log image-processing algorithm | |
Gu et al. | A full-time deep learning-based alert approach for bridge–ship collision using visible spectrum and thermal infrared cameras | |
Salmane et al. | A novel evidence based model for detecting dangerous situations in level crossing environments | |
Benedek et al. | Building extraction and change detection in multitemporal aerial and satellite images in a joint stochastic approach |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20090805 Termination date: 20180929 |
|
CF01 | Termination of patent right due to non-payment of annual fee |