CN110782483A - Multi-view multi-target tracking method and system based on distributed camera network - Google Patents
Multi-view multi-target tracking method and system based on distributed camera network Download PDFInfo
- Publication number
- CN110782483A CN110782483A CN201911012807.6A CN201911012807A CN110782483A CN 110782483 A CN110782483 A CN 110782483A CN 201911012807 A CN201911012807 A CN 201911012807A CN 110782483 A CN110782483 A CN 110782483A
- Authority
- CN
- China
- Prior art keywords
- detected target
- information
- current frame
- camera
- frame view
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/292—Multi-camera tracking
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/52—Surveillance or monitoring of activities, e.g. for recognising suspicious objects
- G06V20/53—Recognition of crowd images, e.g. recognition of crowd congestion
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/181—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/188—Capturing isolated or intermittent images triggered by the occurrence of a predetermined event, e.g. an object reaching a predetermined position
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30241—Trajectory
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/255—Detecting or recognising potential candidate objects based on visual cues, e.g. shapes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/94—Hardware or software architectures specially adapted for image or video understanding
- G06V10/95—Hardware or software architectures specially adapted for image or video understanding structured as a network, e.g. client-server architectures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
- G06V40/171—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Computation (AREA)
- Signal Processing (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- Databases & Information Systems (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Image Analysis (AREA)
Abstract
The utility model discloses a multi-view multi-target tracking method and a system based on a distributed camera network, comprising: acquiring a current frame view acquired by each camera in the distributed camera network; extracting a rectangular bounding box of a detected target from the current frame view; extracting visual appearance information of the detected target from the image inside the rectangular bounding box by using a pre-trained convolutional neural network; converting the image coordinates of the detected target in the current frame view into ground coordinates; an output step: constructing a data association matrix based on the visual appearance information and the ground coordinates of the detected target; and processing the data association matrix by adopting a Hungarian algorithm, and outputting the result of successful matching or failed matching of the detected target and the known track in the current frame view.
Description
Technical Field
The disclosure relates to the technical field of multi-target tracking, in particular to a multi-view multi-target tracking method and system based on a distributed camera network.
Background
The statements in this section merely provide background information related to the present disclosure and may not constitute prior art.
multi-Object Tracking (Multiple Object Tracking) technology has many applications in today's society, such as monitoring, surveillance, crowd behavior analysis, etc.
In the course of implementing the present disclosure, the inventors found that the following technical problems exist in the prior art:
multi-target tracking remains a challenging task because it requires simultaneous solution to the problems of target detection, trajectory estimation, data association, and re-identification. In order to detect the target, various sensors such as radar, laser, sonar, camera and the like can be used according to the requirements of specific tasks, and corresponding detection algorithms are also required to be equipped, and the detection of the target is one of the difficulties of multi-target tracking. Another challenging problem of multi-target tracking is occlusion. The target can be shielded by other objects and can also be shielded by the current View (Field Of View), and frequent shielding easily causes the target to be lost and influences the tracking precision.
Disclosure of Invention
In order to solve the defects of the prior art, the disclosure provides a multi-view multi-target tracking method and system based on a distributed camera network;
in a first aspect, the present disclosure provides a distributed camera network based multi-view multi-target tracking method;
the multi-view multi-target tracking method based on the distributed camera network comprises the following steps:
acquiring a current frame view acquired by each camera in the distributed camera network;
extracting a rectangular bounding box of a detected target from the current frame view;
extracting visual appearance information of the detected target from the image inside the rectangular bounding box by using a pre-trained convolutional neural network; converting the image coordinates of the detected target in the current frame view into ground coordinates;
an output step: constructing a data association matrix based on the visual appearance information and the ground coordinates of the detected target; and processing the data association matrix by adopting a Hungarian algorithm, and outputting the result of successful matching or failed matching of the detected target and the known track in the current frame view.
In a second aspect, the present disclosure further provides a distributed camera network-based multi-view multi-target tracking system;
a multi-view multi-target tracking system based on a distributed camera network comprises:
an acquisition module configured to: acquiring a current frame view acquired by each camera in the distributed camera network;
a pre-processing module configured to: extracting a rectangular bounding box of a detected target from the current frame view;
an extraction module configured to: extracting visual appearance information of the detected target from the image inside the rectangular bounding box by using a pre-trained convolutional neural network; converting the image coordinates of the detected target in the current frame view into ground coordinates;
an output module configured to: constructing a data association matrix based on the visual appearance information and the ground coordinates of the detected target; and processing the data association matrix by adopting a Hungarian algorithm, and outputting the result of successful matching or failed matching of the detected target and the known track in the current frame view.
In a third aspect, the present disclosure also provides an electronic device comprising a memory and a processor, and computer instructions stored on the memory and executed on the processor, wherein the computer instructions, when executed by the processor, perform the steps of the method of the first aspect.
In a fourth aspect, the present disclosure also provides a computer-readable storage medium for storing computer instructions which, when executed by a processor, perform the steps of the method of the first aspect.
Compared with the prior art, the beneficial effect of this disclosure is:
the method adopts a data incidence matrix generated by combining visual appearance information and ground coordinates, and adopts the data incidence matrix to realize the matching of the detected target and the known track, so that the matching accuracy can be improved;
compared with the original DeepsORT method, the method has stronger robustness in processing the target re-identification and occlusion problems.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this application, illustrate embodiments of the application and, together with the description, serve to explain the application and are not intended to limit the application.
Fig. 1 is an overall structure of a distributed multi-view multi-target tracking system of the first embodiment.
Detailed Description
It should be noted that the following detailed description is exemplary and is intended to provide further explanation of the disclosure. Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this application belongs.
It is noted that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of example embodiments according to the present application. As used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, and it should be understood that when the terms "comprises" and/or "comprising" are used in this specification, they specify the presence of stated features, steps, operations, devices, components, and/or combinations thereof, unless the context clearly indicates otherwise.
The embodiment I provides a multi-view multi-target tracking method based on a distributed camera network;
the multi-view multi-target tracking method based on the distributed camera network comprises the following steps:
s1: acquiring a current frame view acquired by each camera in the distributed camera network;
s2: extracting a rectangular bounding box of a detected target from the current frame view;
s3: extracting visual appearance information of the detected target from the image inside the rectangular bounding box by using a pre-trained convolutional neural network; converting the image coordinates of the detected target in the current frame view into ground coordinates;
s4: an output step: constructing a data association matrix based on the visual appearance information and the ground coordinates of the detected target; and processing the data association matrix by adopting a Hungarian algorithm, and outputting the result of successful matching or failed matching of the detected target and the known track in the current frame view.
As one or more embodiments, in S4, the specific steps of the outputting step include:
calculating the Mahalanobis distance between the ground coordinates of the detected target and the terminal coordinates of each storage track in the current frame view;
calculating M cosine distances between the visual appearance information of the detected target and the visual appearance information of the adjacent previous M frames of the current frame view; storing the minimum value of the M cosine distances as the final cosine distance;
when both the mahalanobis distance and the final cosine distance are smaller than a set threshold, performing weighted summation on the mahalanobis distance and the final cosine distance to obtain a data association matrix;
and inputting the data association matrix into a Hungarian algorithm, and outputting the result of successful matching or failed matching of the detected target and the known track in the current frame view by the Hungarian algorithm.
As one or more embodiments, the method further comprises:
s5: if the image coordinate information of the detected target successfully matched with the current camera and the corresponding track number ID exist in the current camera; repeatedly and iteratively exchanging the successfully matched information in the current camera and the successfully associated information in the adjacent camera, and calculating the average consistency to obtain a converged information vector and a converged information matrix;
calculating posterior pose information based on the converged information vector and the converged information matrix; therefore, multi-view and multi-target tracking is realized; then, position information of the detected object of the next frame view is predicted.
As one or more embodiments, the method further comprises:
s6: if the ground coordinates of the detected target which fails to be matched with the corresponding track number ID exist in the current camera; calculating Euclidean distances between the coordinate information of the detected target which fails to be matched in the current camera and the coordinate information of each storage track terminal point in the views shot by the rest other cameras;
and if the Euclidean distance is smaller than the set threshold value, matching the ground coordinates of the detected target in the current camera with corresponding tracks in the views shot by the other remaining cameras.
As one or more embodiments, a distributed camera network, refers to: distributed as opposed to centralized, which is where data is centralized for processing at a central processor; distributed is where data is processed separately on individual processors, which may communicate with each other.
As one or more embodiments, in S2, the rectangular bounding box of the detected object is extracted from the current frame view by using the YOLOv3 network.
As one or more embodiments, in S3, the training step of the pre-trained convolutional neural network includes:
constructing a convolutional neural network; constructing a training set; the training set is an image with known visual appearance information;
inputting the training set into a convolutional neural network, and training the convolutional neural network;
and obtaining the trained convolutional neural network.
For example: the training set is as follows: a large-scale pedestrian re-identification dataset containing over 1100000 images of 1261 pedestrians.
As one or more embodiments, in S3, the visual appearance information specifically refers to 128-dimensional normalized features of the convolutional neural network output. For example, it may be a profile feature.
As one or more embodiments, in S3, the image coordinate information of the detected object in the current frame view is converted into ground coordinates; the method comprises the following specific steps:
the pixel coordinates of the middle point of the bottom edge of the bounding box of the person in the image are used as the position information of the person, and the position information is converted into ground coordinates through a homography matrix (homography matrix), and the homography matrix is calibrated through a camera.
As one or more embodiments, in S5, the specific step of calculating the average consistency includes:
based on an ICF algorithm, information exchange is carried out between adjacent cameras through repeated iteration to obtain a converged information vector and a converged information matrix:
wherein e represents a constant,
the information vector is represented by a vector of information,
representing an information matrix.
As one or more embodiments, in S5, the specific step of calculating posterior pose information includes:
Wherein N is the number of cameras.
As one or more embodiments, in S5, position information of a detected object of a next frame view is predicted; the method comprises the following specific steps:
Wherein i refers to the ith node, t is the tth frame, Φ is the linear state transition matrix, and Q is the process noise covariance.
The general structure of the distributed multi-target tracking method provided by the disclosure is shown in fig. 1. Firstly, the YOLOv3 is used to detect the target of each camera, and the algorithm can extract the rectangular bounding box of the detected target with higher frame rate. Then, the visual appearance information of the target is obtained by a pre-trained convolutional neural network. The Hungarian algorithm is used for combining the visual appearance information and the position information of the target to carry out data association. The position information of the object can be fused from the plurality of view information by using an information weighted Consensus Filter (information weighted Consensus Filter). The method comprises the following specific steps:
1. target detection
Object detection refers to acquiring different objects in an image and determining their type and location. The target detection method based on deep learning has strong robustness to illumination change, shielding problems and complex environments, and has two main research directions: a two-stage process and a one-stage process. The two-stage method first predicts the number of candidate frames that may have targets, then adjusts the size of the frames and classifies the frames to obtain the precise location, size and category of the targets, such as Faster R-CNN. The first step of the one-stage method is omitted, and the position and the category of the target are directly predicted, such as YOLOv3, compared with the two stages, the one-stage method is generally faster and has equivalent performance. Therefore, we chose YOLOv3 as the target detector.
2. Data association
A simple Hungarian algorithm is used for data correlation, visual appearance information of a target is a 128-dimensional feature vector obtained by a trained convolutional neural network, position information of the target is obtained by converting image coordinates to the ground by using a corrected homography matrix (homography matrix), and a final correlation matrix is obtained by weighting the visual appearance information and the target position information:
c
i,j=λd
(1)(i,j)+(1-λ)d
(2)(i,j) (1)
where i, j represent the ith and jth measurements, respectively, λ is an adjustable weighting parameter, d
(1)For measuring the Mahalanobis distance between the position and the last frame of each stored track of the current view, d
(2)The minimum cosine distance between the appearance information and the appearance information stored in each track is measured.
It should be understood that the homography matrix refers to a homography matrix obtained through camera calibration, and the homography matrix can convert pixel coordinates into ground coordinates.
Furthermore, we use a threshold function to ignore irrelevant candidates:
where k is equal to 1 or 2, representing appearance information or position information, the association between the ith trajectory and the jth tracked object is only allowed if both the mahalanobis and cosine distances are less than the corresponding thresholds, i.e., b is set to 1.
3. Track processing using multi-view information
The track processing step is used for ID management, including restoring an old track or creating a new track. The old track is recovered, namely, after a person walks out of the visual field for 30 frames, the track of the person is deleted, and when the person returns again, the original ID is given to the person; creating a new trajectory is a new person entering the field of view assigned a new ID. When a detection value failing to be matched is found in the current view, the position information of the detection value is matched with the position information of the last frame of the track by using the Euclidean distance, the old track is recovered, and if a matched candidate is found, the ID of the track is given to the detection value.
If the match fails again, the algorithm checks whether a match occurs in the other views, i.e. a detected match with the same match failure in the other views, for the generation of a new track, and if the distance meets the threshold requirement, a new ID is initialized for them. In addition, the algorithm will remove the trajectory that disappears more than 30S in the current view to reduce the interference.
4. Information weighted consensus filter for multi-view information fusion
An information weighted coherent filter (ICF) is an efficient distributed state estimation method. Here, we use ICF for multi-view information fusion to estimate the location of the target. ICF mainly comprises three steps: state prediction, measurement update and weighting are consistent. In the aspect of state prediction, in S6, we use a linear uniform velocity model to predict the state variable x and the information matrix W of the target next step.
Wherein i refers to the ith node, t is the tth frame, Φ is the linear state transition matrix, and Q is the process noise covariance. The predicted location information is sent to a data correlation module to be matched with the measured values. At measurement update, we use the current measurement value z
iTo calculate an information vector v
iAnd an information matrix V
i。
In the formula
H
i,R
iAnd N is a prior state vector, an information matrix, an observation matrix, a measurement noise covariance and the number of cameras respectively. For weighted agreement, each camera sends and accepts information vectors v to its neighboring cameras
iAnd an information matrix V
iK iterations are passed until the filter converges.
Where e represents a constant.
Wherein N is the number of cameras.
The second embodiment also provides a multi-view multi-target tracking system based on the distributed camera network;
a multi-view multi-target tracking system based on a distributed camera network comprises:
an acquisition module configured to: acquiring a current frame view acquired by each camera in the distributed camera network;
a pre-processing module configured to: extracting a rectangular bounding box of a detected target from the current frame view;
an extraction module configured to: extracting visual appearance information of the detected target from the image inside the rectangular bounding box by using a pre-trained convolutional neural network; converting the image coordinates of the detected target in the current frame view into ground coordinates;
an output module configured to: constructing a data association matrix based on the visual appearance information and the ground coordinates of the detected target; and processing the data association matrix by adopting a Hungarian algorithm, and outputting the result of successful matching or failed matching of the detected target and the known track in the current frame view.
In a third embodiment, the present embodiment further provides an electronic device, which includes a memory, a processor, and computer instructions stored in the memory and executed on the processor, where the computer instructions, when executed by the processor, implement the steps of the method according to the first aspect.
In a fourth embodiment, the present embodiment further provides a computer-readable storage medium for storing computer instructions, and the computer instructions, when executed by a processor, perform the steps of the method according to the first aspect.
The above description is only a preferred embodiment of the present application and is not intended to limit the present application, and various modifications and changes may be made by those skilled in the art. Any modification, equivalent replacement, improvement and the like made within the spirit and principle of the present application shall be included in the protection scope of the present application.
Claims (10)
1. The multi-view multi-target tracking method based on the distributed camera network is characterized by comprising the following steps:
acquiring a current frame view acquired by each camera in the distributed camera network;
extracting a rectangular bounding box of a detected target from the current frame view;
extracting visual appearance information of the detected target from the image inside the rectangular bounding box by using a pre-trained convolutional neural network; converting the image coordinates of the detected target in the current frame view into ground coordinates;
an output step: constructing a data association matrix based on the visual appearance information and the ground coordinates of the detected target; and processing the data association matrix by adopting a Hungarian algorithm, and outputting the result of successful matching or failed matching of the detected target and the known track in the current frame view.
2. The method of claim 1, wherein said outputting step comprises the specific steps of:
calculating the Mahalanobis distance between the ground coordinates of the detected target and the terminal coordinates of each storage track in the current frame view;
calculating M cosine distances between the visual appearance information of the detected target and the visual appearance information of the adjacent previous M frames of the current frame view; storing the minimum value of the M cosine distances as the final cosine distance;
when both the mahalanobis distance and the final cosine distance are smaller than a set threshold, performing weighted summation on the mahalanobis distance and the final cosine distance to obtain a data association matrix;
and inputting the data association matrix into a Hungarian algorithm, and outputting the result of successful matching or failed matching of the detected target and the known track in the current frame view by the Hungarian algorithm.
3. The method of claim 1, further comprising:
if the image coordinate information of the detected target successfully matched with the current camera and the corresponding track number ID exist in the current camera; repeatedly and iteratively exchanging the successfully matched information in the current camera and the successfully associated information in the adjacent camera, and calculating the average consistency to obtain a converged information vector and a converged information matrix;
calculating posterior pose information based on the converged information vector and the converged information matrix; therefore, multi-view and multi-target tracking is realized; then, position information of the detected object of the next frame view is predicted.
4. The method of claim 3, further comprising:
if the ground coordinates of the detected target which fails to be matched with the corresponding track number ID exist in the current camera; calculating Euclidean distances between the coordinate information of the detected target which fails to be matched in the current camera and the coordinate information of each storage track terminal point in the views shot by the rest other cameras;
and if the Euclidean distance is smaller than the set threshold value, matching the ground coordinates of the detected target in the current camera with corresponding tracks in the views shot by the other remaining cameras.
5. The method of claim 3, wherein the extracting the rectangular bounding box of the detected object from the current frame view is performed by using a YOLOv3 network to extract the rectangular bounding box of the detected object from the current frame view.
6. The method of claim 3, wherein the pre-trained convolutional neural network comprises the specific training steps of:
constructing a convolutional neural network; constructing a training set; the training set is an image with known visual appearance information;
inputting the training set into a convolutional neural network, and training the convolutional neural network;
and obtaining the trained convolutional neural network.
7. The method of claim 1, wherein image coordinates of the detected object in the current frame view are converted into ground coordinates; the method comprises the following specific steps:
the pixel coordinates of the middle point of the bottom edge of the bounding box of the person in the image are used as the position information of the person, and the position information is converted into ground coordinates through a homography matrix, wherein the homography matrix is calibrated through a camera.
8. A multi-view multi-target tracking system based on a distributed camera network is characterized by comprising:
an acquisition module configured to: acquiring a current frame view acquired by each camera in the distributed camera network;
a pre-processing module configured to: extracting a rectangular bounding box of a detected target from the current frame view;
an extraction module configured to: extracting visual appearance information of the detected target from the image inside the rectangular bounding box by using a pre-trained convolutional neural network; converting the image coordinates of the detected target in the current frame view into ground coordinates;
an output module configured to: constructing a data association matrix based on the visual appearance information and the ground coordinates of the detected target; and processing the data association matrix by adopting a Hungarian algorithm, and outputting the result of successful matching or failed matching of the detected target and the known track in the current frame view.
9. An electronic device comprising a memory and a processor and computer instructions stored on the memory and executable on the processor, the computer instructions when executed by the processor performing the steps of the method of any of claims 1 to 7.
10. A computer-readable storage medium storing computer instructions which, when executed by a processor, perform the steps of the method of any one of claims 1 to 7.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911012807.6A CN110782483B (en) | 2019-10-23 | 2019-10-23 | Multi-view multi-target tracking method and system based on distributed camera network |
LU102028A LU102028B1 (en) | 2019-10-23 | 2020-09-03 | Multiple view multiple target tracking method and system based on distributed camera network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911012807.6A CN110782483B (en) | 2019-10-23 | 2019-10-23 | Multi-view multi-target tracking method and system based on distributed camera network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110782483A true CN110782483A (en) | 2020-02-11 |
CN110782483B CN110782483B (en) | 2022-03-15 |
Family
ID=69386547
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201911012807.6A Active CN110782483B (en) | 2019-10-23 | 2019-10-23 | Multi-view multi-target tracking method and system based on distributed camera network |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN110782483B (en) |
LU (1) | LU102028B1 (en) |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111626194A (en) * | 2020-05-26 | 2020-09-04 | 佛山市南海区广工大数控装备协同创新研究院 | Pedestrian multi-target tracking method using depth correlation measurement |
CN111738075A (en) * | 2020-05-18 | 2020-10-02 | 深圳奥比中光科技有限公司 | Joint point tracking method and system based on pedestrian detection |
CN112070807A (en) * | 2020-11-11 | 2020-12-11 | 湖北亿咖通科技有限公司 | Multi-target tracking method and electronic device |
CN112215873A (en) * | 2020-08-27 | 2021-01-12 | 国网浙江省电力有限公司电力科学研究院 | Method for tracking and positioning multiple targets in transformer substation |
CN112633205A (en) * | 2020-12-28 | 2021-04-09 | 北京眼神智能科技有限公司 | Pedestrian tracking method and device based on head and shoulder detection, electronic equipment and storage medium |
CN113674317A (en) * | 2021-08-10 | 2021-11-19 | 深圳市捷顺科技实业股份有限公司 | Vehicle tracking method and device of high-order video |
CN114089675A (en) * | 2021-11-23 | 2022-02-25 | 长春工业大学 | Machine control method and control system based on man-machine distance |
CN114299128A (en) * | 2021-12-30 | 2022-04-08 | 咪咕视讯科技有限公司 | Multi-view positioning detection method and device |
CN114596337A (en) * | 2022-03-03 | 2022-06-07 | 捻果科技(深圳)有限公司 | Self-recognition target tracking method and system based on linkage of multiple camera positions |
CN116758119A (en) * | 2023-06-27 | 2023-09-15 | 重庆比特数图科技有限公司 | Multi-target circulation detection tracking method and system based on motion compensation and linkage |
CN117726656A (en) * | 2024-02-08 | 2024-03-19 | 开拓导航控制技术股份有限公司 | Target tracking method, device, system and medium based on super-resolution image |
CN117853759A (en) * | 2024-03-08 | 2024-04-09 | 山东海润数聚科技有限公司 | Multi-target tracking method, system, equipment and storage medium |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170286774A1 (en) * | 2016-04-04 | 2017-10-05 | Xerox Corporation | Deep data association for online multi-class multi-object tracking |
CN107292911A (en) * | 2017-05-23 | 2017-10-24 | 南京邮电大学 | A kind of multi-object tracking method merged based on multi-model with data correlation |
CN109447121A (en) * | 2018-09-27 | 2019-03-08 | 清华大学 | A kind of Visual Sensor Networks multi-object tracking method, apparatus and system |
CN109816690A (en) * | 2018-12-25 | 2019-05-28 | 北京飞搜科技有限公司 | Multi-target tracking method and system based on depth characteristic |
CN109919981A (en) * | 2019-03-11 | 2019-06-21 | 南京邮电大学 | A kind of multi-object tracking method of the multiple features fusion based on Kalman filtering auxiliary |
CN109934844A (en) * | 2019-01-28 | 2019-06-25 | 中国人民解放军战略支援部队信息工程大学 | A kind of multi-object tracking method and system merging geospatial information |
-
2019
- 2019-10-23 CN CN201911012807.6A patent/CN110782483B/en active Active
-
2020
- 2020-09-03 LU LU102028A patent/LU102028B1/en active IP Right Grant
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170286774A1 (en) * | 2016-04-04 | 2017-10-05 | Xerox Corporation | Deep data association for online multi-class multi-object tracking |
CN107292911A (en) * | 2017-05-23 | 2017-10-24 | 南京邮电大学 | A kind of multi-object tracking method merged based on multi-model with data correlation |
CN109447121A (en) * | 2018-09-27 | 2019-03-08 | 清华大学 | A kind of Visual Sensor Networks multi-object tracking method, apparatus and system |
CN109816690A (en) * | 2018-12-25 | 2019-05-28 | 北京飞搜科技有限公司 | Multi-target tracking method and system based on depth characteristic |
CN109934844A (en) * | 2019-01-28 | 2019-06-25 | 中国人民解放军战略支援部队信息工程大学 | A kind of multi-object tracking method and system merging geospatial information |
CN109919981A (en) * | 2019-03-11 | 2019-06-21 | 南京邮电大学 | A kind of multi-object tracking method of the multiple features fusion based on Kalman filtering auxiliary |
Non-Patent Citations (2)
Title |
---|
GUOLIANG LIU 等: "Human Action Recognition Using a Distributed RGB-Depth Camera Network", 《IEEE》 * |
赵广辉 等: "基于卡尔曼滤波的多目标跟踪方法", 《计算机科学》 * |
Cited By (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111738075A (en) * | 2020-05-18 | 2020-10-02 | 深圳奥比中光科技有限公司 | Joint point tracking method and system based on pedestrian detection |
CN111626194A (en) * | 2020-05-26 | 2020-09-04 | 佛山市南海区广工大数控装备协同创新研究院 | Pedestrian multi-target tracking method using depth correlation measurement |
CN111626194B (en) * | 2020-05-26 | 2024-02-02 | 佛山市南海区广工大数控装备协同创新研究院 | Pedestrian multi-target tracking method using depth correlation measurement |
CN112215873A (en) * | 2020-08-27 | 2021-01-12 | 国网浙江省电力有限公司电力科学研究院 | Method for tracking and positioning multiple targets in transformer substation |
CN112070807B (en) * | 2020-11-11 | 2021-02-05 | 湖北亿咖通科技有限公司 | Multi-target tracking method and electronic device |
CN112070807A (en) * | 2020-11-11 | 2020-12-11 | 湖北亿咖通科技有限公司 | Multi-target tracking method and electronic device |
CN112633205A (en) * | 2020-12-28 | 2021-04-09 | 北京眼神智能科技有限公司 | Pedestrian tracking method and device based on head and shoulder detection, electronic equipment and storage medium |
CN113674317B (en) * | 2021-08-10 | 2024-04-26 | 深圳市捷顺科技实业股份有限公司 | Vehicle tracking method and device for high-level video |
CN113674317A (en) * | 2021-08-10 | 2021-11-19 | 深圳市捷顺科技实业股份有限公司 | Vehicle tracking method and device of high-order video |
CN114089675A (en) * | 2021-11-23 | 2022-02-25 | 长春工业大学 | Machine control method and control system based on man-machine distance |
CN114089675B (en) * | 2021-11-23 | 2023-06-09 | 长春工业大学 | Machine control method and system based on man-machine distance |
CN114299128A (en) * | 2021-12-30 | 2022-04-08 | 咪咕视讯科技有限公司 | Multi-view positioning detection method and device |
CN114596337A (en) * | 2022-03-03 | 2022-06-07 | 捻果科技(深圳)有限公司 | Self-recognition target tracking method and system based on linkage of multiple camera positions |
CN116758119A (en) * | 2023-06-27 | 2023-09-15 | 重庆比特数图科技有限公司 | Multi-target circulation detection tracking method and system based on motion compensation and linkage |
CN116758119B (en) * | 2023-06-27 | 2024-04-19 | 重庆比特数图科技有限公司 | Multi-target circulation detection tracking method and system based on motion compensation and linkage |
CN117726656A (en) * | 2024-02-08 | 2024-03-19 | 开拓导航控制技术股份有限公司 | Target tracking method, device, system and medium based on super-resolution image |
CN117726656B (en) * | 2024-02-08 | 2024-06-04 | 开拓导航控制技术股份有限公司 | Target tracking method, device, system and medium based on super-resolution image |
CN117853759A (en) * | 2024-03-08 | 2024-04-09 | 山东海润数聚科技有限公司 | Multi-target tracking method, system, equipment and storage medium |
CN117853759B (en) * | 2024-03-08 | 2024-05-10 | 山东海润数聚科技有限公司 | Multi-target tracking method, system, equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
LU102028B1 (en) | 2021-03-03 |
CN110782483B (en) | 2022-03-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110782483B (en) | Multi-view multi-target tracking method and system based on distributed camera network | |
CN108470332B (en) | Multi-target tracking method and device | |
Kendall et al. | Modelling uncertainty in deep learning for camera relocalization | |
Zhang et al. | Motion analysis | |
Boniardi et al. | Robot localization in floor plans using a room layout edge extraction network | |
JP4985516B2 (en) | Information processing apparatus, information processing method, and computer program | |
Jellal et al. | LS-ELAS: Line segment based efficient large scale stereo matching | |
CN112184757A (en) | Method and device for determining motion trail, storage medium and electronic device | |
Bashar et al. | Multiple object tracking in recent times: A literature review | |
Chang et al. | Fast Random‐Forest‐Based Human Pose Estimation Using a Multi‐scale and Cascade Approach | |
Klinger et al. | Probabilistic multi-person localisation and tracking in image sequences | |
CN109636828A (en) | Object tracking methods and device based on video image | |
CN107194950B (en) | Multi-person tracking method based on slow feature analysis | |
Urdiales et al. | An improved deep learning architecture for multi-object tracking systems | |
Wang et al. | Effective multiple pedestrian tracking system in video surveillance with monocular stationary camera | |
CN115035158A (en) | Target tracking method and device, electronic equipment and storage medium | |
CN107665495B (en) | Object tracking method and object tracking device | |
Nguyen et al. | 3d pedestrian tracking using local structure constraints | |
Ginhoux et al. | Model-based object tracking using stereo vision | |
JP4879257B2 (en) | Moving object tracking device, moving object tracking method, and moving object tracking program | |
ZHOU et al. | Background subtraction based on co-occurrence pixel-block pairs for robust object detection in dynamic scenes | |
CN108346158B (en) | Multi-target tracking method and system based on main block data association | |
Franchi et al. | Tracking hundreds of people in densely crowded scenes with particle filtering supervising deep convolutional neural networks | |
Palma-Amestoy et al. | Bayesian spatiotemporal context integration sources in robot vision systems | |
KR102589987B1 (en) | Method and Apparatus for Tracking of Online Multi-Object with Visual and Radar Features |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |