[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

WO2021114702A1 - Target tracking method, apparatus and system, and computer-readable storage medium - Google Patents

Target tracking method, apparatus and system, and computer-readable storage medium Download PDF

Info

Publication number
WO2021114702A1
WO2021114702A1 PCT/CN2020/109081 CN2020109081W WO2021114702A1 WO 2021114702 A1 WO2021114702 A1 WO 2021114702A1 CN 2020109081 W CN2020109081 W CN 2020109081W WO 2021114702 A1 WO2021114702 A1 WO 2021114702A1
Authority
WO
WIPO (PCT)
Prior art keywords
camera
detection
frame
target
tracking
Prior art date
Application number
PCT/CN2020/109081
Other languages
French (fr)
Chinese (zh)
Inventor
任培铭
刘金杰
乐振浒
张翔
林诰
Original Assignee
中国银联股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中国银联股份有限公司 filed Critical 中国银联股份有限公司
Publication of WO2021114702A1 publication Critical patent/WO2021114702A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/292Multi-camera tracking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30241Trajectory

Definitions

  • the invention belongs to the field of image processing, and specifically relates to a target tracking method, device, system and computer readable storage medium.
  • target tracking applied in the field of video surveillance has gradually become one of the hot spots in the field of computer vision research.
  • Tracking the movement trajectory of a target object usually requires acquiring an image of the surveillance area of the camera, performing target detection on the image to identify the target, and tracking the identified target object to obtain the complete trajectory of the target object.
  • Due to the complexity of the surveillance scene and the limited field of view of a single camera in order to achieve global surveillance, the cooperation of multiple cameras may be required to cover the surveillance area globally.
  • the existing multi-camera-based target tracking methods need to analyze images and achieve target tracking through deep learning methods. With the increase in the number of cameras, the demand for computing resources and communication resources increase significantly at the same time, causing a technical bottleneck for target tracking.
  • the present invention provides the following solutions.
  • a target tracking method which includes: acquiring the current frame to be measured of multiple cameras arranged in a monitoring area; performing target detection on the current frame to be measured of each of the multiple cameras in turn, and obtaining each The detection frame set corresponding to the camera; target tracking is performed according to the detection frame set corresponding to each camera, and the global target trajectory is determined according to the tracking result.
  • it further includes: determining a plurality of frame numbers to be tested, and iteratively acquiring the current frame to be tested of multiple cameras according to the number of frame numbers to be tested in time series, so as to perform target tracking iteratively; wherein, according to The initial frame number to be tested in the multiple frame numbers to be tested corresponds to the initial global target trajectory; the subsequent frame number to be tested in the frame numbers to be tested corresponds to the iteratively updated global target trajectory.
  • performing target detection on the current frame under test of each camera includes: inputting the current frame under test of each camera into a target detection model for target detection; wherein the target detection model is based on neural network training The resulting pedestrian detection model.
  • the method further includes: determining the center point of each detection frame in the detection frame set corresponding to each camera according to the viewing position of each camera Perform projection transformation to determine the ground coordinates of each detection frame.
  • the viewing areas of the multiple cameras overlap at least partially, and the method further includes: dividing the working area of each camera in the ground coordinate system according to the viewing area of each camera; wherein, the work of each camera The areas do not overlap each other. If the ground coordinates of any detection frame corresponding to the first camera of the multiple cameras exceeds the corresponding working area, any detection frame is removed from the detection frame set of the first camera.
  • the method further includes: cutting off non-critical areas in the working area of each camera.
  • tracking according to the detection frame set corresponding to each camera includes: adopting a multi-target tracking algorithm, and performing multi-target tracking based on the detection frame set corresponding to each camera, and determining the part corresponding to each camera Tracking information: Among them, the parameters used in multi-target tracking are determined based on the historical frame to be measured of each camera.
  • the multi-target tracking algorithm is a deepsort algorithm.
  • it further includes: adding an identity to each detection frame according to the local tracking information corresponding to each camera; and determining the iteratively updated global target trajectory based on the identity and ground coordinates of each detection frame.
  • it further includes: determining the association relationship between the multiple cameras according to the working areas of the multiple cameras; determining the new detection frame and the disappearing detection frame in the corresponding working area according to the local tracking information of each camera ; According to the association relationship between multiple cameras, the newly added detection frames and disappearance detection frames in different working areas are associated to obtain associated information; the iteratively updated global target trajectory is determined according to the associated information.
  • a target tracking device including: an acquisition unit for acquiring the current frame to be measured of multiple cameras arranged in a monitoring area; a detection unit for sequentially checking the current frame of each of the multiple cameras Target detection is performed on the frame to be tested to obtain a detection frame set corresponding to each camera; the tracking unit is used to perform target tracking according to the detection frame set corresponding to each camera, and determine the global target trajectory according to the tracking result.
  • it further includes: a frame selection unit for determining multiple frame numbers to be tested, and iteratively acquires the current frames to be tested from multiple cameras in time series according to the multiple frame numbers to be tested, so as to perform iteratively Target tracking; wherein, the initial global target trajectory is obtained according to the initial frame sequence number to be tested among the multiple frame numbers to be tested; the global target trajectory after iterative update is obtained corresponding to the subsequent frame sequence numbers to be tested in the multiple frame sequence numbers to be tested.
  • the detection unit is further used to: input the current frame to be measured of each camera into the target detection model for target detection; wherein the target detection model is a pedestrian detection model obtained based on neural network training.
  • the detection unit is further configured to: after obtaining the detection frame set corresponding to each camera, determine the value of each detection frame in the detection frame set corresponding to each camera according to the viewing position of each camera. Projection transformation is performed on the center point of the bottom of the frame to determine the ground coordinates of each detection frame.
  • the viewing areas of multiple cameras overlap at least partially, and the device is further used to: divide the working area of each camera in the ground coordinate system according to the viewing area of each camera; The working areas do not overlap each other. If the ground coordinates of any one of the detection frames corresponding to the first camera among the multiple cameras exceeds the corresponding working area, then any one of the detection frames is removed from the detection frame set of the first camera.
  • the detection unit is also used to: cut off non-critical areas in the working area of each camera.
  • the tracking unit is also used to: adopt a multi-target tracking algorithm, and perform multi-target tracking based on the detection frame set corresponding to each camera, and determine the local tracking information corresponding to each camera; The parameters used in tracking are determined based on the historical frames to be measured for each camera.
  • the multi-target tracking algorithm is a deepsort algorithm.
  • the tracking unit is also used to: add an identity to each detection frame according to the local tracking information corresponding to each camera; to determine the iteratively updated global based on the identity and ground coordinates of each detection frame Target trajectory.
  • the tracking unit is further configured to: determine the association relationship between the multiple cameras according to the work areas of the multiple cameras; determine the new detection frame in the corresponding work area according to the local tracking information of each camera And disappear detection frame; according to the association relationship between multiple cameras, the newly added detection frame and disappearance detection frame in different working areas are associated to obtain the associated information; the global target track after iterative update is determined according to the associated information.
  • a target tracking system including: a plurality of cameras arranged in a monitoring area, and a target tracking device respectively communicatively connected with the plurality of cameras; wherein the target tracking device is configured to perform as in the first aspect Methods.
  • a target tracking device including: one or more multi-core processors; a memory for storing one or more programs; when one or more programs are executed by one or more multi-core processors,
  • One or more multi-core processors realize: obtain the current frame to be tested of multiple cameras set in the monitoring area; perform target detection on the current frame to be tested of each of the multiple cameras in turn to obtain the corresponding detection of each camera Frame set; target tracking is performed according to the detection frame set corresponding to each camera, and the global target trajectory is determined according to the tracking result.
  • a computer-readable storage medium stores a program.
  • the program is executed by a multi-core processor, the multi-core processor is caused to execute the method of the first aspect.
  • the current frame to be tested from each camera is detected in sequence, and then based on the detection result corresponding to each camera in the monitoring area
  • Global tracking can realize global tracking of target objects in multi-channel surveillance videos based on less computing resources, and can realize target tracking based on multiple cameras based on less computing resources.
  • Fig. 1 is a schematic flowchart of a target tracking method according to an embodiment of the present invention
  • FIG. 2 is a schematic diagram of the ground of a monitoring area according to an embodiment of the present invention.
  • FIG. 3 is a schematic diagram of viewfinder images of multiple cameras according to an embodiment of the present invention.
  • FIG. 4 is a schematic diagram of current frames to be measured of multiple cameras according to an embodiment of the present invention.
  • FIG. 5 is a schematic diagram of a set of detection frames corresponding to multiple cameras according to an embodiment of the present invention.
  • Fig. 6 is a schematic diagram of a global target trajectory according to an embodiment of the present invention.
  • Fig. 7 is a schematic structural diagram of a target tracking device according to an embodiment of the present invention.
  • FIG. 8 is a schematic structural diagram of a target tracking device according to another embodiment of the present invention.
  • Fig. 9 is a schematic diagram of a computer-readable storage medium according to an embodiment of the present invention.
  • FIG. 1 schematically shows a schematic flowchart of a target tracking method 100 according to an embodiment of the present invention
  • the method 100 may include:
  • Step S101 Obtain current frames to be tested of multiple cameras set in the monitoring area
  • the monitoring area refers to the sum of the viewing areas of multiple cameras.
  • the multiple cameras include at least two cameras, and the viewing areas of the multiple cameras are adjacent to each other or at least partially overlapped, so that the target object to be tracked can be Move in the monitoring area and then appear in the viewing area of any one or more cameras.
  • the current frames to be measured of the multiple cameras are respectively extracted from the surveillance videos of the multiple cameras, and the current frames to be measured of each camera have the same acquisition time.
  • the target to be tracked in this disclosure is preferably a pedestrian.
  • the target to be tracked may also be other movable objects, such as animals, vehicles, etc., which is not specifically limited in the present disclosure.
  • FIG. 2 shows a schematic monitoring scene in which a camera 201 and a camera 202 are set
  • FIG. 3 shows a viewfinder screen of the above-mentioned camera 201 and the camera 202.
  • the surveillance video of the camera 201 can be parsed as a sequence of image frames (A 1 , A 2 ,..., A N )
  • the surveillance video of the camera 202 can be parsed as a sequence of image frames (B 1 , B 2 ,..., B N ), where the above analysis can be performed online or offline in real time.
  • the method 100 may further include: determining a plurality of frame numbers to be measured, and iteratively acquiring the current frame to be measured of multiple cameras according to the plurality of frame numbers to be measured in time series, so as to iteratively perform target tracking ; Among them, the initial global target trajectory is obtained corresponding to the initial frame number to be tested among the multiple frame numbers to be tested; the global target trajectory after iterative update is obtained corresponding to the subsequent frame number to be tested in the multiple frame numbers to be tested. This can reduce the amount of calculations and improve the real-time performance of global tracking,
  • the method 100 may further include:
  • Step S102 Perform target detection on the current frame to be tested of each of the multiple cameras in turn, to obtain a set of detection frames corresponding to each camera;
  • performing target detection on the current frame under test of each camera includes: inputting the current frame under test of each camera into a target detection model for target detection; wherein the target detection model is based on neural network training The resulting pedestrian detection model.
  • the current frames A n and B n of the camera 201 and the camera 202 to be tested are shown, and then the pre-processed current frame A n is input into any deep learning-based pedestrian detection model.
  • the purpose of obtaining the pedestrian detection frame is to obtain the position information and size information of all pedestrians in the current frames A n and B n to be tested.
  • the pedestrian detection model may be, for example, a YOLO (Unified Real-Time Object Detection, You Only Look Once) model, etc., which is not specifically limited in the present disclosure.
  • YOLO Unified Real-Time Object Detection, You Only Look Once
  • the detection frame set corresponding to each camera further includes: according to the viewing position of each camera and the center of each detection frame in the detection frame set corresponding to each camera The points undergo projection transformation to determine the ground coordinates of each detection frame in the detection frame set corresponding to each camera. In this way, the targets identified in the viewing range of each camera can be combined into a unified coordinate system.
  • the position of the center point of the bottom of each detection frame corresponding to each camera in Figure 5 can be obtained, and the position of the center point of the bottom of the frame of each detection frame can be converted to obtain the actual ground position of the target object in the monitoring scene.
  • Fig. 6 shows the ground coordinates of each detection frame obtained through projection transformation. Specifically, it can be seen that the ground aisle under the viewing angle of each camera is an approximate trapezoidal area, so for the detection frame set corresponding to each camera, the center point of the bottom of each detection frame can be obtained through the trapezoid-rectangular transformation.
  • the coordinates in the standard rectangular area Secondly, the standard rectangular area is rotated according to the actual layout of the monitoring scene.
  • the rotated coordinates of the center point of the bottom of each detection frame are calculated by the rotation matrix, and finally the rotation is rotated according to the actual layout of the monitoring scene. After the coordinates are translated and zoomed, the final coordinate position is obtained.
  • the viewing areas of the multiple cameras overlap at least partially, and the method further includes: dividing the working area of each camera in the ground coordinate system according to the viewing area of each camera; wherein, the work of each camera The areas do not overlap each other. If the ground coordinates of any detection frame corresponding to the first camera of the multiple cameras exceeds the corresponding working area, any detection frame is removed from the detection frame set of the first camera.
  • the working area of each camera can be divided.
  • the working area of the camera 201 is the X area
  • the working area of the camera 202 is the Y area, so that the working area of each camera is Adjacent.
  • the ground coordinates of each detection frame corresponding to each camera need to be located in the working area of the camera, and removed if it is not in the working area of the camera.
  • the detection frame a is removed from the detection frame set corresponding to the camera 201 3 , get (a 1 , a 2 ) for subsequent operations.
  • the method further includes: cutting off non-critical areas in the working area of each camera. Specifically, whether it is a critical area can be determined based on the specific layout of the monitoring scene. For example, the ceiling area that cannot be passed by pedestrians can be directly cut off, which can reduce the amount of calculation for target tracking.
  • the method 100 may further include:
  • Step S103 Perform target tracking according to the detection frame set corresponding to each camera, and update the global target trajectory according to the tracking result.
  • target detection can be performed according to the initial current frame to be measured A 1 and B 1 to determine the initial global target trajectory. Further, target detection may be performed according to the current frames A n and B n to be tested subsequently obtained, and target tracking may be performed iteratively according to the target detection result, so as to iteratively update the global target trajectory.
  • tracking according to the detection frame set corresponding to each camera includes: adopting a multi-target tracking algorithm, and performing multi-target tracking based on the detection frame set corresponding to each camera, and determining the part corresponding to each camera Tracking information: Among them, the parameters used in multi-target tracking are determined based on the historical frame to be measured of each camera. This enables multi-target tracking in the monitoring area.
  • the multi-target tracking algorithm is a target tracking algorithm based on a single camera, such as the DeepSORT algorithm (Simple Online and Realtime Tracking with a Deep Association Metric), so you can get the information of each camera. Local tracking information.
  • the parameters used in multi-target tracking are determined based on the historical frame to be measured of each camera.
  • the target frame to be tracked can be determined when any target appears in the working area of a certain camera for the first time, and based on the multi-target
  • the detection algorithm and the identified target frame track the subsequent frame to be measured of the camera, and determine the local tracking information of the target in the working area of the camera.
  • the multi-target tracking algorithm is a deepsort algorithm.
  • other target tracking algorithms can also be used, and those skilled in the art can understand that what the present disclosure intends to emphasize is not which target tracking algorithm is specifically used.
  • updating the global target trajectory according to the tracking result also includes: adding an identity to each detection frame according to the local tracking information corresponding to each camera; based on the identity, using the ground coordinate pair of each detection frame The global target trajectory is updated.
  • the curve part shows the current global target trajectory, that is, the global target trajectory determined in the last iteration, and the points a 1 , a 2 and points b respectively represents the ground coordinates of the multiple detection frames shown in FIG. 5.
  • the detection frame a 2 is labeled "target 2" and the ground coordinates of the point a 2 are added to "target 2" existing track (i.e., the "target 2" dashed curve in FIG. 6)
  • the camera 201 corresponding local trace information indicative of a 1-point detection frame matches the target does not exist, add a 1 compared with a detection frame labeled " Goal 3" and create a new trajectory of "Goal 3".
  • updating the global target trajectory according to the tracking result further includes: determining the association relationship between the multiple cameras by the working areas of the multiple cameras; determining the new information in the corresponding working area according to the local tracking information of each camera Add detection frames and disappearance detection frames; associate new detection frames and disappearance detection frames in different work areas according to the association relationship between multiple cameras to obtain associated information; update the global target trajectory according to the associated information.
  • the association relationship between the multiple cameras is, for example, that the area X and the area Y are adjacent to each other at a specified position, so that when the target moves, different work areas can be straddled from adjacent positions based on the above association relationship.
  • the association information refers to the association between a new detection frame in a certain work area and a disappearance detection frame in another work area, that is, corresponding to the same identity.
  • the disappearance order of multiple tracking targets can be obtained at the adjoining boundary of one of the work areas, and the disappearance order of the pair appears in the adjoining boundary in the other work area.
  • the multiple newly-added targets at the location are assigned corresponding identifiers and continue to be tracked,
  • the point b in the area Y represents the ground coordinates of the detection frame b shown in FIG. 5. If the local tracking information corresponding to the camera 201 indicates that there is no matching target at the detection frame point b, that is, there is a new target in the area Y; and the local tracking information corresponding to the camera 201 indicates that the continuously tracked "target 1" is currently detected If the frame disappears, that is, there is a disappearing target in area X, you can mark the detection frame b with "Target 1" and add the ground coordinates of point b to the existing trajectory of "Target 1" (ie, "Target 1" in Figure 6). "Dashed curve), to achieve cross-camera, cross-working area target tracking.
  • the multi-camera-based target tracking method of the present invention by sequentially performing image detection on the current frame to be tested from each camera, and then performing global tracking in the monitoring area based on the detection result corresponding to each camera, it can be based on Fewer computing resources enable global tracking of target objects in multi-channel surveillance videos, reducing the demand for computing resources. For example, there is no need to separately provide GPU computing resources for tracking the target object in each local area for each camera, but fewer computing resources can be provided for global tracking of the target object in the monitoring area.
  • an embodiment of the present invention also provides a target tracking device for executing the target tracking method provided in any of the foregoing embodiments.
  • Fig. 7 is a schematic structural diagram of a target tracking device provided by an embodiment of the present invention.
  • the apparatus 700 includes:
  • the acquiring unit 701 is configured to acquire current frames to be measured of multiple cameras arranged in the monitoring area;
  • the detection unit 702 is configured to sequentially perform target detection on the current frame to be tested of each of the multiple cameras to obtain a set of detection frames corresponding to each camera;
  • the tracking unit 703 is configured to perform target tracking according to the detection frame set corresponding to each camera, and determine the global target trajectory according to the tracking result.
  • the apparatus 700 further includes: a frame selection unit, configured to determine multiple frame numbers to be tested, and iteratively obtain current frames to be tested from multiple cameras in a time sequence according to the multiple frame numbers to be tested, so as to iterate To perform target tracking; among them, the initial global target trajectory is obtained according to the initial test frame sequence number among the multiple test frame numbers; the global target trajectory after iterative update is obtained according to the subsequent test frame sequence numbers among the multiple test frame sequence numbers. .
  • a frame selection unit configured to determine multiple frame numbers to be tested, and iteratively obtain current frames to be tested from multiple cameras in a time sequence according to the multiple frame numbers to be tested, so as to iterate To perform target tracking; among them, the initial global target trajectory is obtained according to the initial test frame sequence number among the multiple test frame numbers; the global target trajectory after iterative update is obtained according to the subsequent test frame sequence numbers among the multiple test frame sequence numbers.
  • the detection unit 702 is further configured to: input the current frame to be measured of each camera into a target detection model for target detection; wherein the target detection model is a pedestrian detection model obtained based on neural network training.
  • the detection unit 702 is further configured to: after obtaining the detection frame set corresponding to each camera, determine each detection frame in the detection frame set corresponding to each camera according to the viewing position of each camera. Projection transformation is performed on the center point of the bottom of the frame to determine the ground coordinates of each detection frame.
  • the viewing areas of multiple cameras overlap at least partially, and the device 700 is further configured to: divide the working area of each camera in the ground coordinate system according to the viewing area of each camera; wherein, each camera If the ground coordinates of any detection frame corresponding to the first camera of the multiple cameras are beyond the corresponding working area, any detection frame is removed from the detection frame set of the first camera.
  • the detection unit 702 is also used to cut off non-critical areas in the working area of each camera.
  • the tracking unit 703 is further configured to: adopt a multi-target tracking algorithm and perform multi-target tracking based on the detection frame set corresponding to each camera, and determine the local tracking information corresponding to each camera; The parameters used for target tracking are determined based on the historical frames to be tested for each camera.
  • the multi-target tracking algorithm is a deepsort algorithm.
  • the tracking unit 703 is further configured to: add an identity to each detection frame according to the local tracking information corresponding to each camera; to determine the iteratively updated information based on the identity and ground coordinates of each detection frame Global target trajectory.
  • the tracking unit 703 is further configured to: determine the association relationship between the multiple cameras according to the work areas of the multiple cameras; determine the new detection in the corresponding work area according to the local tracking information of each camera Frames and disappearance detection frames; associate new detection frames and disappearance detection frames in different work areas according to the association relationship between multiple cameras to obtain association information; determine the iteratively updated global target trajectory according to the association information.
  • the current frame to be measured from each camera is detected in sequence, and then global tracking is performed in the monitoring area based on the detection result corresponding to each camera, which can be based on Fewer computing resources enable global tracking of target objects in multi-channel surveillance videos, reducing the demand for computing resources. For example, there is no need to separately provide GPU computing resources for tracking the target object in each local area for each camera, but fewer computing resources can be provided for global tracking of the target object in the monitoring area.
  • an embodiment of the present invention also provides a target tracking system, which specifically includes: a plurality of cameras arranged in a monitoring area, and a target tracking device respectively communicatively connected with the plurality of cameras; wherein the target tracking device is It is configured to execute the target tracking method provided in any of the above embodiments.
  • a target tracking device of the present invention may at least include one or more processors and at least one memory.
  • the memory stores a program, and when the program is executed by the processor, the processor is caused to perform the steps shown in FIG. 1: acquiring the current frame to be measured of multiple cameras arranged in the monitoring area ;Sequentially perform target detection on the current frame under test of each camera in the multiple cameras to obtain the detection frame set corresponding to each camera; perform target tracking according to the detection frame set corresponding to each camera, and determine the global target trajectory according to the tracking result.
  • the target tracking device 8 according to this embodiment of the present invention will be described below with reference to FIG. 8.
  • the device 8 shown in FIG. 8 is only an example, and should not bring any limitation to the function and application scope of the embodiment of the present invention.
  • the apparatus 8 may be in the form of a general-purpose computing device, including but not limited to: at least one processor 10, at least one memory 20, and a bus 60 connecting different device components.
  • the bus 60 includes a data bus, an address bus, and a control bus.
  • the memory 20 may include a volatile memory, such as a random access memory (RAM) 21 and/or a cache memory 22, and may further include a read-only memory (ROM) 23.
  • RAM random access memory
  • ROM read-only memory
  • the memory 20 may also include a program module 24.
  • program module 24 includes, but is not limited to, an operating device, one or more application programs, other program modules, and program data. Each of these examples or a certain combination may include a network. The realization of the environment.
  • the apparatus 5 can also communicate with one or more external devices 2 (for example, a keyboard, a pointing device, a Bluetooth device, etc.), and can also communicate with one or more other devices. Such communication can be performed through an input/output (I/O) interface 40 and displayed on the display unit 30.
  • the device 5 may also communicate with one or more networks (for example, a local area network (LAN), a wide area network (WAN), and/or a public network, such as the Internet) through the network adapter 50. As shown in the figure, the network adapter 50 communicates with other modules in the device 5 through the bus 60.
  • LAN local area network
  • WAN wide area network
  • public network such as the Internet
  • Fig. 9 shows a computer-readable storage medium for executing the method as described above.
  • various aspects of the present invention can also be implemented in the form of a computer-readable storage medium, which includes program code.
  • program code When the program code is executed by a processor, the program code is used for The processor is caused to execute the method described above.
  • the computer-readable storage medium may adopt any combination of one or more readable media.
  • the readable medium may be a readable signal medium or a readable storage medium.
  • the readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor device, device, or device, or a combination of any of the above. More specific examples (non-exhaustive list) of readable storage media include: electrical connections with one or more wires, portable disks, hard disks, random access memory (RAM), read-only memory (ROM), erasable Type programmable read only memory (EPROM or flash memory), optical fiber, portable compact disk read only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
  • a computer-readable storage medium 90 As shown in FIG. 9, a computer-readable storage medium 90 according to an embodiment of the present invention is described. It can adopt a portable compact disk read-only memory (CD-ROM) and include program codes, and can be stored in a terminal device, such as a personal computer. Run on.
  • the computer-readable storage medium of the present invention is not limited to this.
  • the readable storage medium can be any tangible medium that contains or stores a program.
  • the program can be used by or in combination with an instruction execution device, device, or device. .
  • the program code used to perform the operations of the present invention can be written in any combination of one or more programming languages.
  • the programming languages include object-oriented programming languages—such as Java, Python, C++, etc., as well as conventional Procedural programming language-such as "C" language or similar programming language.
  • the program code can be executed entirely on the user's computing device, partly executed on the user's device and partly executed on the remote computing device, or entirely executed on the remote computing device or server.
  • the remote computing device can be connected to the user computing device through any kind of network-including a local area network (LAN) or a wide area network (WAN), or it can be connected to an external computing device (for example, using Internet services). Provider to connect via the Internet).
  • LAN local area network
  • WAN wide area network
  • Internet services for example, using Internet services

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)
  • Closed-Circuit Television Systems (AREA)
  • Burglar Alarm Systems (AREA)

Abstract

A target tracking method, apparatus and system, and a computer-readable storage medium. The method comprises: obtaining current frames to be detected of multiple cameras disposed in a monitoring area (step S101); performing target detection on the current frame to be detected of each of the multiple cameras in sequence to obtain a bounding box set corresponding to each camera (step S102); and performing target tracking according to the bounding box set corresponding to each camera, and determining a global target trajectory according to a tracking result (step S103). Using the method above can reduce computing resources for target tracking based on multiple cameras.

Description

一种目标跟踪方法、装置、系统及计算机可读存储介质Target tracking method, device, system and computer readable storage medium 技术领域Technical field
本发明属于图像处理领域,具体涉及一种目标跟踪方法、装置、系统及计算机可读存储介质。The invention belongs to the field of image processing, and specifically relates to a target tracking method, device, system and computer readable storage medium.
背景技术Background technique
本部分旨在为权利要求书中陈述的本发明的实施方式提供背景或上下文。此处的描述不因为包括在本部分中就承认是现有技术。This section is intended to provide background or context for the embodiments of the invention stated in the claims. The description here is not recognized as prior art just because it is included in this section.
目前,随着视频监控技术的普及以及不断提升的安防需求,应用于视频监控领域中的目标跟踪逐渐成为计算机视觉研究领域的热点之一。追踪目标对象的移动轨迹通常需要获取摄像头的监控区域图像,对图像进行目标检测以识别目标,并对识别出的目标对象进行跟踪从而可以得到目标对象的完整轨迹。由于监控场景的复杂性,且单个摄像头视野范围是有限的,所以为了实现全局监控,可能需要多个摄像头的配合才能进行监控区域的全局覆盖。然而,现有的基于多摄像头的目标跟踪方法需要通过深度学习方法分析图像并实现目标跟踪,随着摄像头数量的增加,计算资源需求和通信资源需求同时大幅增加,造成目标跟踪的技术瓶颈。At present, with the popularization of video surveillance technology and the ever-increasing security requirements, target tracking applied in the field of video surveillance has gradually become one of the hot spots in the field of computer vision research. Tracking the movement trajectory of a target object usually requires acquiring an image of the surveillance area of the camera, performing target detection on the image to identify the target, and tracking the identified target object to obtain the complete trajectory of the target object. Due to the complexity of the surveillance scene and the limited field of view of a single camera, in order to achieve global surveillance, the cooperation of multiple cameras may be required to cover the surveillance area globally. However, the existing multi-camera-based target tracking methods need to analyze images and achieve target tracking through deep learning methods. With the increase in the number of cameras, the demand for computing resources and communication resources increase significantly at the same time, causing a technical bottleneck for target tracking.
发明内容Summary of the invention
针对上述现有技术中存在的问题,提出了一种目标跟踪方法、装置及计算机可读存储介质,利用这种方法、装置及计算机可读存储介质,能够解决上述问题。In view of the above-mentioned problems in the prior art, a target tracking method, device, and computer-readable storage medium are proposed. By using this method, device and computer-readable storage medium, the above-mentioned problems can be solved.
本发明提供了以下方案。The present invention provides the following solutions.
第一方面,提供一种目标跟踪方法,包括:获取设置于监控区域内的多个摄像头的当前待测帧;依次对多个摄像头中每个摄像头的当前待测帧进行目标 检测,得到每个摄像头对应的检测框集合;根据每个摄像头对应的检测框集合进行目标跟踪,根据跟踪结果确定全局目标轨迹。In a first aspect, a target tracking method is provided, which includes: acquiring the current frame to be measured of multiple cameras arranged in a monitoring area; performing target detection on the current frame to be measured of each of the multiple cameras in turn, and obtaining each The detection frame set corresponding to the camera; target tracking is performed according to the detection frame set corresponding to each camera, and the global target trajectory is determined according to the tracking result.
在一些可能的实施方式中,还包括:确定多个待测帧序号,根据多个待测帧序号按时序地迭代获取多个摄像头的当前待测帧,从而迭代地执行目标跟踪;其中,根据多个待测帧序号中初始待测帧序号对应得到初始的全局目标轨迹;根据多个待测帧序号中后续待测帧序号对应得到迭代更新后的全局目标轨迹。In some possible implementation manners, it further includes: determining a plurality of frame numbers to be tested, and iteratively acquiring the current frame to be tested of multiple cameras according to the number of frame numbers to be tested in time series, so as to perform target tracking iteratively; wherein, according to The initial frame number to be tested in the multiple frame numbers to be tested corresponds to the initial global target trajectory; the subsequent frame number to be tested in the frame numbers to be tested corresponds to the iteratively updated global target trajectory.
在一些可能的实施方式中,对每个摄像头的当前待测帧进行目标检测,包括:将每个摄像头的当前待测帧输入目标检测模型进行目标检测;其中,目标检测模型是基于神经网络训练得到的行人检测模型。In some possible implementation manners, performing target detection on the current frame under test of each camera includes: inputting the current frame under test of each camera into a target detection model for target detection; wherein the target detection model is based on neural network training The resulting pedestrian detection model.
在一些可能的实施方式中,在得到每个摄像头对应的检测框集合之后,还包括:根据每个摄像头的取景位置对每个摄像头对应的检测框集合中的每个检测框的框底中心点进行投影变换,从而确定每个检测框的地面坐标。In some possible implementation manners, after the detection frame set corresponding to each camera is obtained, the method further includes: determining the center point of each detection frame in the detection frame set corresponding to each camera according to the viewing position of each camera Perform projection transformation to determine the ground coordinates of each detection frame.
在一些可能的实施方式中,多个摄像头的取景区域至少部分地重叠,方法还包括:根据每个摄像头的取景区域在地面坐标系中划分每个摄像头的工作区域;其中,每个摄像头的工作区域互不重叠,若多个摄像头中的第一摄像头对应的任意一个检测框的地面坐标超出对应的工作区域,则在第一摄像头的检测框集合中去除任意一个检测框。In some possible implementation manners, the viewing areas of the multiple cameras overlap at least partially, and the method further includes: dividing the working area of each camera in the ground coordinate system according to the viewing area of each camera; wherein, the work of each camera The areas do not overlap each other. If the ground coordinates of any detection frame corresponding to the first camera of the multiple cameras exceeds the corresponding working area, any detection frame is removed from the detection frame set of the first camera.
在一些可能的实施方式中,方法还包括:将每个摄像头的工作区域中的非关键区域截去。In some possible implementation manners, the method further includes: cutting off non-critical areas in the working area of each camera.
在一些可能的实施方式中,根据每个摄像头对应的检测框集合进行跟踪,包括:采用多目标跟踪算法,并基于每个摄像头对应的检测框集合进行多目标跟踪,确定每个摄像头对应的局部跟踪信息;其中,多目标跟踪采用的参数基于每个摄像头的历史待测帧而确定。In some possible implementation manners, tracking according to the detection frame set corresponding to each camera includes: adopting a multi-target tracking algorithm, and performing multi-target tracking based on the detection frame set corresponding to each camera, and determining the part corresponding to each camera Tracking information: Among them, the parameters used in multi-target tracking are determined based on the historical frame to be measured of each camera.
在一些可能的实施方式中,多目标跟踪算法为deepsort算法。In some possible implementation manners, the multi-target tracking algorithm is a deepsort algorithm.
在一些可能的实施方式中,还包括:根据每个摄像头对应的局部跟踪信息为每个检测框添加身份标识;基于每个检测框的身份标识和地面坐标确定迭代更新后的全局目标轨迹。In some possible implementation manners, it further includes: adding an identity to each detection frame according to the local tracking information corresponding to each camera; and determining the iteratively updated global target trajectory based on the identity and ground coordinates of each detection frame.
在一些可能的实施方式中,还包括:根据多个摄像头的工作区域确定多个摄像头之间的关联关系;根据每个摄像头的局部跟踪信息确定对应工作区域中的新增检测框和消失检测框;根据多个摄像头之间的关联关系对处于不同工作区域中的新增检测框和消失检测框进行关联,得到关联信息;根据关联信息确定迭代更新后的全局目标轨迹。In some possible implementation manners, it further includes: determining the association relationship between the multiple cameras according to the working areas of the multiple cameras; determining the new detection frame and the disappearing detection frame in the corresponding working area according to the local tracking information of each camera ; According to the association relationship between multiple cameras, the newly added detection frames and disappearance detection frames in different working areas are associated to obtain associated information; the iteratively updated global target trajectory is determined according to the associated information.
第二方面,提供一种目标跟踪装置,包括:获取单元,用于获取设置于监控区域内的多个摄像头的当前待测帧;检测单元,用于依次对多个摄像头中每个摄像头的当前待测帧进行目标检测,得到每个摄像头对应的检测框集合;跟踪单元,用于根据每个摄像头对应的检测框集合进行目标跟踪,根据跟踪结果确定全局目标轨迹。In a second aspect, a target tracking device is provided, including: an acquisition unit for acquiring the current frame to be measured of multiple cameras arranged in a monitoring area; a detection unit for sequentially checking the current frame of each of the multiple cameras Target detection is performed on the frame to be tested to obtain a detection frame set corresponding to each camera; the tracking unit is used to perform target tracking according to the detection frame set corresponding to each camera, and determine the global target trajectory according to the tracking result.
在一些可能的实施方式中,还包括:选帧单元,用于确定多个待测帧序号,根据多个待测帧序号按时序地迭代获取多个摄像头的当前待测帧,从而迭代地执行目标跟踪;其中,根据多个待测帧序号中初始待测帧序号对应得到初始的全局目标轨迹;根据多个待测帧序号中后续待测帧序号对应得到迭代更新后的全局目标轨迹。In some possible implementation manners, it further includes: a frame selection unit for determining multiple frame numbers to be tested, and iteratively acquires the current frames to be tested from multiple cameras in time series according to the multiple frame numbers to be tested, so as to perform iteratively Target tracking; wherein, the initial global target trajectory is obtained according to the initial frame sequence number to be tested among the multiple frame numbers to be tested; the global target trajectory after iterative update is obtained corresponding to the subsequent frame sequence numbers to be tested in the multiple frame sequence numbers to be tested.
在一些可能的实施方式中,检测单元,还用于:将每个摄像头的当前待测帧输入目标检测模型进行目标检测;其中,目标检测模型是基于神经网络训练得到的行人检测模型。In some possible implementation manners, the detection unit is further used to: input the current frame to be measured of each camera into the target detection model for target detection; wherein the target detection model is a pedestrian detection model obtained based on neural network training.
在一些可能的实施方式中,检测单元,还用于:在得到每个摄像头对应的检测框集合之后,根据每个摄像头的取景位置对每个摄像头对应的检测框集合中的每个检测框的框底中心点进行投影变换,从而确定每个检测框的地面坐标。In some possible implementation manners, the detection unit is further configured to: after obtaining the detection frame set corresponding to each camera, determine the value of each detection frame in the detection frame set corresponding to each camera according to the viewing position of each camera. Projection transformation is performed on the center point of the bottom of the frame to determine the ground coordinates of each detection frame.
在一些可能的实施方式中,多个摄像头的取景区域至少部分地重叠,装置还用于:根据每个摄像头的取景区域在地面坐标系中划分每个摄像头的工作区域;其中,每个摄像头的工作区域互不重叠,若多个摄像头中的第一摄像头对应的任意一个检测框的地面坐标超出对应的工作区域,则在第一摄像头的检测框集合中去除任意一个检测框。In some possible implementation manners, the viewing areas of multiple cameras overlap at least partially, and the device is further used to: divide the working area of each camera in the ground coordinate system according to the viewing area of each camera; The working areas do not overlap each other. If the ground coordinates of any one of the detection frames corresponding to the first camera among the multiple cameras exceeds the corresponding working area, then any one of the detection frames is removed from the detection frame set of the first camera.
在一些可能的实施方式中,检测单元,还用于:将每个摄像头的工作区域中的非关键区域截去。In some possible implementation manners, the detection unit is also used to: cut off non-critical areas in the working area of each camera.
在一些可能的实施方式中,跟踪单元,还用于:采用多目标跟踪算法,并基于每个摄像头对应的检测框集合进行多目标跟踪,确定每个摄像头对应的局部跟踪信息;其中,多目标跟踪采用的参数基于每个摄像头的历史待测帧而确定。In some possible implementations, the tracking unit is also used to: adopt a multi-target tracking algorithm, and perform multi-target tracking based on the detection frame set corresponding to each camera, and determine the local tracking information corresponding to each camera; The parameters used in tracking are determined based on the historical frames to be measured for each camera.
在一些可能的实施方式中,多目标跟踪算法为deepsort算法。In some possible implementation manners, the multi-target tracking algorithm is a deepsort algorithm.
在一些可能的实施方式中,跟踪单元,还用于:根据每个摄像头对应的局部跟踪信息为每个检测框添加身份标识;基于每个检测框的身份标识和地面坐标确定迭代更新后的全局目标轨迹。In some possible implementations, the tracking unit is also used to: add an identity to each detection frame according to the local tracking information corresponding to each camera; to determine the iteratively updated global based on the identity and ground coordinates of each detection frame Target trajectory.
在一些可能的实施方式中,跟踪单元,还用于:根据多个摄像头的工作区域确定多个摄像头之间的关联关系;根据每个摄像头的局部跟踪信息确定对应工作区域中的新增检测框和消失检测框;根据多个摄像头之间的关联关系对处于不同工作区域中的新增检测框和消失检测框进行关联,得到关联信息;根据关联信息确定迭代更新后的全局目标轨迹。In some possible implementation manners, the tracking unit is further configured to: determine the association relationship between the multiple cameras according to the work areas of the multiple cameras; determine the new detection frame in the corresponding work area according to the local tracking information of each camera And disappear detection frame; according to the association relationship between multiple cameras, the newly added detection frame and disappearance detection frame in different working areas are associated to obtain the associated information; the global target track after iterative update is determined according to the associated information.
第三方面,提供一种目标跟踪系统,包括:设置于监控区域内的多个摄像头,以及与多个摄像头分别通信连接的目标跟踪装置;其中,目标跟踪装置被配置用于执行如第一方面的方法。In a third aspect, a target tracking system is provided, including: a plurality of cameras arranged in a monitoring area, and a target tracking device respectively communicatively connected with the plurality of cameras; wherein the target tracking device is configured to perform as in the first aspect Methods.
第四方面,提供一种目标跟踪装置,包括:一个或者多个多核处理器;存储器,用于存储一个或多个程序;当一个或多个程序被一个或者多个多核处理器执行时,使得一个或多个多核处理器实现:获取设置于监控区域内的多个摄像头的当前待测帧;依次对多个摄像头中每个摄像头的当前待测帧进行目标检测,得到每个摄像头对应的检测框集合;根据每个摄像头对应的检测框集合进行目标跟踪,根据跟踪结果确定全局目标轨迹。In a fourth aspect, a target tracking device is provided, including: one or more multi-core processors; a memory for storing one or more programs; when one or more programs are executed by one or more multi-core processors, One or more multi-core processors realize: obtain the current frame to be tested of multiple cameras set in the monitoring area; perform target detection on the current frame to be tested of each of the multiple cameras in turn to obtain the corresponding detection of each camera Frame set; target tracking is performed according to the detection frame set corresponding to each camera, and the global target trajectory is determined according to the tracking result.
第五方面,提供一种计算机可读存储介质,计算机可读存储介质存储有程序,当程序被多核处理器执行时,使得多核处理器执行如第一方面的方法。In a fifth aspect, a computer-readable storage medium is provided, and the computer-readable storage medium stores a program. When the program is executed by a multi-core processor, the multi-core processor is caused to execute the method of the first aspect.
本申请实施例采用的上述至少一个技术方案能够达到以下有益效果:本实施例中,通过依次对来各个摄像头的当前待测帧进行图像检测,然后基于对应于各个摄像头的检测结果在监控区域中进行全局追踪,可以基于较少的计算资源实现对多路监控视频中的目标对象实现全局的跟踪,能够基于较少的计算资源实现基于多摄像头的目标跟踪。The above-mentioned at least one technical solution adopted in the embodiment of this application can achieve the following beneficial effects: In this embodiment, the current frame to be tested from each camera is detected in sequence, and then based on the detection result corresponding to each camera in the monitoring area Global tracking can realize global tracking of target objects in multi-channel surveillance videos based on less computing resources, and can realize target tracking based on multiple cameras based on less computing resources.
应当理解,上述说明仅是本发明技术方案的概述,以便能够更清楚地了解本发明的技术手段,从而可依照说明书的内容予以实施。为了让本发明的上述和其它目的、特征和优点能够更明显易懂,以下特举例说明本发明的具体实施方式。It should be understood that the above description is only an overview of the technical solution of the present invention, so that the technical means of the present invention can be understood more clearly, so that it can be implemented in accordance with the content of the description. In order to make the above and other objects, features and advantages of the present invention more obvious and understandable, the following examples illustrate the specific embodiments of the present invention.
附图说明Description of the drawings
通过阅读下文的示例性实施例的详细描述,本领域普通技术人员将明白本文所述的有点和益处以及其他优点和益处。附图仅用于示出示例性实施例的目的,而并不认为是对本发明的限制。而且在整个附图中,用相同的标号表示相同的部件。在附图中:By reading the detailed description of the exemplary embodiments below, those of ordinary skill in the art will understand the advantages and benefits described herein as well as other advantages and benefits. The drawings are only used for the purpose of illustrating exemplary embodiments, and are not considered as a limitation to the present invention. Moreover, the same reference numerals are used to denote the same components throughout the drawings. In the attached picture:
图1为根据本发明一实施例的目标跟踪方法的流程示意图;Fig. 1 is a schematic flowchart of a target tracking method according to an embodiment of the present invention;
图2为根据本发明一实施例的监控区域的地面示意图;2 is a schematic diagram of the ground of a monitoring area according to an embodiment of the present invention;
图3为根据本发明一实施例的多个摄像头的取景画面示意图;3 is a schematic diagram of viewfinder images of multiple cameras according to an embodiment of the present invention;
图4为根据本发明一实施例的多个摄像头的当前待测帧的示意图;4 is a schematic diagram of current frames to be measured of multiple cameras according to an embodiment of the present invention;
图5为根据本发明一实施例的多个摄像头对应的检测框集合的示意图;5 is a schematic diagram of a set of detection frames corresponding to multiple cameras according to an embodiment of the present invention;
图6为根据本发明一实施例的全局目标轨迹的示意图;Fig. 6 is a schematic diagram of a global target trajectory according to an embodiment of the present invention;
图7为根据本发明一实施例的目标跟踪装置的结构示意图;Fig. 7 is a schematic structural diagram of a target tracking device according to an embodiment of the present invention;
图8为根据本发明另一实施例的目标跟踪装置的结构示意图;FIG. 8 is a schematic structural diagram of a target tracking device according to another embodiment of the present invention;
图9为根据本发明一实施例的计算机可读存储介质的示意图。Fig. 9 is a schematic diagram of a computer-readable storage medium according to an embodiment of the present invention.
在附图中,相同或对应的标号表示相同或对应的部分。In the drawings, the same or corresponding reference numerals indicate the same or corresponding parts.
具体实施方式Detailed ways
下面将参照附图更详细地描述本公开的示例性实施例。虽然附图中显示了本公开的示例性实施例,然而应当理解,可以以各种形式实现本公开而不应被这里阐述的实施例所限制。相反,提供这些实施例是为了能够更透彻地理解本公开,并且能够将本公开的范围完整的传达给本领域的技术人员。Hereinafter, exemplary embodiments of the present disclosure will be described in more detail with reference to the accompanying drawings. Although the drawings show exemplary embodiments of the present disclosure, it should be understood that the present disclosure can be implemented in various forms and should not be limited by the embodiments set forth herein. On the contrary, these embodiments are provided to enable a more thorough understanding of the present disclosure and to fully convey the scope of the present disclosure to those skilled in the art.
在本发明中,应理解,诸如“包括”或“具有”等术语旨在指示本说明书中所公开的特征、数字、步骤、行为、部件、部分或其组合的存在,并且不旨在排除一个或多个其他特征、数字、步骤、行为、部件、部分或其组合存在的可能性。In the present invention, it should be understood that terms such as "including" or "having" are intended to indicate the existence of the features, numbers, steps, actions, components, parts, or combinations thereof disclosed in this specification, and are not intended to exclude one Or the possibility of the existence of multiple other features, numbers, steps, behaviors, components, parts or combinations thereof.
另外还需要说明的是,在不冲突的情况下,本发明中的实施例及实施例中的特征可以相互组合。下面将参考附图并结合实施例来详细说明本发明。In addition, it should be noted that the embodiments of the present invention and the features in the embodiments can be combined with each other if there is no conflict. Hereinafter, the present invention will be described in detail with reference to the drawings and in conjunction with the embodiments.
在对监控区域内活动目标进行跟踪时,可以通过依次对来自于各个摄像头的当前待测帧进行图像检测,然后基于对应于各个摄像头的检测结果在监控区域中进行全局追踪,进而基于较少的计算资源实现对多路监控视频中的目标对象实现全局的跟踪,降低对计算资源的需求。When tracking moving targets in the monitoring area, you can perform image detection on the current frame to be tested from each camera in turn, and then perform global tracking in the monitoring area based on the detection results corresponding to each camera, and then based on less Computing resources realize global tracking of target objects in multi-channel surveillance videos, reducing the demand for computing resources.
在介绍了本发明的基本原理之后,下面具体介绍本发明的各种非限制性实施方式。After introducing the basic principles of the present invention, various non-limiting embodiments of the present invention will be described in detail below.
图1示意性地示出了根据本发明实施方式的目标跟踪方法100的流程示意图,FIG. 1 schematically shows a schematic flowchart of a target tracking method 100 according to an embodiment of the present invention,
如图1所示,该方法100可以包括:As shown in FIG. 1, the method 100 may include:
步骤S101、获取设置于监控区域内的多个摄像头的当前待测帧;Step S101: Obtain current frames to be tested of multiple cameras set in the monitoring area;
具体地,监控区域是指多个摄像头的取景区域的总和,多个摄像头包括至少两个摄像头,并且上述多个摄像头的取景区域彼此相邻接或至少部分地重叠,从而待跟踪的目标对象能够在监控区域中移动进而出现在任意一个或多个摄像头的取景区域内。其中,从多个摄像头的监控视频中分别提取多个摄像头的当前待测帧,其中每个摄像头的当前待测帧具有相同的采集时间。可选地,本公 开中的待跟踪目标优选为行人,本领域技术人员可以理解,上述待跟踪目标也可以是其他可移动的物体,比如动物、车辆等,本公开对此不作具体限制。Specifically, the monitoring area refers to the sum of the viewing areas of multiple cameras. The multiple cameras include at least two cameras, and the viewing areas of the multiple cameras are adjacent to each other or at least partially overlapped, so that the target object to be tracked can be Move in the monitoring area and then appear in the viewing area of any one or more cameras. Among them, the current frames to be measured of the multiple cameras are respectively extracted from the surveillance videos of the multiple cameras, and the current frames to be measured of each camera have the same acquisition time. Optionally, the target to be tracked in this disclosure is preferably a pedestrian. Those skilled in the art can understand that the target to be tracked may also be other movable objects, such as animals, vehicles, etc., which is not specifically limited in the present disclosure.
例如,在复杂监控场景下,比如在楼道、大型商场、机房等场所,通常会使用大量的摄像头对各个区域进行监控,并得到多路监控视频。图2示出一种示意性监控场景,在该监控场景中设置有摄像头201和摄像头202,如图3示出上述摄像头201和摄像头202的取景画面。其中,摄像头201的监控视频可解析为图像帧序列(A 1,A 2,...,A N),摄像头202的监控视频可解析为图像帧序列(B 1,B 2,...,B N),其中上述解析可以实时在线进行或离线进行。基于此,可以按时序从上述多个图像帧序列中依次提取两个摄像头的当前待测帧A n和B n以进行本公开所示出的目标跟踪,其中,下标n的取值可以是n=1,2,…,N。 For example, in complex surveillance scenarios, such as corridors, large shopping malls, computer rooms, etc., a large number of cameras are usually used to monitor various areas and obtain multiple surveillance videos. FIG. 2 shows a schematic monitoring scene in which a camera 201 and a camera 202 are set, and FIG. 3 shows a viewfinder screen of the above-mentioned camera 201 and the camera 202. Among them, the surveillance video of the camera 201 can be parsed as a sequence of image frames (A 1 , A 2 ,..., A N ), and the surveillance video of the camera 202 can be parsed as a sequence of image frames (B 1 , B 2 ,..., B N ), where the above analysis can be performed online or offline in real time. Based on this, the current frames A n and B n of the two cameras to be tested can be sequentially extracted from the above-mentioned multiple image frame sequences in time sequence to perform the target tracking shown in the present disclosure, where the value of the subscript n may be n=1,2,...,N.
在一些可能的实施例中,该方法100还可以包括:确定多个待测帧序号,根据多个待测帧序号按时序地迭代获取多个摄像头的当前待测帧,从而迭代地执行目标跟踪;其中,根据多个待测帧序号中初始待测帧序号对应得到初始的全局目标轨迹;根据多个待测帧序号中后续待测帧序号对应得到迭代更新后的全局目标轨迹。这样可以减少运算量,提高全局跟踪实时性,In some possible embodiments, the method 100 may further include: determining a plurality of frame numbers to be measured, and iteratively acquiring the current frame to be measured of multiple cameras according to the plurality of frame numbers to be measured in time series, so as to iteratively perform target tracking ; Among them, the initial global target trajectory is obtained corresponding to the initial frame number to be tested among the multiple frame numbers to be tested; the global target trajectory after iterative update is obtained corresponding to the subsequent frame number to be tested in the multiple frame numbers to be tested. This can reduce the amount of calculations and improve the real-time performance of global tracking,
具体地,可以根据预设取帧策略确定多个待测帧序号。例如,针对每秒24帧的监控视频,可以每跨5帧从摄像头201和摄像头202的监控视频中获取一次当前待测帧A n和B n,其中下标n的取值可以是n=1,6,11,…,并依次类推。然而,也可以采取其他间隔帧数,或者,也可以采取逐帧检测的方式,本公开对此不作具体限定。基于此,可以基于初始待测帧序号(n=1)对应的当前待测帧A 1和B 1到初始的全局目标轨迹,进一步可以根据后续待测帧序号(n=6,11,...等)对应的当前待测帧A n和B n进行迭代的目标跟踪,从而得到迭代更新后的全局目标轨迹。 Specifically, multiple frame numbers to be tested can be determined according to a preset frame fetching strategy. For example, for a surveillance video of 24 frames per second, the current frames A n and B n to be tested can be obtained from the surveillance video of the camera 201 and the camera 202 every 5 frames, where the value of the subscript n can be n=1 , 6, 11,..., and so on. However, other interval frame numbers may also be adopted, or a frame-by-frame detection method may also be adopted, which is not specifically limited in the present disclosure. Based on this, the current frame to be tested A 1 and B 1 corresponding to the initial frame number to be tested (n = 1) to the initial global target trajectory can be further based on the subsequent frame number to be tested (n = 6, 11,... Etc.) Iterative target tracking is performed on the corresponding current frames A n and B n to be tested, so as to obtain the global target trajectory after iterative update.
如图1所示,该方法100还可以包括:As shown in FIG. 1, the method 100 may further include:
步骤S102、依次对多个摄像头中每个摄像头的当前待测帧进行目标检测,得到每个摄像头对应的检测框集合;Step S102: Perform target detection on the current frame to be tested of each of the multiple cameras in turn, to obtain a set of detection frames corresponding to each camera;
在一个可能的实施方式中,对每个摄像头的当前待测帧进行目标检测,包括:将每个摄像头的当前待测帧输入目标检测模型进行目标检测;其中,目标检测模型是基于神经网络训练得到的行人检测模型。In a possible implementation manner, performing target detection on the current frame under test of each camera includes: inputting the current frame under test of each camera into a target detection model for target detection; wherein the target detection model is based on neural network training The resulting pedestrian detection model.
例如,如图4所示,示出了摄像头201和摄像头202的当前待测帧A n和B n,然后,在任意基于深度学习的行人检测模型中输入预处理后的当前待测帧A n和B n进行检测,输出针对每个摄像头的一系列行人检测框。获取行人检测框的目的在于获取当前待测帧A n和B n中所有行人的位置信息和尺寸信息。行人检测模型比如可以是YOLO(统一实时目标检测,You Only Look Once)模型等,本公开对此不作具体限制。如图5所示,示出了对多个当前待测帧A n和B n进行检测得到的多个检测框集合,其中摄像头201对应的检测框集合(a 1,a 2,a 3),摄像头202对应的检测框集合(b)。 For example, as shown in FIG. 4, the current frames A n and B n of the camera 201 and the camera 202 to be tested are shown, and then the pre-processed current frame A n is input into any deep learning-based pedestrian detection model. Perform detection with B n and output a series of pedestrian detection frames for each camera. The purpose of obtaining the pedestrian detection frame is to obtain the position information and size information of all pedestrians in the current frames A n and B n to be tested. The pedestrian detection model may be, for example, a YOLO (Unified Real-Time Object Detection, You Only Look Once) model, etc., which is not specifically limited in the present disclosure. As shown in FIG. 5, multiple detection frame sets obtained by detecting multiple current frames A n and B n to be tested are shown, where the detection frame sets (a 1 , a 2 , a 3 ) corresponding to the camera 201 are shown, The detection frame set corresponding to the camera 202 (b).
在一个可能的实施方式中,在得到每个摄像头对应的检测框集合之后,还包括:根据每个摄像头的取景位置、以及每个摄像头对应的检测框集合中的每个检测框的框底中心点进行投影变换,从而确定每个摄像头对应的检测框集合中每个检测框的地面坐标。这样,可以将每个摄像头取景范围内识别的目标组合到统一的坐标系中。In a possible implementation manner, after the detection frame set corresponding to each camera is obtained, it further includes: according to the viewing position of each camera and the center of each detection frame in the detection frame set corresponding to each camera The points undergo projection transformation to determine the ground coordinates of each detection frame in the detection frame set corresponding to each camera. In this way, the targets identified in the viewing range of each camera can be combined into a unified coordinate system.
例如,可以获取图5中每个摄像头对应的每个检测框的框底中心点位置,对该每个检测框的框底中心点位置进行转换,得到目标对象在监控场景中的实际地面位置,图6示出了通过投影转换获得的每个检测框的地面坐标。具体而言,可以看出,每个摄像头视角下的地面过道是一个近似梯形区域,因此针对每个摄像头对应的检测框集合,首先可以通过梯形-矩形转换得到每个检测框的框底中心点在标准矩形区域中的坐标,其次根据监控场景的实际布局对标准矩形区域进行旋转,通过旋转矩阵计算得到每个检测框的框底中心点的旋转后坐标,最后根据监控场景的实际布局对旋转后坐标进行平移和缩放,得到最终的坐标位置。For example, the position of the center point of the bottom of each detection frame corresponding to each camera in Figure 5 can be obtained, and the position of the center point of the bottom of the frame of each detection frame can be converted to obtain the actual ground position of the target object in the monitoring scene. Fig. 6 shows the ground coordinates of each detection frame obtained through projection transformation. Specifically, it can be seen that the ground aisle under the viewing angle of each camera is an approximate trapezoidal area, so for the detection frame set corresponding to each camera, the center point of the bottom of each detection frame can be obtained through the trapezoid-rectangular transformation. The coordinates in the standard rectangular area. Secondly, the standard rectangular area is rotated according to the actual layout of the monitoring scene. The rotated coordinates of the center point of the bottom of each detection frame are calculated by the rotation matrix, and finally the rotation is rotated according to the actual layout of the monitoring scene. After the coordinates are translated and zoomed, the final coordinate position is obtained.
在一个可能的实施方式中,多个摄像头的取景区域至少部分地重叠,方法还包括:根据每个摄像头的取景区域在地面坐标系中划分每个摄像头的工作区 域;其中,每个摄像头的工作区域互不重叠,若多个摄像头中的第一摄像头对应的任意一个检测框的地面坐标超出对应的工作区域,则在第一摄像头的检测框集合中去除任意一个检测框。In a possible implementation manner, the viewing areas of the multiple cameras overlap at least partially, and the method further includes: dividing the working area of each camera in the ground coordinate system according to the viewing area of each camera; wherein, the work of each camera The areas do not overlap each other. If the ground coordinates of any detection frame corresponding to the first camera of the multiple cameras exceeds the corresponding working area, any detection frame is removed from the detection frame set of the first camera.
例如,如图2所示,为了使得监控场景中不存在监控盲区,摄像头201和摄像头202的取景区域实际上存在重叠。基于此,为了有效避免坐标显示冲突的问题,可以对每个摄像头进行工作区域的划分,比如,摄像头201的工作区域为X区域,摄像头202的工作区域为Y区域,使得每个摄像头的工作区域相邻接。进一步地,每个摄像头对应的每个检测框的地面坐标需位于该摄像头的工作区域内,若不在该摄像头负责的工作区域内则除去。比如,由于摄像头201对应的检测框集合(a 1,a 2,a 3)中的检测框a 3的地面坐标在X区域之外,因此,在摄像头201对应的检测框集合中去除检测框a 3,得到(a 1,a 2)进行后续的操作。 For example, as shown in FIG. 2, in order to prevent a surveillance blind spot in the surveillance scene, the viewing areas of the camera 201 and the camera 202 actually overlap. Based on this, in order to effectively avoid the problem of coordinate display conflicts, the working area of each camera can be divided. For example, the working area of the camera 201 is the X area, and the working area of the camera 202 is the Y area, so that the working area of each camera is Adjacent. Further, the ground coordinates of each detection frame corresponding to each camera need to be located in the working area of the camera, and removed if it is not in the working area of the camera. For example, because the ground coordinates of the detection frame a 3 in the detection frame set (a 1 , a 2 , a 3 ) corresponding to the camera 201 are outside the X area, the detection frame a is removed from the detection frame set corresponding to the camera 201 3 , get (a 1 , a 2 ) for subsequent operations.
在一个可能的实施方式中,方法还包括:将每个摄像头的工作区域中的非关键区域截去。具体地,可以基于监控场景的具体布局确定是否为关键区域,比如,对于行人无法通过的天花板区域,就可以直接截去,这样能够减少目标跟踪的运算量。In a possible implementation, the method further includes: cutting off non-critical areas in the working area of each camera. Specifically, whether it is a critical area can be determined based on the specific layout of the monitoring scene. For example, the ceiling area that cannot be passed by pedestrians can be directly cut off, which can reduce the amount of calculation for target tracking.
如图1所示,该方法100还可以包括:As shown in FIG. 1, the method 100 may further include:
步骤S103、根据每个摄像头对应的检测框集合进行目标跟踪,根据跟踪结果更新全局目标轨迹。Step S103: Perform target tracking according to the detection frame set corresponding to each camera, and update the global target trajectory according to the tracking result.
具体地,如上文所述,针对每个摄像头,可以根据初始的当前待测帧A 1和B 1进行目标检测,确定初始的全局目标轨迹。进一步地,可以根据后续获取的当前待测帧A n和B n进行目标检测,并根据目标检测结果迭代地进行目标跟踪,从而对全局目标轨迹进行迭代更新。 Specifically, as described above, for each camera, target detection can be performed according to the initial current frame to be measured A 1 and B 1 to determine the initial global target trajectory. Further, target detection may be performed according to the current frames A n and B n to be tested subsequently obtained, and target tracking may be performed iteratively according to the target detection result, so as to iteratively update the global target trajectory.
在一个可能的实施方式中,根据每个摄像头对应的检测框集合进行跟踪,包括:采用多目标跟踪算法,并基于每个摄像头对应的检测框集合进行多目标跟踪,确定每个摄像头对应的局部跟踪信息;其中,多目标跟踪采用的参数基于每个摄像头的历史待测帧而确定。这样能够实现监控区域中的多目标跟踪。In a possible implementation, tracking according to the detection frame set corresponding to each camera includes: adopting a multi-target tracking algorithm, and performing multi-target tracking based on the detection frame set corresponding to each camera, and determining the part corresponding to each camera Tracking information: Among them, the parameters used in multi-target tracking are determined based on the historical frame to be measured of each camera. This enables multi-target tracking in the monitoring area.
具体地,多目标跟踪算法是基于单摄像头的目标跟踪算法,例如DeepSORT算法(基于深度特征关联的简单在线实时跟踪算法,Simple Online and Realtime Tracking with a Deep Association Metric),因此可以得到每个摄像头的局部跟踪信息。其中,多目标跟踪采用的参数基于每个摄像头的历史待测帧而确定,具体而言,可以在任意一个目标初次出现在某个摄像头的工作区域时确定待跟踪的目标框,并基于多目标检测算法和已经标注身份的目标框对该摄像头的后续待测帧进行跟踪,确定该目标在该摄像头工作区域中的局部跟踪信息。Specifically, the multi-target tracking algorithm is a target tracking algorithm based on a single camera, such as the DeepSORT algorithm (Simple Online and Realtime Tracking with a Deep Association Metric), so you can get the information of each camera. Local tracking information. Among them, the parameters used in multi-target tracking are determined based on the historical frame to be measured of each camera. Specifically, the target frame to be tracked can be determined when any target appears in the working area of a certain camera for the first time, and based on the multi-target The detection algorithm and the identified target frame track the subsequent frame to be measured of the camera, and determine the local tracking information of the target in the working area of the camera.
在一个可能的实施方式中,多目标跟踪算法为deepsort算法。当然,也可以采用其他的目标跟踪算法,本领域的技术人员可以理解,本公开所要强调的不是具体采用何种目标跟踪算法。In a possible implementation, the multi-target tracking algorithm is a deepsort algorithm. Of course, other target tracking algorithms can also be used, and those skilled in the art can understand that what the present disclosure intends to emphasize is not which target tracking algorithm is specifically used.
在一个可能的实施方式中,根据跟踪结果更新全局目标轨迹,还包括:根据每个摄像头对应的局部跟踪信息为每个检测框添加身份标识;基于身份标识,利用每个检测框的地面坐标对全局目标轨迹进行更新。In a possible implementation, updating the global target trajectory according to the tracking result also includes: adding an identity to each detection frame according to the local tracking information corresponding to each camera; based on the identity, using the ground coordinate pair of each detection frame The global target trajectory is updated.
例如,如图6所示,其中的曲线部分示出了当前已有的全局目标轨迹,也即是在上一次迭代过程中确定的全局目标轨迹,且其中的点a 1、点a 2和点b分别表示图5中所示出多个检测框的地面坐标。其中,若摄像头201对应的局部跟踪信息指示检测框a 2和已有的“目标2”特征匹配,则为检测框a 2标注“目标2”并将点a 2的地面坐标加入“目标2”的现有轨迹中(即图6中的“目标2”虚曲线),若摄像头201对应的局部跟踪信息指示检测框点a 1并不存在匹配目标,则为检测框a 1新增一个标注“目标3”,并新创建“目标3”的轨迹。 For example, as shown in Figure 6, the curve part shows the current global target trajectory, that is, the global target trajectory determined in the last iteration, and the points a 1 , a 2 and points b respectively represents the ground coordinates of the multiple detection frames shown in FIG. 5. Among them, if the local tracking information corresponding to the camera 201 indicates that the detection frame a 2 matches the existing "target 2" feature, then the detection frame a 2 is labeled "target 2" and the ground coordinates of the point a 2 are added to "target 2" existing track (i.e., the "target 2" dashed curve in FIG. 6), if the camera 201 corresponding local trace information indicative of a 1-point detection frame matches the target does not exist, add a 1 compared with a detection frame labeled " Goal 3" and create a new trajectory of "Goal 3".
在一个可能的实施方式中,根据跟踪结果更新全局目标轨迹,还包括:多个摄像头的工作区域确定多个摄像头之间的关联关系;根据每个摄像头的局部跟踪信息确定对应工作区域中的新增检测框和消失检测框;根据多个摄像头之间的关联关系对处于不同工作区域中的新增检测框和消失检测框进行关联,得到关联信息;根据关联信息更新全局目标轨迹。In a possible implementation manner, updating the global target trajectory according to the tracking result further includes: determining the association relationship between the multiple cameras by the working areas of the multiple cameras; determining the new information in the corresponding working area according to the local tracking information of each camera Add detection frames and disappearance detection frames; associate new detection frames and disappearance detection frames in different work areas according to the association relationship between multiple cameras to obtain associated information; update the global target trajectory according to the associated information.
具体地,其中多个摄像头之间的关联关系比如是区域X和区域Y在指定位置相邻接,从而在目标移动时能够基于上述关联关系从邻接位置处跨越不同的工作区域。其中,关联信息是指某一工作区域中的新增检测框和另一工作区域中的消失检测框实现关联,也即对应为同一身份标识。换句话说,针对具有邻接边界的两个工作区域,可以在其中一个工作区域的邻接边界处先获取多个跟踪目标的消失次序,在另一工作区域中按照上述消失次序对出现于该邻接边界处的多个新增目标进行对应的标识分配并持续跟踪,Specifically, the association relationship between the multiple cameras is, for example, that the area X and the area Y are adjacent to each other at a specified position, so that when the target moves, different work areas can be straddled from adjacent positions based on the above association relationship. Among them, the association information refers to the association between a new detection frame in a certain work area and a disappearance detection frame in another work area, that is, corresponding to the same identity. In other words, for two working areas with adjoining boundaries, the disappearance order of multiple tracking targets can be obtained at the adjoining boundary of one of the work areas, and the disappearance order of the pair appears in the adjoining boundary in the other work area. The multiple newly-added targets at the location are assigned corresponding identifiers and continue to be tracked,
例如,如图6所示,其中区域Y中的点b表示图5中所示出检测框b的地面坐标。若摄像头201对应的局部跟踪信息指示检测框点b并不存在匹配目标,也即在区域Y中存在新增目标;并且摄像头201对应的局部跟踪信息指示所持续跟踪的“目标1”在当前检测帧消失,也即在区域X中存在消失目标,则可以为检测框b标注“目标1”并将点b的地面坐标加入“目标1”的现有轨迹中(即图6中的“目标1”虚曲线),实现跨摄像头、跨工作区域的目标跟踪。For example, as shown in FIG. 6, the point b in the area Y represents the ground coordinates of the detection frame b shown in FIG. 5. If the local tracking information corresponding to the camera 201 indicates that there is no matching target at the detection frame point b, that is, there is a new target in the area Y; and the local tracking information corresponding to the camera 201 indicates that the continuously tracked "target 1" is currently detected If the frame disappears, that is, there is a disappearing target in area X, you can mark the detection frame b with "Target 1" and add the ground coordinates of point b to the existing trajectory of "Target 1" (ie, "Target 1" in Figure 6). "Dashed curve), to achieve cross-camera, cross-working area target tracking.
这样,根据本发明实施方式的基于多摄像头的目标跟踪方法,通过依次对来各个摄像头的当前待测帧进行图像检测,然后基于对应于各个摄像头的检测结果在监控区域中进行全局追踪,可以基于较少的计算资源实现对多路监控视频中的目标对象实现全局的跟踪,降低对计算资源的需求。例如,无需为各个摄像头单独提供用于跟踪各个局部区域中的目标对象的GPU计算资源,而可以提供较少的计算资源以用于在监控区域中进行目标对象的全局跟踪。In this way, according to the multi-camera-based target tracking method of the present invention, by sequentially performing image detection on the current frame to be tested from each camera, and then performing global tracking in the monitoring area based on the detection result corresponding to each camera, it can be based on Fewer computing resources enable global tracking of target objects in multi-channel surveillance videos, reducing the demand for computing resources. For example, there is no need to separately provide GPU computing resources for tracking the target object in each local area for each camera, but fewer computing resources can be provided for global tracking of the target object in the monitoring area.
基于相同的技术构思,本发明实施例还提供一种目标跟踪装置,用于执行上述任一实施例所提供的目标跟踪方法。图7为本发明实施例提供的一种目标跟踪装置结构示意图。Based on the same technical concept, an embodiment of the present invention also provides a target tracking device for executing the target tracking method provided in any of the foregoing embodiments. Fig. 7 is a schematic structural diagram of a target tracking device provided by an embodiment of the present invention.
如图7所示,装置700包括:As shown in FIG. 7, the apparatus 700 includes:
获取单元701,用于获取设置于监控区域内的多个摄像头的当前待测帧;The acquiring unit 701 is configured to acquire current frames to be measured of multiple cameras arranged in the monitoring area;
检测单元702,用于依次对多个摄像头中每个摄像头的当前待测帧进行目标检测,得到每个摄像头对应的检测框集合;The detection unit 702 is configured to sequentially perform target detection on the current frame to be tested of each of the multiple cameras to obtain a set of detection frames corresponding to each camera;
跟踪单元703,用于根据每个摄像头对应的检测框集合进行目标跟踪,根据跟踪结果确定全局目标轨迹。The tracking unit 703 is configured to perform target tracking according to the detection frame set corresponding to each camera, and determine the global target trajectory according to the tracking result.
在一些可能的实施方式中,装置700还包括:选帧单元,用于确定多个待测帧序号,根据多个待测帧序号按时序地迭代获取多个摄像头的当前待测帧,从而迭代地执行目标跟踪;其中,根据多个待测帧序号中初始待测帧序号对应得到初始的全局目标轨迹;根据多个待测帧序号中后续待测帧序号对应得到迭代更新后的全局目标轨迹。In some possible implementation manners, the apparatus 700 further includes: a frame selection unit, configured to determine multiple frame numbers to be tested, and iteratively obtain current frames to be tested from multiple cameras in a time sequence according to the multiple frame numbers to be tested, so as to iterate To perform target tracking; among them, the initial global target trajectory is obtained according to the initial test frame sequence number among the multiple test frame numbers; the global target trajectory after iterative update is obtained according to the subsequent test frame sequence numbers among the multiple test frame sequence numbers. .
在一些可能的实施方式中,检测单元702,还用于:将每个摄像头的当前待测帧输入目标检测模型进行目标检测;其中,目标检测模型是基于神经网络训练得到的行人检测模型。In some possible implementation manners, the detection unit 702 is further configured to: input the current frame to be measured of each camera into a target detection model for target detection; wherein the target detection model is a pedestrian detection model obtained based on neural network training.
在一些可能的实施方式中,检测单元702,还用于:在得到每个摄像头对应的检测框集合之后,根据每个摄像头的取景位置对每个摄像头对应的检测框集合中的每个检测框的框底中心点进行投影变换,从而确定每个检测框的地面坐标。In some possible implementation manners, the detection unit 702 is further configured to: after obtaining the detection frame set corresponding to each camera, determine each detection frame in the detection frame set corresponding to each camera according to the viewing position of each camera. Projection transformation is performed on the center point of the bottom of the frame to determine the ground coordinates of each detection frame.
在一些可能的实施方式中,多个摄像头的取景区域至少部分地重叠,装置700还用于:根据每个摄像头的取景区域在地面坐标系中划分每个摄像头的工作区域;其中,每个摄像头的工作区域互不重叠,若多个摄像头中的第一摄像头对应的任意一个检测框的地面坐标超出对应的工作区域,则在第一摄像头的检测框集合中去除任意一个检测框。In some possible implementation manners, the viewing areas of multiple cameras overlap at least partially, and the device 700 is further configured to: divide the working area of each camera in the ground coordinate system according to the viewing area of each camera; wherein, each camera If the ground coordinates of any detection frame corresponding to the first camera of the multiple cameras are beyond the corresponding working area, any detection frame is removed from the detection frame set of the first camera.
在一些可能的实施方式中,检测单元702,还用于:将每个摄像头的工作区域中的非关键区域截去。In some possible implementation manners, the detection unit 702 is also used to cut off non-critical areas in the working area of each camera.
在一些可能的实施方式中,跟踪单元703,还用于:采用多目标跟踪算法,并基于每个摄像头对应的检测框集合进行多目标跟踪,确定每个摄像头对应的局部跟踪信息;其中,多目标跟踪采用的参数基于每个摄像头的历史待测帧而确定。In some possible implementation manners, the tracking unit 703 is further configured to: adopt a multi-target tracking algorithm and perform multi-target tracking based on the detection frame set corresponding to each camera, and determine the local tracking information corresponding to each camera; The parameters used for target tracking are determined based on the historical frames to be tested for each camera.
在一些可能的实施方式中,多目标跟踪算法为deepsort算法。In some possible implementation manners, the multi-target tracking algorithm is a deepsort algorithm.
在一些可能的实施方式中,跟踪单元703,还用于:根据每个摄像头对应的局部跟踪信息为每个检测框添加身份标识;基于每个检测框的身份标识和地面坐标确定迭代更新后的全局目标轨迹。In some possible implementation manners, the tracking unit 703 is further configured to: add an identity to each detection frame according to the local tracking information corresponding to each camera; to determine the iteratively updated information based on the identity and ground coordinates of each detection frame Global target trajectory.
在一些可能的实施方式中,跟踪单元703,还用于:根据多个摄像头的工作区域确定多个摄像头之间的关联关系;根据每个摄像头的局部跟踪信息确定对应工作区域中的新增检测框和消失检测框;根据多个摄像头之间的关联关系对处于不同工作区域中的新增检测框和消失检测框进行关联,得到关联信息;根据关联信息确定迭代更新后的全局目标轨迹。In some possible implementation manners, the tracking unit 703 is further configured to: determine the association relationship between the multiple cameras according to the work areas of the multiple cameras; determine the new detection in the corresponding work area according to the local tracking information of each camera Frames and disappearance detection frames; associate new detection frames and disappearance detection frames in different work areas according to the association relationship between multiple cameras to obtain association information; determine the iteratively updated global target trajectory according to the association information.
这样,根据本发明实施方式的基于多摄像头的目标跟踪装置,通过依次对来各个摄像头的当前待测帧进行图像检测,然后基于对应于各个摄像头的检测结果在监控区域中进行全局追踪,可以基于较少的计算资源实现对多路监控视频中的目标对象实现全局的跟踪,降低对计算资源的需求。例如,无需为各个摄像头单独提供用于跟踪各个局部区域中的目标对象的GPU计算资源,而可以提供较少的计算资源以用于在监控区域中进行目标对象的全局跟踪。In this way, according to the multi-camera-based target tracking device of the present invention, the current frame to be measured from each camera is detected in sequence, and then global tracking is performed in the monitoring area based on the detection result corresponding to each camera, which can be based on Fewer computing resources enable global tracking of target objects in multi-channel surveillance videos, reducing the demand for computing resources. For example, there is no need to separately provide GPU computing resources for tracking the target object in each local area for each camera, but fewer computing resources can be provided for global tracking of the target object in the monitoring area.
需要说明的是,本申请实施例中的装置可以实现前述方法的实施例的各个过程,并达到相同的效果和功能,这里不再赘述。It should be noted that the device in the embodiment of the present application can implement each process of the foregoing method embodiment, and achieve the same effect and function, which will not be repeated here.
基于相同的技术构思,本发明实施例还提供一种目标跟踪系统,具体包括:设置于监控区域内的多个摄像头,以及与多个摄像头分别通信连接的目标跟踪装置;其中,目标跟踪装置被配置用于执行上述任一实施例所提供的目标跟踪方法。Based on the same technical concept, an embodiment of the present invention also provides a target tracking system, which specifically includes: a plurality of cameras arranged in a monitoring area, and a target tracking device respectively communicatively connected with the plurality of cameras; wherein the target tracking device is It is configured to execute the target tracking method provided in any of the above embodiments.
基于相同的技术构思,所属技术领域的技术人员能够理解,本发明的各个方面可以实现为设备、方法或计算机可读存储介质。因此,本发明的各个方面可以具体实现为以下形式,即:完全的硬件实施方式、完全的软件实施方式(包括固件、微代码等),或硬件和软件方面结合的实施方式,这里可以统称为“电路”、“模块”或“设备”。Based on the same technical concept, those skilled in the art can understand that various aspects of the present invention can be implemented as devices, methods, or computer-readable storage media. Therefore, various aspects of the present invention can be specifically implemented in the following forms, namely: complete hardware implementation, complete software implementation (including firmware, microcode, etc.), or a combination of hardware and software implementations, which can be collectively referred to herein as "Circuit", "Module" or "Equipment".
在一些可能的实施方式中,本发明的一种目标跟踪装置可以至少包括一个或多个处理器、以及至少一个存储器。其中,所述存储器存储有程序,当 所述程序被所述处理器执行时,使得所述处理器执行如图1所示的步骤:获取设置于监控区域内的多个摄像头的当前待测帧;依次对多个摄像头中每个摄像头的当前待测帧进行目标检测,得到每个摄像头对应的检测框集合;根据每个摄像头对应的检测框集合进行目标跟踪,根据跟踪结果确定全局目标轨迹。In some possible implementation manners, a target tracking device of the present invention may at least include one or more processors and at least one memory. Wherein, the memory stores a program, and when the program is executed by the processor, the processor is caused to perform the steps shown in FIG. 1: acquiring the current frame to be measured of multiple cameras arranged in the monitoring area ;Sequentially perform target detection on the current frame under test of each camera in the multiple cameras to obtain the detection frame set corresponding to each camera; perform target tracking according to the detection frame set corresponding to each camera, and determine the global target trajectory according to the tracking result.
下面参照图8来描述根据本发明的这种实施方式的目标跟踪装置8。图8显示的装置8仅仅是一个示例,不应对本发明实施例的功能和使用范围带来任何限制。The target tracking device 8 according to this embodiment of the present invention will be described below with reference to FIG. 8. The device 8 shown in FIG. 8 is only an example, and should not bring any limitation to the function and application scope of the embodiment of the present invention.
如图8所示,装置8可以以通用计算设备的形式表现,包括但不限于:至少一个处理器10、至少一个存储器20、连接不同设备组件的总线60。As shown in FIG. 8, the apparatus 8 may be in the form of a general-purpose computing device, including but not limited to: at least one processor 10, at least one memory 20, and a bus 60 connecting different device components.
总线60包括数据总线、地址总线和控制总线。The bus 60 includes a data bus, an address bus, and a control bus.
存储器20可以包括易失性存储器,例如随机存取存储器(RAM)21和/或高速缓存存储器22,还可以进一步包括只读存储器(ROM)23。The memory 20 may include a volatile memory, such as a random access memory (RAM) 21 and/or a cache memory 22, and may further include a read-only memory (ROM) 23.
存储器20还可以包括程序模块24,这样的程序模块24包括但不限于:操作设备、一个或者多个应用程序、其它程序模块以及程序数据,这些示例中的每一个或某种组合中可能包括网络环境的实现。The memory 20 may also include a program module 24. Such program module 24 includes, but is not limited to, an operating device, one or more application programs, other program modules, and program data. Each of these examples or a certain combination may include a network. The realization of the environment.
装置5还可以与一个或多个外部设备2(例如键盘、指向设备、蓝牙设备等)通信,也可与一个或者多个其他设备进行通信。这种通信可以通过输入/输出(I/O)接口40进行,并在显示单元30上进行显示。并且,装置5还可以通过网络适配器50与一个或者多个网络(例如局域网(LAN),广域网(WAN)和/或公共网络,例如因特网)通信。如图所示,网络适配器50通过总线60与装置5中的其它模块通信。应当明白,尽管图中未示出,但可以结合装置5使用其它硬件和/或软件模块,包括但不限于:微代码、设备驱动器、冗余处理单元、外部磁盘驱动阵列、RAID设备、磁带驱动器以及数据备份存储设备等。The apparatus 5 can also communicate with one or more external devices 2 (for example, a keyboard, a pointing device, a Bluetooth device, etc.), and can also communicate with one or more other devices. Such communication can be performed through an input/output (I/O) interface 40 and displayed on the display unit 30. In addition, the device 5 may also communicate with one or more networks (for example, a local area network (LAN), a wide area network (WAN), and/or a public network, such as the Internet) through the network adapter 50. As shown in the figure, the network adapter 50 communicates with other modules in the device 5 through the bus 60. It should be understood that although not shown in the figure, other hardware and/or software modules can be used in conjunction with the device 5, including but not limited to: microcode, device drivers, redundant processing units, external disk drive arrays, RAID devices, tape drives And data backup storage devices, etc.
图9示出了一种计算机可读存储介质,用于执行如上所述的方法。Fig. 9 shows a computer-readable storage medium for executing the method as described above.
在一些可能的实施方式中,本发明的各个方面还可以实现为一种计算机可读存储介质的形式,其包括程序代码,当所述程序代码在被处理器执行时,所述程序代码用于使所述处理器执行上面描述的方法。In some possible implementation manners, various aspects of the present invention can also be implemented in the form of a computer-readable storage medium, which includes program code. When the program code is executed by a processor, the program code is used for The processor is caused to execute the method described above.
上面描述的方法包括了上面的附图中示出和未示出的多个操作和步骤,这里将不再赘述。The above-described method includes multiple operations and steps shown and not shown in the above drawings, which will not be repeated here.
所述计算机可读存储介质可以采用一个或多个可读介质的任意组合。可读介质可以是可读信号介质或者可读存储介质。可读存储介质例如可以是——但不限于——电、磁、光、电磁、红外线、或半导体的设备、设备或器件,或者任意以上的组合。可读存储介质的更具体的例子(非穷举的列表)包括:具有一个或多个导线的电连接、便携式盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、光纤、便携式紧凑盘只读存储器(CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。The computer-readable storage medium may adopt any combination of one or more readable media. The readable medium may be a readable signal medium or a readable storage medium. The readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor device, device, or device, or a combination of any of the above. More specific examples (non-exhaustive list) of readable storage media include: electrical connections with one or more wires, portable disks, hard disks, random access memory (RAM), read-only memory (ROM), erasable Type programmable read only memory (EPROM or flash memory), optical fiber, portable compact disk read only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
如图9所示,描述了根据本发明的实施方式的计算机可读存储介质90,其可以采用便携式紧凑盘只读存储器(CD-ROM)并包括程序代码,并可以在终端设备,例如个人电脑上运行。然而,本发明的计算机可读存储介质不限于此,在本文件中,可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行设备、设备或者器件使用或者与其结合使用。As shown in FIG. 9, a computer-readable storage medium 90 according to an embodiment of the present invention is described. It can adopt a portable compact disk read-only memory (CD-ROM) and include program codes, and can be stored in a terminal device, such as a personal computer. Run on. However, the computer-readable storage medium of the present invention is not limited to this. In this document, the readable storage medium can be any tangible medium that contains or stores a program. The program can be used by or in combination with an instruction execution device, device, or device. .
可以以一种或多种程序设计语言的任意组合来编写用于执行本发明操作的程序代码,所述程序设计语言包括面向对象的程序设计语言—诸如Java、Python、C++等,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算设备上执行、部分地在用户设备上执行部分在远程计算设备上执行、或者完全在远程计算设备或服务器上执行。在涉及远程计算设备的情形中,远程计算设备可以通过任意种类的网络——包括局域网(LAN)或广域网(WAN)—连接到用户计算设备,或者,可以连接到外部计算设备(例如利用因特网服务提供商来通过因特网连接)。The program code used to perform the operations of the present invention can be written in any combination of one or more programming languages. The programming languages include object-oriented programming languages—such as Java, Python, C++, etc., as well as conventional Procedural programming language-such as "C" language or similar programming language. The program code can be executed entirely on the user's computing device, partly executed on the user's device and partly executed on the remote computing device, or entirely executed on the remote computing device or server. In the case of remote computing devices, the remote computing device can be connected to the user computing device through any kind of network-including a local area network (LAN) or a wide area network (WAN), or it can be connected to an external computing device (for example, using Internet services). Provider to connect via the Internet).
此外,尽管在附图中以特定顺序描述了本发明方法的操作,但是,这并非要求或者暗示必须按照该特定顺序来执行这些操作,或是必须执行全部所示的操作才能实现期望的结果。附加地或备选地,可以省略某些步骤,将多个步骤合并为一个步骤执行,和/或将一个步骤分解为多个步骤执行。In addition, although the operations of the method of the present invention are described in a specific order in the drawings, this does not require or imply that these operations must be performed in the specific order, or that all the operations shown must be performed to achieve the desired result. Additionally or alternatively, some steps may be omitted, multiple steps may be combined into one step for execution, and/or one step may be decomposed into multiple steps for execution.
虽然已经参考若干具体实施方式描述了本发明的精神和原理,但是应该理解,本发明并不限于所公开的具体实施方式,对各方面的划分也不意味着这些方面中的特征不能组合以进行受益,这种划分仅是为了表述的方便。本发明旨在涵盖所附权利要求的精神和范围内所包括的各种修改和等同布置。Although the spirit and principle of the present invention have been described with reference to several specific embodiments, it should be understood that the present invention is not limited to the disclosed specific embodiments, and the division of various aspects does not mean that the features in these aspects cannot be combined for performance. Benefit, this division is only for the convenience of presentation. The present invention is intended to cover various modifications and equivalent arrangements included within the spirit and scope of the appended claims.

Claims (23)

  1. 一种目标跟踪方法,其特征在于,包括:A target tracking method is characterized in that it comprises:
    获取设置于监控区域内的多个摄像头的当前待测帧;Obtain the current frame to be tested of multiple cameras set in the monitoring area;
    依次对所述多个摄像头中每个摄像头的当前待测帧进行目标检测,得到每个摄像头对应的检测框集合;Performing target detection on the current frame to be tested of each camera in the plurality of cameras in turn, to obtain a set of detection frames corresponding to each camera;
    根据所述每个摄像头对应的检测框集合进行目标跟踪,根据跟踪结果确定全局目标轨迹。Target tracking is performed according to the detection frame set corresponding to each camera, and the global target trajectory is determined according to the tracking result.
  2. 根据权利要求1所述的方法,其特征在于,还包括:The method according to claim 1, further comprising:
    确定多个待测帧序号,根据所述多个待测帧序号按时序地迭代获取所述多个摄像头的当前待测帧,从而迭代地执行所述目标跟踪;Determine a plurality of frame numbers to be tested, and iteratively acquire current frames to be tested of the plurality of cameras in time series according to the frame numbers to be tested, so as to iteratively perform the target tracking;
    其中,根据所述多个待测帧序号中初始待测帧序号对应得到初始的所述全局目标轨迹;根据所述多个待测帧序号中后续待测帧序号对应得到迭代更新后的所述全局目标轨迹。Wherein, the initial global target trajectory is correspondingly obtained according to the initial frame sequence number to be tested among the plurality of frame sequence numbers to be tested; the iteratively updated trajectory is obtained according to the subsequent frame sequence numbers under test among the plurality of frame sequence numbers to be tested Global target trajectory.
  3. 根据权利要求2所述的方法,其特征在于,对所述每个摄像头的当前待测帧进行目标检测,包括:The method according to claim 2, characterized in that, performing target detection on the current frame to be measured of each camera comprises:
    将所述每个摄像头的当前待测帧输入目标检测模型进行所述目标检测;Input the current frame to be measured of each camera into a target detection model to perform the target detection;
    其中,所述目标检测模型是基于神经网络训练得到的行人检测模型。Wherein, the target detection model is a pedestrian detection model obtained based on neural network training.
  4. 根据权利要求2所述的方法,其特征在于,在得到每个摄像头对应的检测框集合之后,还包括:The method according to claim 2, characterized in that, after obtaining the detection frame set corresponding to each camera, the method further comprises:
    根据每个摄像头的取景位置对所述每个摄像头对应的检测框集合中的每个检测框的框底中心点进行投影变换,从而确定所述每个检测框的地面坐标。Projective transformation is performed on the frame bottom center point of each detection frame in the detection frame set corresponding to each camera according to the viewing position of each camera, so as to determine the ground coordinates of each detection frame.
  5. 根据权利要求4所述的方法,其特征在于,所述多个摄像头的取景区域至少部分地重叠,所述方法还包括:The method according to claim 4, wherein the viewing areas of the plurality of cameras overlap at least partially, and the method further comprises:
    根据所述每个摄像头的取景区域在地面坐标系中划分所述每个摄像头的工作区域;Dividing the working area of each camera in a ground coordinate system according to the viewing area of each camera;
    其中,所述每个摄像头的工作区域互不重叠,若所述多个摄像头中的第一摄像头对应的任意一个检测框的地面坐标超出对应的工作区域,则在所述第一摄像头的检测框集合中去除所述任意一个检测框。Wherein, the working areas of each camera do not overlap each other. If the ground coordinate of any one of the detection frames corresponding to the first camera of the multiple cameras exceeds the corresponding working area, the detection frame of the first camera Remove any one of the detection frames from the set.
  6. 根据权利要求5所述的方法,其特征在于,所述方法还包括:The method according to claim 5, wherein the method further comprises:
    将所述每个摄像头的工作区域中的非关键区域截去。The non-critical area in the working area of each camera is cut off.
  7. 根据权利要求2所述的方法,其特征在于,根据所述每个摄像头对应的检测框集合进行跟踪,包括:The method according to claim 2, wherein the tracking according to the detection frame set corresponding to each camera comprises:
    采用多目标跟踪算法,并基于所述每个摄像头对应的检测框集合进行多目标跟踪,确定每个摄像头对应的局部跟踪信息;Adopting a multi-target tracking algorithm, and performing multi-target tracking based on the detection frame set corresponding to each camera, and determining the local tracking information corresponding to each camera;
    其中,所述多目标跟踪采用的参数基于所述每个摄像头的历史待测帧而确定。Wherein, the parameters used in the multi-target tracking are determined based on the historical frame to be measured of each camera.
  8. 根据权利要求7所述的方法,其特征在于,所述多目标跟踪算法为deepsort算法。The method according to claim 7, wherein the multi-target tracking algorithm is a deepsort algorithm.
  9. 根据权利要求7所述的方法,其特征在于,还包括:The method according to claim 7, further comprising:
    根据所述每个摄像头对应的局部跟踪信息为所述每个检测框添加身份标识;Adding an identity mark to each detection frame according to the local tracking information corresponding to each camera;
    基于所述每个检测框的身份标识和地面坐标确定迭代更新后的所述全局目标轨迹。The iteratively updated global target trajectory is determined based on the identity and ground coordinates of each detection frame.
  10. 根据权利要求7所述的方法,其特征在于,还包括:The method according to claim 7, further comprising:
    根据所述多个摄像头的工作区域确定所述多个摄像头之间的关联关系;Determining the association relationship between the multiple cameras according to the working areas of the multiple cameras;
    根据所述每个摄像头的局部跟踪信息确定对应工作区域中的新增检测框和消失检测框;Determine the newly added detection frame and the disappearance detection frame in the corresponding work area according to the local tracking information of each camera;
    根据所述多个摄像头之间的关联关系对处于不同工作区域中的新增检测框和消失检测框进行关联,得到关联信息;Associating newly-added detection frames and disappearing detection frames in different working areas according to the association relationship between the multiple cameras to obtain association information;
    根据所述关联信息确定迭代更新后的所述全局目标轨迹。The iteratively updated global target trajectory is determined according to the associated information.
  11. 一种目标跟踪装置,其特征在于,包括:A target tracking device is characterized in that it comprises:
    获取单元,用于获取设置于监控区域内的多个摄像头的当前待测帧;An acquiring unit for acquiring the current frame to be measured of multiple cameras arranged in the monitoring area;
    检测单元,用于依次对所述多个摄像头中每个摄像头的当前待测帧进行目标检测,得到每个摄像头对应的检测框集合;A detection unit, configured to perform target detection on the current frame to be tested of each camera in the plurality of cameras in turn, to obtain a set of detection frames corresponding to each camera;
    跟踪单元,用于根据所述每个摄像头对应的检测框集合进行目标跟踪,根据跟踪结果确定全局目标轨迹。The tracking unit is configured to perform target tracking according to the detection frame set corresponding to each camera, and determine the global target trajectory according to the tracking result.
  12. 根据权利要求11所述的装置,其特征在于,还包括:The device according to claim 11, further comprising:
    选帧单元,用于确定多个待测帧序号,根据所述多个待测帧序号按时序地迭代获取所述多个摄像头的当前待测帧,从而迭代地执行所述目标跟踪;The frame selection unit is configured to determine a plurality of frame numbers to be tested, and iteratively acquire the current frames to be tested of the plurality of cameras according to the number of frame numbers to be tested in time series, so as to iteratively perform the target tracking;
    其中,根据所述多个待测帧序号中初始待测帧序号对应得到初始的所述全局目标轨迹;根据所述多个待测帧序号中后续待测帧序号对应得到迭代更新后的所述全局目标轨迹。Wherein, the initial global target trajectory is correspondingly obtained according to the initial frame sequence number to be tested among the plurality of frame sequence numbers to be tested; the iteratively updated trajectory is obtained according to the subsequent frame sequence numbers under test among the plurality of frame sequence numbers to be tested Global target trajectory.
  13. 根据权利要求12所述的装置,其特征在于,所述检测单元,还用于:The device according to claim 12, wherein the detection unit is further configured to:
    将所述每个摄像头的当前待测帧输入目标检测模型进行所述目标检测;Input the current frame to be measured of each camera into a target detection model to perform the target detection;
    其中,所述目标检测模型是基于神经网络训练得到的行人检测模型。Wherein, the target detection model is a pedestrian detection model obtained based on neural network training.
  14. 根据权利要求12所述的装置,其特征在于,所述检测单元,还用于:The device according to claim 12, wherein the detection unit is further configured to:
    在得到每个摄像头对应的检测框集合之后,根据每个摄像头的取景位置对所述每个摄像头对应的检测框集合中的每个检测框的框底中心点进行投影变换,从而确定所述每个检测框的地面坐标。After the detection frame set corresponding to each camera is obtained, the center point of the frame bottom of each detection frame in the detection frame set corresponding to each camera is projected and transformed according to the viewing position of each camera, so as to determine the each camera. The ground coordinates of the detection frame.
  15. 根据权利要求14所述的装置,其特征在于,所述多个摄像头的取景区域至少部分地重叠,所述装置还用于:The device according to claim 14, wherein the viewing areas of the plurality of cameras overlap at least partially, and the device is further configured to:
    根据所述每个摄像头的取景区域在地面坐标系中划分所述每个摄像头的工作区域;Dividing the working area of each camera in a ground coordinate system according to the viewing area of each camera;
    其中,所述每个摄像头的工作区域互不重叠,若所述多个摄像头中的第一摄像头对应的任意一个检测框的地面坐标超出对应的工作区域,则在所述第一摄像头的检测框集合中去除所述任意一个检测框。Wherein, the working areas of each camera do not overlap each other. If the ground coordinate of any one of the detection frames corresponding to the first camera of the multiple cameras exceeds the corresponding working area, the detection frame of the first camera Remove any one of the detection frames from the set.
  16. 根据权利要求15所述的装置,其特征在于,所述检测单元,还用于:The device according to claim 15, wherein the detection unit is further configured to:
    将所述每个摄像头的工作区域中的非关键区域截去。The non-critical area in the working area of each camera is cut off.
  17. 根据权利要求12所述的装置,其特征在于,所述跟踪单元,还用于:The device according to claim 12, wherein the tracking unit is further configured to:
    采用多目标跟踪算法,并基于所述每个摄像头对应的检测框集合进行多目标跟踪,确定每个摄像头对应的局部跟踪信息;Adopting a multi-target tracking algorithm, and performing multi-target tracking based on the detection frame set corresponding to each camera, and determining the local tracking information corresponding to each camera;
    其中,所述多目标跟踪采用的参数基于所述每个摄像头的历史待测帧而确定。Wherein, the parameters used in the multi-target tracking are determined based on the historical frame to be measured of each camera.
  18. 根据权利要求17所述的装置,其特征在于,所述多目标跟踪算法为deepsort算法。The device according to claim 17, wherein the multi-target tracking algorithm is a deepsort algorithm.
  19. 根据权利要求17所述的装置,其特征在于,所述跟踪单元,还用于:The device according to claim 17, wherein the tracking unit is further configured to:
    根据所述每个摄像头对应的局部跟踪信息为所述每个检测框添加身份标识;Adding an identity mark to each detection frame according to the local tracking information corresponding to each camera;
    基于所述每个检测框的身份标识和地面坐标确定迭代更新后的所述全局目标轨迹。The iteratively updated global target trajectory is determined based on the identity and ground coordinates of each detection frame.
  20. 根据权利要求17所述的装置,其特征在于,所述跟踪单元,还用于:The device according to claim 17, wherein the tracking unit is further configured to:
    根据所述多个摄像头的工作区域确定所述多个摄像头之间的关联关系;Determining the association relationship between the multiple cameras according to the working areas of the multiple cameras;
    根据所述每个摄像头的局部跟踪信息确定对应工作区域中的新增检测框和消失检测框;Determine the newly added detection frame and the disappearance detection frame in the corresponding work area according to the local tracking information of each camera;
    根据所述多个摄像头之间的关联关系对处于不同工作区域中的新增检测框和消失检测框进行关联,得到关联信息;Associating newly-added detection frames and disappearing detection frames in different working areas according to the association relationship between the multiple cameras to obtain association information;
    根据所述关联信息确定迭代更新后的所述全局目标轨迹。The iteratively updated global target trajectory is determined according to the associated information.
  21. 一种目标跟踪系统,其特征在于,包括:设置于监控区域内的多个摄像头,以及与所述多个摄像头分别通信连接的目标跟踪装置;A target tracking system, characterized by comprising: a plurality of cameras arranged in a monitoring area, and a target tracking device respectively communicatively connected with the plurality of cameras;
    其中,所述目标跟踪装置被配置用于执行如权利要求1-10中的任一项所述的方法。Wherein, the target tracking device is configured to perform the method according to any one of claims 1-10.
  22. 一种目标跟踪装置,其特征在于,包括:A target tracking device is characterized in that it comprises:
    一个或者多个多核处理器;One or more multi-core processors;
    存储器,用于存储一个或多个程序;Memory, used to store one or more programs;
    当所述一个或多个程序被所述一个或者多个多核处理器执行时,使得所述一个或多个多核处理器实现:When the one or more programs are executed by the one or more multi-core processors, the one or more multi-core processors are caused to realize:
    获取设置于监控区域内的多个摄像头的当前待测帧;Obtain the current frame to be tested of multiple cameras set in the monitoring area;
    依次对所述多个摄像头中每个摄像头的当前待测帧进行目标检测,得到每个摄像头对应的检测框集合;Performing target detection on the current frame to be tested of each camera in the plurality of cameras in turn, to obtain a set of detection frames corresponding to each camera;
    根据所述每个摄像头对应的检测框集合进行目标跟踪,根据跟踪结果确定全局目标轨迹。Target tracking is performed according to the detection frame set corresponding to each camera, and the global target trajectory is determined according to the tracking result.
  23. 一种计算机可读存储介质,所述计算机可读存储介质存储有程序,当所述程序被多核处理器执行时,使得所述多核处理器执行如权利要求1-10中任一项所述的方法。A computer-readable storage medium, the computer-readable storage medium stores a program, and when the program is executed by a multi-core processor, the multi-core processor is caused to execute any one of claims 1-10 method.
PCT/CN2020/109081 2019-12-10 2020-08-14 Target tracking method, apparatus and system, and computer-readable storage medium WO2021114702A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201911258014.2A CN111145213A (en) 2019-12-10 2019-12-10 Target tracking method, device and system and computer readable storage medium
CN201911258014.2 2019-12-10

Publications (1)

Publication Number Publication Date
WO2021114702A1 true WO2021114702A1 (en) 2021-06-17

Family

ID=70518015

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/109081 WO2021114702A1 (en) 2019-12-10 2020-08-14 Target tracking method, apparatus and system, and computer-readable storage medium

Country Status (3)

Country Link
CN (1) CN111145213A (en)
TW (1) TWI795667B (en)
WO (1) WO2021114702A1 (en)

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113436295A (en) * 2021-06-25 2021-09-24 平安科技(深圳)有限公司 Living body breeding monitoring track drawing method, device, equipment and storage medium
CN113592903A (en) * 2021-06-28 2021-11-02 北京百度网讯科技有限公司 Vehicle track recognition method and device, electronic equipment and storage medium
CN113610895A (en) * 2021-08-06 2021-11-05 烟台艾睿光电科技有限公司 Target tracking method and device, electronic equipment and readable storage medium
CN113642454A (en) * 2021-08-11 2021-11-12 汇纳科技股份有限公司 Seat use condition identification method, system, equipment and computer storage medium
CN113673391A (en) * 2021-08-09 2021-11-19 浙江大华技术股份有限公司 Method, system, equipment and storage medium for video scene classification
CN113688278A (en) * 2021-07-13 2021-11-23 北京旷视科技有限公司 Information processing method, device, electronic equipment and computer readable medium
CN113744299A (en) * 2021-09-02 2021-12-03 上海安维尔信息科技股份有限公司 Camera control method and device, electronic equipment and storage medium
CN113743260A (en) * 2021-08-23 2021-12-03 北京航空航天大学 Pedestrian tracking method under dense pedestrian flow condition of subway platform
CN114120188A (en) * 2021-11-19 2022-03-01 武汉大学 Multi-pedestrian tracking method based on joint global and local features
CN114820700A (en) * 2022-04-06 2022-07-29 北京百度网讯科技有限公司 Object tracking method and device
CN115311820A (en) * 2022-07-11 2022-11-08 西安电子科技大学广州研究院 Intelligent security system near water
CN115497303A (en) * 2022-08-19 2022-12-20 招商新智科技有限公司 Expressway vehicle speed detection method and system under complex detection condition
CN115527162A (en) * 2022-05-18 2022-12-27 湖北大学 Multi-pedestrian re-identification method and system based on three-dimensional space
CN115690163A (en) * 2023-01-04 2023-02-03 中译文娱科技(青岛)有限公司 Target tracking method, system and storage medium based on image content
CN116363494A (en) * 2023-05-31 2023-06-30 睿克环境科技(中国)有限公司 Fish quantity monitoring and migration tracking method and system
CN117237418A (en) * 2023-11-15 2023-12-15 成都航空职业技术学院 Moving object detection method and system based on deep learning
CN117058331B (en) * 2023-10-13 2023-12-19 山东建筑大学 Indoor personnel three-dimensional track reconstruction method and system based on single monitoring camera
CN117315028A (en) * 2023-10-12 2023-12-29 北京多维视通技术有限公司 Method, device, equipment and medium for positioning fire point of outdoor fire scene
CN113627497B (en) * 2021-07-27 2024-03-12 武汉大学 Space-time constraint-based cross-camera pedestrian track matching method
WO2024108539A1 (en) * 2022-11-25 2024-05-30 京东方科技集团股份有限公司 Target people tracking method and apparatus
CN118678225A (en) * 2024-08-15 2024-09-20 湖州奥酷斯特智能科技有限公司 Camera blind area evaluation and switching method and system based on real-time position
CN113592903B (en) * 2021-06-28 2024-11-12 北京百度网讯科技有限公司 Vehicle track recognition method, device, electronic equipment and storage medium

Families Citing this family (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111145213A (en) * 2019-12-10 2020-05-12 中国银联股份有限公司 Target tracking method, device and system and computer readable storage medium
CN111815675B (en) * 2020-06-30 2023-07-21 北京市商汤科技开发有限公司 Target object tracking method and device, electronic equipment and storage medium
CN111967498A (en) * 2020-07-20 2020-11-20 重庆大学 Night target detection and tracking method based on millimeter wave radar and vision fusion
CN112200841B (en) * 2020-09-30 2021-08-27 杭州海宴科技有限公司 Cross-domain multi-camera tracking method and device based on pedestrian posture
CN112381132A (en) * 2020-11-11 2021-02-19 上汽大众汽车有限公司 Target object tracking method and system based on fusion of multiple cameras
CN112418064A (en) * 2020-11-19 2021-02-26 上海交通大学 Real-time automatic detection method for number of people in library reading room
CN112560621A (en) * 2020-12-08 2021-03-26 北京大学 Identification method, device, terminal and medium based on animal image
CN112906452A (en) * 2020-12-10 2021-06-04 叶平 Automatic identification, tracking and statistics method and system for antelope buffalo deer
CN112489085A (en) * 2020-12-11 2021-03-12 北京澎思科技有限公司 Target tracking method, target tracking device, electronic device, and storage medium
CN112634332A (en) * 2020-12-21 2021-04-09 合肥讯图信息科技有限公司 Tracking method based on YOLOv4 model and DeepsORT model
CN112614159B (en) * 2020-12-22 2023-04-07 浙江大学 Cross-camera multi-target tracking method for warehouse scene
CN112906483B (en) * 2021-01-25 2024-01-23 中国银联股份有限公司 Target re-identification method, device and computer readable storage medium
CN112819859B (en) * 2021-02-02 2023-06-02 重庆特斯联智慧科技股份有限公司 Multi-target tracking method and device applied to intelligent security
CN113012223B (en) * 2021-02-26 2023-01-24 清华大学 Target flow monitoring method and device, computer equipment and storage medium
CN113223060B (en) * 2021-04-16 2022-04-15 天津大学 Multi-agent cooperative tracking method and device based on data sharing and storage medium
CN113257003A (en) * 2021-05-12 2021-08-13 上海天壤智能科技有限公司 Traffic lane-level traffic flow counting system, method, device and medium thereof
CN113473091B (en) * 2021-07-09 2023-04-18 杭州海康威视数字技术股份有限公司 Camera association method, device, system, electronic equipment and storage medium
CN115086527B (en) * 2022-07-04 2023-05-12 天翼数字生活科技有限公司 Household video tracking and monitoring method, device, equipment and storage medium
CN115619832B (en) * 2022-12-20 2023-04-07 浙江莲荷科技有限公司 Multi-camera collaborative multi-target track confirmation method, system and related device
CN116071686B (en) * 2023-02-27 2023-06-20 中国信息通信研究院 Correlation analysis method, device and system for cameras in industrial Internet
CN117784798B (en) * 2024-02-26 2024-05-31 安徽蔚来智驾科技有限公司 Target tracking method, intelligent device and computer readable storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104331901A (en) * 2014-11-26 2015-02-04 北京邮电大学 TLD-based multi-view target tracking device and method
CN104463900A (en) * 2014-12-31 2015-03-25 天津汉光祥云信息科技有限公司 Method for automatically tracking target among multiple cameras
CN108876821A (en) * 2018-07-05 2018-11-23 北京云视万维科技有限公司 Across camera lens multi-object tracking method and system
CN108986158A (en) * 2018-08-16 2018-12-11 新智数字科技有限公司 A kind of across the scene method for tracing identified again based on target and device and Computer Vision Platform
CN111145213A (en) * 2019-12-10 2020-05-12 中国银联股份有限公司 Target tracking method, device and system and computer readable storage medium

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI502558B (en) * 2013-09-25 2015-10-01 Chunghwa Telecom Co Ltd Traffic Accident Monitoring and Tracking System
CN104933392A (en) * 2014-03-19 2015-09-23 通用汽车环球科技运作有限责任公司 Probabilistic people tracking using multi-view integration
CN106845385A (en) * 2017-01-17 2017-06-13 腾讯科技(上海)有限公司 The method and apparatus of video frequency object tracking
CN108875588B (en) * 2018-05-25 2022-04-15 武汉大学 Cross-camera pedestrian detection tracking method based on deep learning
CN109903260B (en) * 2019-01-30 2023-05-23 华为技术有限公司 Image processing method and image processing apparatus
CN110428448B (en) * 2019-07-31 2021-05-14 腾讯科技(深圳)有限公司 Target detection tracking method, device, equipment and storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104331901A (en) * 2014-11-26 2015-02-04 北京邮电大学 TLD-based multi-view target tracking device and method
CN104463900A (en) * 2014-12-31 2015-03-25 天津汉光祥云信息科技有限公司 Method for automatically tracking target among multiple cameras
CN108876821A (en) * 2018-07-05 2018-11-23 北京云视万维科技有限公司 Across camera lens multi-object tracking method and system
CN108986158A (en) * 2018-08-16 2018-12-11 新智数字科技有限公司 A kind of across the scene method for tracing identified again based on target and device and Computer Vision Platform
CN111145213A (en) * 2019-12-10 2020-05-12 中国银联股份有限公司 Target tracking method, device and system and computer readable storage medium

Cited By (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113436295A (en) * 2021-06-25 2021-09-24 平安科技(深圳)有限公司 Living body breeding monitoring track drawing method, device, equipment and storage medium
CN113436295B (en) * 2021-06-25 2023-09-15 平安科技(深圳)有限公司 Living body culture monitoring track drawing method, device, equipment and storage medium
CN113592903A (en) * 2021-06-28 2021-11-02 北京百度网讯科技有限公司 Vehicle track recognition method and device, electronic equipment and storage medium
CN113592903B (en) * 2021-06-28 2024-11-12 北京百度网讯科技有限公司 Vehicle track recognition method, device, electronic equipment and storage medium
CN113688278A (en) * 2021-07-13 2021-11-23 北京旷视科技有限公司 Information processing method, device, electronic equipment and computer readable medium
CN113627497B (en) * 2021-07-27 2024-03-12 武汉大学 Space-time constraint-based cross-camera pedestrian track matching method
CN113610895A (en) * 2021-08-06 2021-11-05 烟台艾睿光电科技有限公司 Target tracking method and device, electronic equipment and readable storage medium
CN113673391A (en) * 2021-08-09 2021-11-19 浙江大华技术股份有限公司 Method, system, equipment and storage medium for video scene classification
CN113642454A (en) * 2021-08-11 2021-11-12 汇纳科技股份有限公司 Seat use condition identification method, system, equipment and computer storage medium
CN113642454B (en) * 2021-08-11 2024-03-01 汇纳科技股份有限公司 Seat use condition identification method, system, equipment and computer storage medium
CN113743260B (en) * 2021-08-23 2024-03-05 北京航空航天大学 Pedestrian tracking method under condition of dense pedestrian flow of subway platform
CN113743260A (en) * 2021-08-23 2021-12-03 北京航空航天大学 Pedestrian tracking method under dense pedestrian flow condition of subway platform
CN113744299A (en) * 2021-09-02 2021-12-03 上海安维尔信息科技股份有限公司 Camera control method and device, electronic equipment and storage medium
CN114120188A (en) * 2021-11-19 2022-03-01 武汉大学 Multi-pedestrian tracking method based on joint global and local features
CN114120188B (en) * 2021-11-19 2024-04-05 武汉大学 Multi-row person tracking method based on joint global and local features
CN114820700B (en) * 2022-04-06 2023-05-16 北京百度网讯科技有限公司 Object tracking method and device
CN114820700A (en) * 2022-04-06 2022-07-29 北京百度网讯科技有限公司 Object tracking method and device
CN115527162B (en) * 2022-05-18 2023-07-18 湖北大学 Multi-pedestrian re-identification method and system based on three-dimensional space
CN115527162A (en) * 2022-05-18 2022-12-27 湖北大学 Multi-pedestrian re-identification method and system based on three-dimensional space
CN115311820A (en) * 2022-07-11 2022-11-08 西安电子科技大学广州研究院 Intelligent security system near water
CN115497303A (en) * 2022-08-19 2022-12-20 招商新智科技有限公司 Expressway vehicle speed detection method and system under complex detection condition
WO2024108539A1 (en) * 2022-11-25 2024-05-30 京东方科技集团股份有限公司 Target people tracking method and apparatus
CN115690163B (en) * 2023-01-04 2023-05-09 中译文娱科技(青岛)有限公司 Target tracking method, system and storage medium based on image content
CN115690163A (en) * 2023-01-04 2023-02-03 中译文娱科技(青岛)有限公司 Target tracking method, system and storage medium based on image content
CN116363494B (en) * 2023-05-31 2023-08-04 睿克环境科技(中国)有限公司 Fish quantity monitoring and migration tracking method and system
CN116363494A (en) * 2023-05-31 2023-06-30 睿克环境科技(中国)有限公司 Fish quantity monitoring and migration tracking method and system
CN117315028A (en) * 2023-10-12 2023-12-29 北京多维视通技术有限公司 Method, device, equipment and medium for positioning fire point of outdoor fire scene
CN117315028B (en) * 2023-10-12 2024-04-30 北京多维视通技术有限公司 Method, device, equipment and medium for positioning fire point of outdoor fire scene
CN117058331B (en) * 2023-10-13 2023-12-19 山东建筑大学 Indoor personnel three-dimensional track reconstruction method and system based on single monitoring camera
CN117237418A (en) * 2023-11-15 2023-12-15 成都航空职业技术学院 Moving object detection method and system based on deep learning
CN118678225A (en) * 2024-08-15 2024-09-20 湖州奥酷斯特智能科技有限公司 Camera blind area evaluation and switching method and system based on real-time position

Also Published As

Publication number Publication date
TW202123171A (en) 2021-06-16
CN111145213A (en) 2020-05-12
TWI795667B (en) 2023-03-11

Similar Documents

Publication Publication Date Title
WO2021114702A1 (en) Target tracking method, apparatus and system, and computer-readable storage medium
CN109242913B (en) Method, device, equipment and medium for calibrating relative parameters of collector
CN109344755B (en) Video action recognition method, device, equipment and storage medium
WO2021223367A1 (en) Single lens-based multi-pedestrian online tracking method and apparatus, device, and storage medium
JP2020042816A (en) Object detection method, device, apparatus, storage media, and vehicle
US20240144793A1 (en) Information processing system, method and computer readable medium for determining whether moving bodies appearing in first and second videos are the same or not using histogram
US20120099765A1 (en) Method and system of video object tracking
TWI798815B (en) Target re-identification method, device, and computer readable storage medium
CN108648140B (en) Image splicing method, system, equipment and storage medium
WO2018213702A1 (en) Augmented reality system
US20160313799A1 (en) Method and apparatus for identifying operation event
CN111091584A (en) Target tracking method, device, equipment and storage medium
CN114998768B (en) Intelligent construction site management system and method based on unmanned aerial vehicle
JP2017130042A (en) Video processing apparatus, video processing method, and program
CN112291478A (en) Method, device and equipment for monitoring high-altitude falling object and storage medium
CN111210457A (en) Aircraft listing method combining video analysis and positioning information
CN114170556A (en) Target track tracking method and device, storage medium and electronic equipment
US10187610B2 (en) Controlling display based on an object position in an imaging space
CN113312951A (en) Dynamic video target tracking system, related method, device and equipment
WO2023036274A1 (en) Video processing method and apparatus, electronic device, medium and program product
CN113762017B (en) Action recognition method, device, equipment and storage medium
CN113012223B (en) Target flow monitoring method and device, computer equipment and storage medium
CN112184766B (en) Object tracking method and device, computer equipment and storage medium
JP2024520375A (en) Semi-automated data collection and correlation for multi-camera tracking
CN114461078A (en) Man-machine interaction method based on artificial intelligence

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20900355

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20900355

Country of ref document: EP

Kind code of ref document: A1