WO2021031954A1 - 对象数量确定方法、装置、存储介质与电子设备 - Google Patents
对象数量确定方法、装置、存储介质与电子设备 Download PDFInfo
- Publication number
- WO2021031954A1 WO2021031954A1 PCT/CN2020/108677 CN2020108677W WO2021031954A1 WO 2021031954 A1 WO2021031954 A1 WO 2021031954A1 CN 2020108677 W CN2020108677 W CN 2020108677W WO 2021031954 A1 WO2021031954 A1 WO 2021031954A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- objects
- processed
- value
- preset threshold
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/52—Surveillance or monitoring of activities, e.g. for recognising suspicious objects
Definitions
- the present disclosure relates to the field of computer vision technology, and in particular to a method for determining the number of objects, a device for determining the number of objects, a computer-readable storage medium, and electronic equipment.
- the traditional method is to count the number of inflow and outflow objects at the entrance and exit of the target area, such as setting gates or infrared sensing equipment at the entrance and exit of the scenic spot, and setting barrier gate equipment at the entrance and exit of the parking lot.
- this method cannot count the number of objects in the open area. For example, the number of tourists in open scenic spots, the number of vehicles on the street, etc., and only the total number of objects in the target area can be counted, and the distribution of objects cannot be determined.
- the present disclosure provides a method for determining the number of objects, a device for determining the number of objects, a computer-readable storage medium, and electronic equipment, thereby improving the prior art at least to a certain extent.
- the density of objects is high, the accuracy of determining the number of objects is low.
- a method for determining the number of objects including: recognizing objects in an image to be processed, using the number of recognized objects as a first value; comparing the first value with a preset Threshold; if the first value is less than the preset threshold, the number of objects in the image to be processed is determined as the first value; if the first value is greater than the preset threshold, then Density detection is performed on the objects in the image to be processed to obtain a second value related to the number of objects, and the number of objects in the image to be processed is determined as the second value.
- the method further includes: acquiring a target image, dividing the target image into a plurality of regions, and using an image of each region as the image to be processed.
- each of the regions has a corresponding preset threshold.
- the recognizing the object in the image to be processed includes: recognizing the object in the image to be processed through a pre-trained first neural network model.
- the first neural network model includes the YOLO model (You Only Look Once, an algorithm framework for real-time target detection, including v1, v2, v3, etc.).
- the present disclosure Either version can be used).
- the performing density detection on the object in the image to be processed includes: performing density detection on the object in the image to be processed through a pre-trained second neural network model .
- the second neural network model includes: a first branch network for performing a first convolution process on the image to be processed to obtain a first characteristic image; and a second branch
- the network is used to perform the second convolution processing on the image to be processed to obtain the second characteristic image
- the third branch network is used to perform the third convolution processing on the image to be processed to obtain the third characteristic image
- the layer is used to merge the first feature image, the second feature image, and the third feature image into a final feature image
- the output layer is used to map the final feature image to a density image.
- a device for determining the number of objects including: a recognition module for recognizing objects in an image to be processed, and using the number of recognized objects as a first value; and a comparison module using To compare the first value with a preset threshold; a first determining module is configured to determine the number of objects in the image to be processed as the first if the first value is less than the preset threshold A value; a second determination module, configured to perform density detection on objects in the image to be processed if the first value is greater than the preset threshold to obtain a second value about the number of objects, and The number of the objects in the image to be processed is determined as the second value.
- the device further includes: an acquisition module configured to acquire a target image, divide the target image into a plurality of regions, and use an image of each region as the to-be Process images.
- each of the regions has a corresponding preset threshold.
- the recognition module is configured to recognize the object in the image to be processed through a pre-trained first neural network model.
- the first neural network model includes a YOLO model.
- the second determination module includes: a density detection unit, configured to perform density detection on the object in the image to be processed through a pre-trained second neural network model.
- the second neural network model includes: a first branch network for performing a first convolution process on the image to be processed to obtain a first characteristic image; and a second branch
- the network is used to perform the second convolution processing on the image to be processed to obtain the second characteristic image
- the third branch network is used to perform the third convolution processing on the image to be processed to obtain the third characteristic image
- the layer is used to merge the first feature image, the second feature image, and the third feature image into a final feature image
- the output layer is used to map the final feature image to a density image.
- a computer-readable storage medium having a computer program stored thereon, and when the computer program is executed by a processor, the method described in any one of the above is implemented.
- an electronic device including: a processor; and a memory for storing executable instructions of the processor; wherein the processor is configured to execute the executable instructions Perform any of the methods described above.
- the object in the image Recognize the object in the image to be processed, and determine whether the object in the image is sparse or dense based on the relationship between the first value obtained by the recognition and the preset threshold, so as to determine whether to use the first value as the final result or the density
- the second value obtained from the detection is used as the final result.
- the first value is greater than the preset threshold, the objects in the image are dense and may be blocked.
- the density detection method is used, and the second value obtained is used as the final structure to determine the number of objects more accurately.
- the combination of object recognition and density detection has high flexibility. By adjusting the preset threshold, this exemplary embodiment can be applied to various different scenarios, and has high applicability. .
- Fig. 1 shows a flowchart of a method for determining the number of objects in this exemplary embodiment
- Figure 2 shows the scenic spot monitoring image to be processed
- Figure 3 shows a visualized effect diagram of tourist recognition on surveillance images of scenic spots
- Fig. 4 shows a structure diagram of a neural network model in this exemplary embodiment
- Fig. 5 shows a schematic diagram of dividing regions of a target image in this exemplary embodiment
- Fig. 6 shows a flowchart of another method for determining the number of objects in this exemplary embodiment
- Fig. 7 shows a structural block diagram of a device for determining the number of objects in this exemplary embodiment
- FIG. 8 shows a computer-readable storage medium for implementing the above method in this exemplary embodiment
- Fig. 9 shows an electronic device for implementing the above-mentioned method in this exemplary embodiment.
- Example embodiments will now be described more fully with reference to the accompanying drawings.
- the example embodiments can be implemented in various forms, and should not be construed as being limited to the examples set forth herein; on the contrary, the provision of these embodiments makes the present disclosure more comprehensive and complete, and fully conveys the concept of the example embodiments To those skilled in the art.
- the described features, structures or characteristics may be combined in one or more embodiments in any suitable way.
- Exemplary embodiments of the present disclosure first provide a method for determining the number of objects in an image.
- the application scenarios of the method include but are not limited to: counting the number of people in areas such as scenic spots and shopping malls; counting vehicles in areas such as parking lots and streets Monitoring the number of ships in ports, docks and other areas; monitoring the number of livestock in the livestock farm.
- the following takes the scenario of counting the number of people in a scenic spot as an example, and the method content is also applicable to other scenarios.
- Fig. 1 shows the method flow of this exemplary embodiment, which may include steps S110 to S140:
- Step S110 Recognize the objects in the image to be processed, and use the number of recognized objects as the first value.
- the image to be processed may be a surveillance image of a scenic spot or a GIS image (Geographic Information System, where the GIS image includes a satellite view of the ground surface, a population heat map, etc.), etc.
- GIS image Geographic Information System
- the GIS image includes a satellite view of the ground surface, a population heat map, etc.
- pull the video stream of surveillance cameras in the scenic area through a background computer or server.
- web cameras provide protocols such as rtmp (Real Time Messaging Protocol), http (Hyper Text Transfer Protocol, Hypertext Transfer Protocol), etc.
- OpenCV Open Source Computer Vision Library
- a deep learning technology may be used to recognize objects in the image to be processed through a pre-trained first neural network model.
- the first neural network model can use the YOLO model, the YOLO model can be trained through the open source dense pedestrian detection data set, or the pictures in the application scene can be manually labeled to obtain the data set (for example, from a large number of scenic spot monitoring images Out of tourists) and conduct training.
- the YOLO model takes the scenic spot monitoring image as input, and the bounding box information of all tourists in the image as output.
- input Figure 2 into the YOLO model, and the output visualization effect can be referred to as shown in Figure 3.
- the YOLO model The tourists in the image are identified, and the (x, y, w, h) of the bounding box of each visitor is actually obtained.
- x and y represent the position coordinates of the center of the bounding box in the image
- w and h represent the width of the bounding box.
- the first neural network model can also use R-CNN (Region-Convolutional Neural Network, or improved versions of Fast R-CNN, Faster R-CNN, etc.), SSD (Single Shot MultiBox Detector, single-step Multi-frame target detection) and other target detection algorithm models.
- the contour of an object may also be detected from the image to be processed, and an object whose contour shape is close to the shape of the object is recognized as an object.
- the number of objects recognized from the image to be processed is the first value.
- Step S120 comparing the first value with a preset threshold.
- the first value obtained in step S110 is close to the true number of objects, that is, the reliability of the first value is higher. ; In the case of a large number of objects, there may be a problem that several objects are occluded, or the image resolution of a single object is low, making the object difficult to identify, and the reliability of the first value is low.
- the first neural network model is used to identify the tourists in the surveillance image, and there are many cases of missed detection in the central area with dense tourists.
- the first value is credible by comparing the relative size of the first value and the preset threshold. If the first value is less than the preset threshold, the objects in the image to be processed are relatively sparse, and the first value is credible ; On the contrary, the objects in the image to be processed are relatively dense, and the first value is unreliable.
- the preset threshold may be determined according to experience, the characteristics of the region corresponding to the image to be processed, the size relationship between the image to be processed and the object, etc., which is not particularly limited in the present disclosure.
- step S130 if the first value is less than the preset threshold, the number of objects in the image to be processed is determined as the first value.
- step S130 when the condition of step S130 is satisfied, the first value is credible, so it can be used as the number of objects in the image to be processed, and the result is output.
- Step S140 if the first value is greater than the preset threshold, density detection is performed on objects in the image to be processed to obtain a second value regarding the number of objects, and the number of objects in the image to be processed is determined as the second value.
- step S140 When the condition of step S140 is met, and the first value is not credible, another method other than object recognition may be used for processing, that is, the method of density detection, to determine the number of objects in the image to be processed.
- Density detection is different from object recognition. It mainly regresses the probability of objects in each region (or each pixel) in the image to be processed, and obtains the number of objects in the image to be processed in a statistical manner, which is the second value mentioned above. In the case of many objects, especially in the case of dense distribution and occlusion, density detection has higher credibility than object recognition. Therefore, the second value can be used as the number of objects in the image to be processed, and the result can be output .
- the case where the first value is equal to the preset threshold can be regarded as a special case meeting the condition of step S130, or as a special case meeting the condition of step S140, so that step S130 or S140 is executed. There is no particular limitation on this.
- the density detection of the object in the image to be processed may be performed through a pre-trained second neural network model.
- the second neural network model may adopt the MCNN model (Multi-column Convolutional Neural Network, multi-column convolutional neural network).
- the MCNN model 400 shows a structure of the MCNN model 400, which may include: an input layer 410 for inputting Process the image; the first branch network 420 is used to perform the first convolution process on the image to be processed to obtain the first feature image; the second branch network 430 is used to perform the second convolution process on the image to be processed to obtain the second feature image The third branch network 440 is used to perform the third convolution processing on the image to be processed to obtain the third feature image; the merging layer 450 is used to merge the first feature image, the second feature image and the third feature image into the final feature Image; output layer 460, used to map the final feature image to a density image.
- the first convolution processing, the second convolution processing and the third convolution processing respectively include a series of operations such as convolution and pooling.
- the parameters used For example, the size of the convolution kernel, pooling parameters, etc.
- the density image the value of each point represents the probability that the point is an object.
- the values of all points are accumulated to obtain the second value representing the number of objects in the image to be processed.
- the training of the MCNN model can be based on open source data sets.
- the image annotation can be the coordinates of each human head.
- the geometric adaptive Gaussian kernel is used to convert the human head coordinates into a probability density image.
- the sum of the probabilities of each human head region is 1.
- the second neural network model can also use other density detection networks, such as a variant of MCNN.
- a fourth branch network is added, or in the first, second, or third branch.
- An intermediate layer is added to the network, or one or more fully connected layers are added, and the present disclosure does not specifically limit this.
- this exemplary embodiment recognizes the object in the image to be processed, and determines whether the object in the image is sparse or dense according to the relationship between the first value obtained by the recognition and the preset threshold, so as to determine whether to use the first value.
- One value is used as the final result, and the second value obtained by density detection is used as the final result.
- the first value is greater than the preset threshold, the objects in the image are dense and may be blocked. In this case, the density detection method is used, and the second value obtained is used as the final structure to determine the number of objects more accurately. , So that the exemplary embodiment has a higher accuracy.
- the combination of object recognition and density detection has high flexibility. By adjusting the preset threshold, this exemplary embodiment can be applied to various different scenarios, and has high applicability. .
- the target image after acquiring the target image, the target image may be divided into multiple regions, and the images of each region may be used as the image to be processed.
- the target image is a complete image that needs to determine the number of objects.
- the captured image contains part of the fixed scene, sky, etc. Inside, there are more interference factors, which will cause a certain disturbance to the number of tourists, and the distribution of tourists in different regions also has a difference between dense and sparse, which can be dealt with separately.
- Figure 2 can be divided into multiple regions based on prior experience, and the method process of Figure 1 is performed on each region image. Finally, the number of objects in each region is added to obtain the target image The total number of objects in.
- the preset threshold used can be the same or different, that is, each area can have a unified preset threshold.
- the thresholds may also have corresponding preset thresholds. For example, in FIG. 5, a smaller preset threshold can be set for areas two and three, and a larger preset threshold for area four.
- the preset threshold for each area can be determined based on experience or calculated based on image features.
- calculate the area of a part of the image where tourists may appear in each area divide it by the image area occupied by each tourist, and estimate the number of tourists in each area
- the number of tourists when it is full and there is no occlusion can be used as the preset threshold, or multiplied by an empirical coefficient less than 1 (such as 0.9) as the preset threshold, etc.
- the present disclosure does not make special limited. Using targeted preset thresholds for each region can more accurately obtain the total number of objects in the target image.
- Fig. 6 shows another process of this exemplary embodiment, including: step S601, acquiring a target image, for example, a surveillance image; step S602, dividing the target image into multiple regions; step S603, using the The image is an image to be processed, and steps S604 to S608 are performed respectively: step S604, through object recognition, detects the number of objects in the image to be processed, and is the first value; step S605, judges the size of the first value and the preset threshold; step S606 If the first value is less than the preset threshold, it is determined that the number of objects in the area is the first value; step S607, if the first value is greater than the preset threshold, the first value is not credible, and the object density of the image to be processed needs to be checked.
- step S608 determine the number of objects in the area as the second value; based on the above process, the number of objects in each area can be obtained, and finally step S609 is performed to accumulate the number of objects in each area to obtain the target image The total number of objects in the target image, which ultimately determines the number of objects in the target image.
- the device 700 may include: a recognition module 710 for recognizing objects in the image to be processed, The number is used as the first value; the comparison module 720 is used to compare the first value with a preset threshold; the first determination module 730 is used to determine the number of objects in the image to be processed as the first value if the first value is less than the preset threshold. A value; a second determination module 740, configured to, if the first value is greater than a preset threshold, perform density detection on objects in the image to be processed to obtain a second value on the number of objects, and determine the number of objects in the image to be processed Is the second value.
- the device 700 for determining the number of objects may further include: an acquisition module (not shown in the figure), configured to acquire a target image, divide the target image into a plurality of regions, and use the image of each region as the target image. Process images.
- each of the foregoing regions has a corresponding preset threshold.
- the recognition module 710 may be used to recognize an object in the image to be processed through a pre-trained first neural network model.
- the first neural network model may be a YOLO model.
- the second determination module 740 may include: a density detection unit (not shown in the figure), configured to perform density detection on objects in the image to be processed through a pre-trained second neural network model.
- the second neural network model may include: a first branch network for performing a first convolution process on the image to be processed to obtain a first feature image; a second branch network for performing a first feature image on the image to be processed The second convolution process is used to obtain the second feature image; the third branch network is used to perform the third convolution process on the image to be processed to obtain the third feature image; the merge layer is used to combine the first feature image and the second feature image And the third feature image are merged into the final feature image; the output layer is used to map the final feature image into a density image.
- Exemplary embodiments of the present disclosure also provide a computer-readable storage medium on which is stored a program product capable of implementing the above method of this specification.
- various aspects of the present disclosure can also be implemented in the form of a program product, which includes program code.
- the program product runs on a terminal device, the program code is used to make the terminal device execute the above-mentioned instructions in this specification.
- the steps according to various exemplary embodiments of the present disclosure are described in the "Exemplary Methods" section.
- a program product 800 for implementing the above method according to an exemplary embodiment of the present disclosure is described. It can adopt a portable compact disk read-only memory (CD-ROM) and include program codes, and can be used in a terminal Running on equipment, such as a personal computer.
- CD-ROM compact disk read-only memory
- the program product of the present disclosure is not limited thereto.
- the readable storage medium can be any tangible medium that contains or stores a program, and the program can be used by or in combination with an instruction execution system, device, or device.
- the program product can adopt any combination of one or more readable media.
- the readable medium may be a readable signal medium or a readable storage medium.
- the readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device, or device, or a combination of any of the above. More specific examples (non-exhaustive list) of readable storage media include: electrical connections with one or more wires, portable disks, hard disks, random access memory (RAM), read only memory (ROM), erasable Type programmable read only memory (EPROM or flash memory), optical fiber, portable compact disk read only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
- the computer-readable signal medium may include a data signal propagated in baseband or as a part of a carrier wave, and readable program code is carried therein. This propagated data signal can take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
- the readable signal medium may also be any readable medium other than a readable storage medium, and the readable medium may send, propagate, or transmit a program for use by or in combination with the instruction execution system, apparatus, or device.
- the program code contained on the readable medium can be transmitted by any suitable medium, including but not limited to wireless, wired, optical cable, RF, etc., or any suitable combination of the foregoing.
- the program code for performing the operations of the present disclosure can be written in any combination of one or more programming languages.
- the programming languages include object-oriented programming languages—such as Java, C++, etc., as well as conventional procedural programming. Language-such as "C" language or similar programming language.
- the program code can be executed entirely on the user's computing device, partly on the user's device, executed as an independent software package, partly on the user's computing device and partly executed on the remote computing device, or entirely on the remote computing device or server Executed on.
- the remote computing device can be connected to a user computing device through any kind of network, including a local area network (LAN) or a wide area network (WAN), or can be connected to an external computing device (for example, using Internet service providers) Business to connect via the Internet).
- LAN local area network
- WAN wide area network
- Internet service providers Internet service providers
- Exemplary embodiments of the present disclosure also provide an electronic device capable of implementing the above method.
- the electronic device 900 according to this exemplary embodiment of the present disclosure will be described below with reference to FIG. 9.
- the electronic device 900 shown in FIG. 9 is only an example, and should not bring any limitation to the function and scope of use of the embodiments of the present disclosure.
- the electronic device 900 may be in the form of a general-purpose computing device.
- the components of the electronic device 900 may include, but are not limited to: the aforementioned at least one processing unit 910, the aforementioned at least one storage unit 920, a bus 930 connecting different system components (including the storage unit 920 and the processing unit 910), and a display unit 940.
- the storage unit 920 stores program codes, and the program codes can be executed by the processing unit 910 so that the processing unit 910 executes the steps according to various exemplary embodiments of the present disclosure described in the "Exemplary Method" section of this specification.
- the processing unit 910 may execute the method steps shown in FIG. 4 or FIG. 5 and the like.
- the storage unit 920 may include a readable medium in the form of a volatile storage unit, such as a random access storage unit (RAM) 921 and/or a cache storage unit 922, and may further include a read-only storage unit (ROM) 923.
- RAM random access storage unit
- ROM read-only storage unit
- the storage unit 920 may also include a program/utility tool 924 having a set of (at least one) program module 925.
- program module 925 includes but is not limited to: an operating system, one or more application programs, other program modules, and program data, Each of these examples or some combination may include the implementation of a network environment.
- the bus 930 may represent one or more of several types of bus structures, including a storage unit bus or a storage unit controller, a peripheral bus, a graphics acceleration port, a processing unit, or a local area using any bus structure among multiple bus structures. bus.
- the electronic device 900 may also communicate with one or more external devices 1000 (such as keyboards, pointing devices, Bluetooth devices, etc.), and may also communicate with one or more devices that enable users to interact with the electronic device 900, and/or communicate with Any device (eg, router, modem, etc.) that enables the electronic device 900 to communicate with one or more other computing devices. This communication can be performed through an input/output (I/O) interface 950.
- the electronic device 900 may also communicate with one or more networks (for example, a local area network (LAN), a wide area network (WAN), and/or a public network, such as the Internet) through the network adapter 960.
- networks for example, a local area network (LAN), a wide area network (WAN), and/or a public network, such as the Internet
- the network adapter 960 communicates with other modules of the electronic device 900 through the bus 930. It should be understood that although not shown in the figure, other hardware and/or software modules can be used in conjunction with the electronic device 900, including but not limited to: microcode, device drivers, redundant processing units, external disk drive arrays, RAID systems, tape drives And data backup storage system, etc.
- the exemplary embodiments described herein can be implemented by software, or can be implemented by combining software with necessary hardware. Therefore, the technical solution according to the embodiments of the present disclosure can be embodied in the form of a software product, which can be stored in a non-volatile storage medium (can be a CD-ROM, U disk, mobile hard disk, etc.) or on the network , Including several instructions to make a computing device (which may be a personal computer, a server, a terminal device, or a network device, etc.) execute the method according to the exemplary embodiment of the present disclosure.
- a computing device which may be a personal computer, a server, a terminal device, or a network device, etc.
- modules or units of the device for action execution are mentioned in the above detailed description, this division is not mandatory.
- the features and functions of two or more modules or units described above may be embodied in one module or unit.
- the features and functions of a module or unit described above can be further divided into multiple modules or units to be embodied.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- General Engineering & Computer Science (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (10)
- 一种对象数量确定方法,其特征在于,包括:对待处理图像中的对象进行识别,将识别到的所述对象的数量作为第一数值;比较所述第一数值和预设阈值;如果所述第一数值小于所述预设阈值,则将所述待处理图像中所述对象的数量确定为所述第一数值;如果所述第一数值大于所述预设阈值,则对所述待处理图像中的对象进行密度检测,得到关于所述对象数量的第二数值,并将所述待处理图像中所述对象的数量确定为所述第二数值。
- 根据权利要求1所述的方法,其特征在于,所述方法还包括:获取目标图像,将所述目标图像划分为多个区域,分别以各所述区域的图像作为所述待处理图像。
- 根据权利要求2所述的方法,其特征在于,各所述区域具有对应的预设阈值。
- 根据权利要求1所述的方法,其特征在于,所述对待处理图像中的对象进行识别,包括:通过预先训练的第一神经网络模型对所述待处理图像中的对象进行识别。
- 根据权利要求4所述的方法,其特征在于,所述第一神经网络模型包括YOLO模型。
- 根据权利要求1所述的方法,其特征在于,所述对所述待处理图像中的对象进行密度检测,包括:通过预先训练的第二神经网络模型对所述待处理图像中的对象进行密度检测。
- 根据权利要求6所述的方法,其特征在于,所述第二神经网络模型包括:第一分支网络,用于对所述待处理图像进行第一卷积处理,得到第一特征图像;第二分支网络,用于对所述待处理图像进行第二卷积处理,得到第二特征图像;第三分支网络,用于对所述待处理图像进行第三卷积处理,得到第三特征图像;合并层,用于将所述第一特征图像、第二特征图像和第三特征图像合并为最终特征图像;输出层,用于将所述最终特征图像映射为密度图像。
- 一种对象数量确定装置,其特征在于,包括:识别模块,用于对待处理图像中的对象进行识别,将识别到的所述对象的数量作为第一数值;比较模块,用于比较所述第一数值和预设阈值;第一确定模块,用于如果所述第一数值小于所述预设阈值,则将所述待处理图像中所述对象的数量确定为所述第一数值;第二确定模块,用于如果所述第一数值大于所述预设阈值,则对所述待处理图像 中的对象进行密度检测,得到关于所述对象数量的第二数值,并将所述待处理图像中所述对象的数量确定为所述第二数值。
- 一种计算机可读存储介质,其上存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现权利要求1-7任一项所述的方法。
- 一种电子设备,其特征在于,包括:处理器;以及存储器,用于存储所述处理器的可执行指令;其中,所述处理器配置为经由执行所述可执行指令来执行权利要求1-7任一项所述的方法。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910769944.8A CN110472599B (zh) | 2019-08-20 | 2019-08-20 | 对象数量确定方法、装置、存储介质与电子设备 |
CN201910769944.8 | 2019-08-20 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021031954A1 true WO2021031954A1 (zh) | 2021-02-25 |
Family
ID=68512644
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/108677 WO2021031954A1 (zh) | 2019-08-20 | 2020-08-12 | 对象数量确定方法、装置、存储介质与电子设备 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN110472599B (zh) |
WO (1) | WO2021031954A1 (zh) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112988823A (zh) * | 2021-03-09 | 2021-06-18 | 北京百度网讯科技有限公司 | 对象处理方法、装置、设备及存储介质 |
CN113283499A (zh) * | 2021-05-24 | 2021-08-20 | 南京航空航天大学 | 一种基于深度学习的三维编制物编制密度检测方法 |
CN113486732A (zh) * | 2021-06-17 | 2021-10-08 | 普联国际有限公司 | 一种人群密度估计方法、装置、设备及存储介质 |
CN113807260A (zh) * | 2021-09-17 | 2021-12-17 | 北京百度网讯科技有限公司 | 数据处理方法、装置、电子设备和存储介质 |
CN114785943A (zh) * | 2022-03-31 | 2022-07-22 | 联想(北京)有限公司 | 一种数据确定方法、设备和计算机可读存储介质 |
CN115384796A (zh) * | 2022-04-01 | 2022-11-25 | 中国民用航空飞行学院 | 一种能够增加旅客转运效率的机场管理系统 |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110472599B (zh) * | 2019-08-20 | 2021-09-03 | 北京海益同展信息科技有限公司 | 对象数量确定方法、装置、存储介质与电子设备 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130063282A1 (en) * | 2010-05-31 | 2013-03-14 | Central Signal, Llc | Roadway detection |
CN106845344A (zh) * | 2016-12-15 | 2017-06-13 | 重庆凯泽科技股份有限公司 | 人群统计方法及装置 |
CN108399388A (zh) * | 2018-02-28 | 2018-08-14 | 福州大学 | 一种中高密度人群数量统计方法 |
CN110472599A (zh) * | 2019-08-20 | 2019-11-19 | 北京海益同展信息科技有限公司 | 对象数量确定方法、装置、存储介质与电子设备 |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1089214A3 (en) * | 1999-09-30 | 2005-01-26 | Matsushita Electric Industrial Co., Ltd. | Apparatus and method for image recognition |
CN101320427A (zh) * | 2008-07-01 | 2008-12-10 | 北京中星微电子有限公司 | 一种具有对目标的辅助监测功能的视频监控方法及系统 |
CN102831613B (zh) * | 2012-08-29 | 2014-11-19 | 武汉大学 | 一种并行的分形网络演化影像分割方法 |
CN107093171B (zh) * | 2016-02-18 | 2021-04-30 | 腾讯科技(深圳)有限公司 | 一种图像处理方法及装置、系统 |
CN108009477B (zh) * | 2017-11-10 | 2020-08-21 | 东软集团股份有限公司 | 图像的人流数量检测方法、装置、存储介质及电子设备 |
CN110008783A (zh) * | 2018-01-04 | 2019-07-12 | 杭州海康威视数字技术股份有限公司 | 基于神经网络模型的人脸活体检测方法、装置及电子设备 |
CN108875587A (zh) * | 2018-05-24 | 2018-11-23 | 北京飞搜科技有限公司 | 目标分布检测方法及设备 |
CN109224442B (zh) * | 2018-09-03 | 2021-06-11 | 腾讯科技(深圳)有限公司 | 虚拟场景的数据处理方法、装置及存储介质 |
CN109389589A (zh) * | 2018-09-28 | 2019-02-26 | 百度在线网络技术(北京)有限公司 | 用于统计人数的方法和装置 |
CN109815868B (zh) * | 2019-01-15 | 2022-02-01 | 腾讯科技(深圳)有限公司 | 一种图像目标检测方法、装置及存储介质 |
-
2019
- 2019-08-20 CN CN201910769944.8A patent/CN110472599B/zh active Active
-
2020
- 2020-08-12 WO PCT/CN2020/108677 patent/WO2021031954A1/zh active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130063282A1 (en) * | 2010-05-31 | 2013-03-14 | Central Signal, Llc | Roadway detection |
CN106845344A (zh) * | 2016-12-15 | 2017-06-13 | 重庆凯泽科技股份有限公司 | 人群统计方法及装置 |
CN108399388A (zh) * | 2018-02-28 | 2018-08-14 | 福州大学 | 一种中高密度人群数量统计方法 |
CN110472599A (zh) * | 2019-08-20 | 2019-11-19 | 北京海益同展信息科技有限公司 | 对象数量确定方法、装置、存储介质与电子设备 |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112988823A (zh) * | 2021-03-09 | 2021-06-18 | 北京百度网讯科技有限公司 | 对象处理方法、装置、设备及存储介质 |
CN112988823B (zh) * | 2021-03-09 | 2024-06-04 | 北京百度网讯科技有限公司 | 对象处理方法、装置、设备及存储介质 |
CN113283499A (zh) * | 2021-05-24 | 2021-08-20 | 南京航空航天大学 | 一种基于深度学习的三维编制物编制密度检测方法 |
CN113486732A (zh) * | 2021-06-17 | 2021-10-08 | 普联国际有限公司 | 一种人群密度估计方法、装置、设备及存储介质 |
CN113807260A (zh) * | 2021-09-17 | 2021-12-17 | 北京百度网讯科技有限公司 | 数据处理方法、装置、电子设备和存储介质 |
CN113807260B (zh) * | 2021-09-17 | 2022-07-12 | 北京百度网讯科技有限公司 | 数据处理方法、装置、电子设备和存储介质 |
CN114785943A (zh) * | 2022-03-31 | 2022-07-22 | 联想(北京)有限公司 | 一种数据确定方法、设备和计算机可读存储介质 |
CN114785943B (zh) * | 2022-03-31 | 2024-03-05 | 联想(北京)有限公司 | 一种数据确定方法、设备和计算机可读存储介质 |
CN115384796A (zh) * | 2022-04-01 | 2022-11-25 | 中国民用航空飞行学院 | 一种能够增加旅客转运效率的机场管理系统 |
Also Published As
Publication number | Publication date |
---|---|
CN110472599A (zh) | 2019-11-19 |
CN110472599B (zh) | 2021-09-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021031954A1 (zh) | 对象数量确定方法、装置、存储介质与电子设备 | |
US10735694B2 (en) | System and method for activity monitoring using video data | |
CN105051754B (zh) | 用于通过监控系统检测人的方法和装置 | |
CN108256404B (zh) | 行人检测方法和装置 | |
WO2021051601A1 (zh) | 利用Mask R-CNN选择检测框的方法及系统、电子装置及存储介质 | |
AU2018379393B2 (en) | Monitoring systems, and computer implemented methods for processing data in monitoring systems, programmed to enable identification and tracking of human targets in crowded environments | |
CN110348522B (zh) | 一种图像检测识别方法及系统、电子设备、图像分类网络优化方法及系统 | |
CN112861575A (zh) | 一种行人结构化方法、装置、设备和存储介质 | |
CN110781836A (zh) | 人体识别方法、装置、计算机设备及存储介质 | |
CN113205037B (zh) | 事件检测的方法、装置、电子设备以及可读存储介质 | |
WO2022041830A1 (zh) | 行人重识别方法和装置 | |
TW202026948A (zh) | 活體檢測方法、裝置以及儲存介質 | |
WO2020094088A1 (zh) | 一种图像抓拍方法、监控相机及监控系统 | |
CN112669344A (zh) | 一种运动物体的定位方法、装置、电子设备及存储介质 | |
JP7273129B2 (ja) | 車線検出方法、装置、電子機器、記憶媒体及び車両 | |
WO2023273041A1 (zh) | 车路协同中目标检测方法、装置和路侧设备 | |
CN113780270B (zh) | 目标检测方法和装置 | |
US20150104067A1 (en) | Method and apparatus for tracking object, and method for selecting tracking feature | |
CN113807361B (zh) | 神经网络、目标检测方法、神经网络训练方法及相关产品 | |
WO2022199360A1 (zh) | 运动物体的定位方法、装置、电子设备及存储介质 | |
CN112634368A (zh) | 场景目标的空间与或图模型生成方法、装置及电子设备 | |
WO2021022698A1 (zh) | 尾随检测方法、装置、电子设备及存储介质 | |
CN114663871A (zh) | 图像识别方法、训练方法、装置、系统及存储介质 | |
CN112541403B (zh) | 一种利用红外摄像头的室内人员跌倒检测方法 | |
CN113869258A (zh) | 交通事件检测方法、装置、电子设备及可读存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20854419 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20854419 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20854419 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 12.09.2022) |