[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CA3094424C - Safety monitoring and early-warning method for man-machine interaction behavior of underground conveyor belt operator - Google Patents

Safety monitoring and early-warning method for man-machine interaction behavior of underground conveyor belt operator Download PDF

Info

Publication number
CA3094424C
CA3094424C CA3094424A CA3094424A CA3094424C CA 3094424 C CA3094424 C CA 3094424C CA 3094424 A CA3094424 A CA 3094424A CA 3094424 A CA3094424 A CA 3094424A CA 3094424 C CA3094424 C CA 3094424C
Authority
CA
Canada
Prior art keywords
human body
points
body key
key
belt
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CA3094424A
Other languages
French (fr)
Other versions
CA3094424A1 (en
Inventor
Yanjing Sun
Kaiwen DONG
Xiaozhou Cheng
Xiao YUN
Xiaofeng Hou
Bowen Wang
Bin Wang
Hongli XU
Xiaojing Chen
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China University of Mining and Technology CUMT
Original Assignee
China University of Mining and Technology CUMT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from CN201910540349.7A external-priority patent/CN110425005B/en
Application filed by China University of Mining and Technology CUMT filed Critical China University of Mining and Technology CUMT
Publication of CA3094424A1 publication Critical patent/CA3094424A1/en
Application granted granted Critical
Publication of CA3094424C publication Critical patent/CA3094424C/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Analysis (AREA)

Abstract

A safety monitoring and early-warning method for a man-machine interaction behavior of an underground conveyor belt operator is disclosed. The present invention determines a belt position in a video based on a camera calibration principle, and delineates a three-dimensional region of interest (ROI) based on the belt position and size. By using a "bottom-up" key-point extraction method, multi-person key-point detection of conveyor belt operators is performed by detection followed by clustering, thus guaranteeing the detection accuracy and improving the detection efficiency. The human body key-points and the ROI are separately projected twice, and a relationship between the human body and the belt position is estimated based on the two projections to screen out unsafe actions during man-machine interaction and give an early warning, so as to avoid a major safety hazard in the belt conveyor system caused by an abnormal contact of the operator with the belt region.

Description

SAFETY MONITORING AND EARLY-WARNING METHOD FOR MAN-MACHINE INTERACTION BEHAVIOR OF UNDERGROUND CONVEYOR
BELT OPERATOR
BACKGROUND OF THE INVENTION
Field of the Invention The present invention relates to the field of monitoring of underground operations, and in particular, to a behavior safety monitoring method for an underground conveyor belt operator.
Description of Related Art China leads the world in the development of coal production industry all along.
However, as a high-risk industry, coal mining has posed great production safety hazards over the years. Belt conveyors in coal mines, as the most common underground transport system currently, directly affect the safety level of coal production during operation. Safety management of the belt conveyor system mainly resorts to a manual monitoring manner at the present stage, which has many limitations such as short operation duration, a narrow functional coverage, and a high cost. Therefore, it is of great significance to develop a safety early-warning system based on video monitoring for the belt conveyors and related operators, thus improving the production safety level of the belt conveyor system.
The current behavior safety early-warning system based on video monitoring pre-warns coal miners of danger merely by analyzing and identifying actions of the miners.
For example, a behavior safety monitoring method based on feature extraction and support vector machine (SVM) classification that was proposed by Yang chaoyu et al.
in 2016, and an underground dangerous region monitoring method based on moving target detection that was proposed by Zhang liya in 2017 both use rectangles to locate underground operators, implementing behavior safety monitoring of the operators. For limitations of the methods using the rectangles, Zhu aichun et al. in 2018 proposed a method for identifying postures of underground coal miners based on an hourglass Date Recue/Date Received 2020-09-24 network with hard example mining for generative adversarial training, which implements positioning and safety identification of underground operators by detecting human body key-points, improving accuracy and robustness of safety identification for the underground operators. The foregoing methods achieve good evaluation and identification effects for unsafe behaviors without man-machine interaction (namely, interaction between human and equipment). However, most underground accidents occur in a process of unsafe man-machine interaction. Therefore, without identification of man-machine interaction behaviors, safety monitoring and early warning cannot be accurately implemented merely by identifying actions or determining positions of the operators. Moreover, an existing algorithm model (for example, the hard example mining for generative adversarial training used by Zhu aichun et al.) is complex in structure and slow in operation speed, and has a problem that the detection speed linearly decreases as the number of detected persons increases, failing to have a wide application prospect.
SUMMARY OF THE INVENTION
Technical Solution To solve the technical problems mentioned above in the prior art, the present invention provides a safety monitoring and early-warning method for a man-machine interaction behavior of an underground conveyor belt operator.
To achieve the foregoing technical objective, the present invention adopts the following technical solutions:
A safety monitoring and early-warning method for a man-machine interaction behavior of an underground conveyor belt operator includes the following steps:
(1) acquiring an underground real-time video stream by using a surveillance camera;
(2) estimating the size of a belt in the video according to a camera calibration principle, and accordingly delineating a three-dimensional region of interest (ROT), namely, a belt dangerous region;
(3) detecting human body key-points of all persons in the video, measuring the Date Recue/Date Received 2021-05-26 degree of correlation between the key-points by using part affinity fields, and clustering key-points belonging to each individual by means of matching and optimization in a bipartite graph, thus achieving the objective of detecting human body key-points of each person in the video;
(4) determining x-axis and y-axis components of each detected human body key-point in a world coordinate system, self-defining a height component z for the human body key-point, and then combining the three components into complete world coordinates of the human body key-point; and
(5) determining, according to a relative position relationship between the belt dangerous region and the human body key-points of each person, whether a man-machine interaction behavior is safe, thus determining whether to give an early warning.
Further, in step (3), with each picture in the video as an input, deep features thereof are extracted to obtain a feature graph F, and then the feature graph F is input to step 1 having two convolutional neural network (CNN) branches; in this step, the first CNN
branch predicts confidence maps St=p1(F) regarding a set of key-points, where 1)1 indicates an inference procedure of this CNN branch in step 1; the second CNN
branch predicts a set of "part affinity fields" L'= 01(F), where 01 indicates an inference procedure of this CNN branch in step 1 and aims to cluster the predicted human body key-points on respective limbs of each individual, to obtain a set of complete information of human body key-points; and afterwards, the prediction results obtained by step 1 based on the two CNN branches are linked in series with the original feature graph F, and then are together input to subsequent steps, so as to obtain a more accurate prediction result, where the subsequent steps are denoted by the following formulas:
= ps(F,S",11-1), Vt 2 = ti(1,S",11 1), \ft 2 where St and Lt respectively indicate confidence maps and part affinity fields obtained by step t, and ID' and Ot indicate inference procedures of the two CNN branches in step t respectively.
Further, a loss function, namely, a mean squared error, is applied for the two CNN

Date Recue/Date Received 2020-09-24 branches separately after each step, and the loss functions for the two CNN
branches in step t are respectively as follows:

I: =EEW(P)IIS'1(P) ¨S;(13)12 ,J-1 P
C
fl, =W(p) 112(p)-1:0(0r c=1. p where f; and fLt are respectively the loss functions for the two CNN branches in step t; p is the coordinates of any position in a picture to be detected; W(p) is a Boolean value, which equals 0 when there is no mark in a training dataset, or otherwise equals 1; Stj(p) indicates a confidence map of the jth human body key-point at the position p in step t; S*J(p) indicates a true position of the confidence map; Ltc(p) indicates a part affinity field at the position p in step t, and L*,(p) indicates a true position of the part affinity field;
a true reference for defining confidence levels of key-points at any position p in the picture is as follows:
Slk (0 = exp[ 1P11 -xi,k2 2 ]
a2 where xj,k indicates true coordinates of the jth human body key-point of the kth person in a marked training sample, and a indicates a constant for controlling the degree of dispersion of a Gaussian graph regarding confidence points; and a maximum value is taken, to obtain a confidence reference S1(p) = max.Sik(p) of the jth human body key-point of the kth person.
Further, each person has 9 human body key-points in total, which are the nose, chest, right shoulder, right hand, left shoulder, left hand, hip bone, right foot, and left foot that represent a person.
Further, in step (5), front-view and top-view projections of the belt dangerous region are determined based on the belt dangerous region obtained in step (2);
for each Date Recue/Date Received 2020-09-24 person in the video, a minimum distance dT from the human body key-points to the top-view projection of the belt dangerous region, a minimum distance dF from the human body key-points to the front-view projection of the belt dangerous region, and the height h of the corresponding human body key-point are calculated; and if dT and dF
are both less than or equal to a safe distance threshold d and h is less than the height of the front-view projection of the belt dangerous region, it is determined that a man-machine interaction behavior of the operator is unsafe and an alarm is given.
Further, in step (5), a Deep Neural Network (DNN) classifier is established and used to classify actions according to the detected information of the human body key-points; the key-point information in each picture is combined into a sample which corresponds to one action type; a large number of marked human body key-point -action samples are used to train the classifier, such that the classifier has the ability to identify human actions in each single picture; and safe distance thresholds di corresponding to different actions are determined according to identification results of the classifier, where the subscript i represents the ith action type.
Further, considering continuity of actions of the operator in a surveillance video, a probability judgment model for multiple successive pictures is added based on the action identification for each single picture: by using M successive pictures as a unit for judgment, returning results of classifying actions in the M pictures one by one with the action classifier, and performing counting for different classification results separately;
and finally, calculating a ratio of the counted number in each result to the total number in all the results, and determining the greatest ratio as a classification result of the actions in the M pictures.
Further, there are three action classification results: falling over, squatting down, and smoking; different safety coefficients yi are assigned to the three action types respectively, and their respective safe distance thresholds are calculated as follows: di =
yi = d, where i = 1, 2, 3; and finally, it is determined, according to the safe distance threshold, whether man-machine interaction of the operator for the current action is safe.
Date Recue/Date Received 2020-09-24 Advantageous Effect The foregoing technical solutions achieve the following advantageous effects:
The present invention determines a belt position in a video based on a camera calibration principle, and delineates a three-dimensional ROT based on the belt position and size. By using a "bottom-up" key-point extraction method, multi-person key-point detection of conveyor belt operators is performed by detection followed by clustering, thus guaranteeing the detection accuracy and improving the detection efficiency. The human body key-points and the ROT are separately projected twice: front-view projection and top-view projection; and a relationship between the human body and the belt position is estimated based on the two projections. A DNN classifier is established and used to classify actions according to the key-point information in each single picture, and action labels of each person are returned. By combining action identification and position judgment, the positions of actions with different unsafe coefficients are determined based on different scales. The present invention screens out unsafe actions during man-machine interaction and gives an early warning, so as to avoid a major safety hazard in the belt conveyor system caused by an abnormal contact of the operator with the belt region.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is an overall flowchart of the present invention;
FIG. 2 is a projection diagram of a belt dangerous region from three angles of view;
FIG. 3 is a schematic diagram of camera calibration;
FIG. 4 is a schematic structural diagram of a key-point predication network;
FIG. 5 is a curve chart showing a relationship between key-point coordinates and confidence levels;
FIG. 6 is a schematic diagram showing transformation of belt coordinates and projection;
FIG. 7 is a schematic simplified diagram of human body key-points;
6 Date Recue/Date Received 2020-09-24 FIG. 8 is a schematic diagram of evaluating an unsafe action by a projection method;
FIG. 9 is a schematic diagram of classification of actions of an underground operator;
FIG. 10 is a schematic diagram of safety judgment when the operator falls over;
FIG. 11 is a schematic diagram of safety judgment when the operator squats down;
FIG. 12 is a schematic diagram of safety judgment when the operator smokes;
and FIG. 13 is a schematic diagram of a specific implementation process of the present invention.
DETAILED DESCRIPTION OF THE INVENTION
The technical solutions of the present invention are described in detail below with reference to the accompanying drawings.
A procedure of a safety monitoring and early-warning method for a man-machine interaction behavior of an underground conveyor belt operator that is provided by the present invention is shown in FIG. 1. For a real-time video stream acquired by a surveillance camera, a three-dimensional ROI model is established for a belt position based on a camera calibration principle. Then, key-points of the conveyor belt operator are detected based on a "bottom-up" approach and actions are classified based on key-point information by using a DNN. Finally, the ROI and the key-points are projected in the front-view and top-view directions, and a positional relationship between the key-points and the ROI is evaluated according to safe distance thresholds corresponding to different actions, to determine a dangerous action and give an early warning.
1. Modeling for a belt dangerous region During belt safety identification for early warning, delineating an unsafe ROI
on the belt is a basic task in a detection phase. The present invention identifies a dangerous action by evaluating a positional relationship between human body key-points and the delineated ROI on the belt. Because a positional relationship between the human being and the belt in the vertical direction cannot be evaluated by means of a 2D
ROI, a false
7 Date Recue/Date Received 2020-09-24 alarm rate inevitably increases if the conventional method is used to delineate the 2D
ROT on the belt. For example, when a miner normally works on a platform higher than the belt, it is highly probable that such an operation is evaluated as an unsafe behavior according to the 2D ROT. To solve the problem, the present invention proposes establishing a 3D ROT model according to the belt position, estimating the size of the belt in a video according to a camera calibration and imaging principle, and accordingly delineating a 3D ROT region, where a diagram of the ROT from three angles of view is shown in FIG. 2.
2. Measurement of a belt size by camera calibration (i) A measurement principle of a belt size is as follows: Based on known internal parameters of a monocular camera and coordinates of a picture-image coordinate system in a monocular lens, a relationship between an image coordinate system and a world coordinate system is determined, so as to establish a three-dimensional model showing a relative position of the belt and surrounding operators.
(ii) The image coordinate system is a coordinate system using a pixel as a unit, with the origin being on the upper left side. A position of each pixel point is denoted with pixels, and therefore such a coordinate system is called an image-pixel coordinate system denoted as (u, v), where u and v respectively indicate the column and the row of a pixel point in a digital image.
(iii) The world coordinate system is a three-dimensional coordinate system defined by a user, and used to describe a relative position of an object and the camera in a three-dimensional space, which is denoted as (X, Y, Z).
It can be learned from Figure 3 that the image coordinate system U01P, a camera coordinate system with 02 as the origin, and the world coordinate system X03Y
are on the upper left corner, where the already known quantities are as follows:
The height of the camera is H; a distance between a world coordinate point corresponding to a pixel coordinate center and the camera on the y axis is 03M; image coordinates of the pixel coordinate center 01 are (ucenter, vcenter); a measured point P
is a projection of a point Q to be measured on the Y axis of the world coordinate system,
8 Date Recue/Date Received 2021-05-26 which has pixel coordinates Pi(0, v); an actual pixel length is xpix and an actual pixel width is ypix; and 0102 indicates a focal length f of the camera. A schematic diagram of calibration is shown in FIG. 3.
A coordinate of Y is calculated as follows:
a = &retail _____ H

1= (OLP' .ypix) = (v¨ a:enter) x ypite 13 =
OsP= tan (0) , where y indicates an included angle between 0102 and PiP, and a is an angle between the camera and the horizontal plane and is denoted by an acute angle formed between 0102 and the Y axis. After the angle 13 is obtained by calculation, the coordinate Y=03P in the vertical direction can be calculated according to the properties of a right triangle.
A coordinate of X is calculated as follows:
02P3. = I/Kw vcente r) X xpisi 2 2 aap=
V) = 02P x P1Q1 PQ 02P
PQ
02 PI can be obtained according to PA 02P1, thus obtaining the coordinate X = PQ in the horizontal direction. Therefore, the true coordinates of the point Q are (X, Y).
3. Detection of underground human body key-points Conventional key-point detection algorithms mostly use a "top-down" approach, which first detects all persons in an image to be detected, and then detects key-points of each person separately, failing to realize high-speed detection for a large number of
9 Date Recue/Date Received 2020-09-24 persons. In contrast, the present invention uses a ''bottom-up" structure to first detect key-points of all persons in a video and then cluster key-points belonging to each individual by means of matching and optimization in a bipartite graph, thus achieving the objective of detecting human body key-points of each person in the video. In this way, the detection speed will not decrease with increase in the number of detected persons, realizing multi-person real-time detection of human body key-points. A key-point detection structure is shown in FIG. 4.
With a color RGB image as an input, deep features thereof are extracted via a layer deep neural network ("VGG 19") to obtain a feature graph shown in FIG.
4.
Afterwards, the feature graph is input to step 1 having two CNN branches. In this step, the first network branch predicts confidence maps St=p1(F) regarding a set of key-points, where pl indicates an inference procedure of the first network branch in step 1. The second network branch predicts a set of "part affinity fields" Lt= 01(F), where 0 indicates an inference procedure of the second network branch in step 1 and aims to cluster the predicted key-points on respective limbs of each individual, to obtain a set of complete information of human body key-points. The prediction results obtained by this step based on the two network branches are linked in series with the original feature graph, and then are together input to subsequent steps t as shown in FIG. 4, so as to obtain a more accurate prediction result. The subsequent inference steps t may be denoted by the following formulas:
S = AF,S",11'1), Vt 2 L ev(F.V-1,11-1), "v't 2 where pt and 0 indicate inference procedures of the two CNN branches in step t respectively. According to FIG. 4, t in the subsequent steps t is 2.
In order to guide the network to iteratively predict the confidence maps and the "part affinity fields" of the key-points, a loss function 1,2 (also referred to as a mean squared error) is applied after each step for each branch, which is used to measure an error between a predicted value and a true value. Herein, a spatial weighted value is used to solve the problem that some datasets are unable to mark the key-points of all persons. The loss functions for the CNN branches in step t may be denoted by the Date Recue/Date Received 2021-05-26 following formulas respectively:
J
=Elvv(P) s.,(10¨s*(0112 2 P
EEw(p)- 111,(1)¨ec(l))12 emi p where Si indicates a true position of the key-point confidence map; Lc*
indicates a true position of the "part affinity field"; W is a Boolean value, which equals 0 when there is no mark in a dataset, or otherwise equals 1, and is used to avoid the detection network from punishing the true key-points without any mark.
A true reference for defining confidence levels of key-points at any position p in the picture is as follows:
f ;2 SIktrO=exp ______________ cr2 where p is coordinates of any position in the picture to be detected, k indicates the kth person in the picture, xj,k indicates true coordinates of the jth key-point of the kth person in a marked training sample, and a indicates a constant for controlling the degree of dispersion of a Gaussian graph regarding confidence points. FIG. 5 is a curve chart showing a relationship between key-point coordinates (k, j) and confidence levels.
A maximum value is taken from the foregoing chart, to obtain a confidence reference S(p) = max.SLk(p) of the jth key-point of the kth person.
4. Method for front-view and top-view projections of the key-point coordinates and the ROI
In the "bottom-up" approach for detection of key-points, coordinate information of all the key-points is output in the end. The above-mentioned monocular visual method can calculate x-axis and yw-axis components in the world coordinate system that correspond to a particular pixel coordinate point in the video, which are enough for front-view and top-view projections. However, if it is unable to calculate a z-axis Date Recue/Date Received 2020-09-24 component of a human body key-point, projection of the key-point in the front-view direction cannot be implemented. In addition, because a target pixel in a monocular visual image does not contain depth information capable of reflecting a 3D
relationship, transformation from the image coordinate system to the world coordinate system cannot be realized. To solve such a problem, the present invention simplifies a human body key-point model as follows: According to the already known xw-axis and yw-axis components of the key-points in the world coordinate system, a height component zw is self-defined for each key-point based on the key-point model shown in FIG. 7, and the defined height component and the known xw-axis and yw-axis components are then combined into complete world coordinates (xw, yw, zw) of the human body key-point.
Front-view and top-view projections of a belt dangerous region ROI
corresponding to the key-point coordinates are shown in FIG. 6.
To shorten a system operation time, the human body key-point model is simplified.
FIG. 7(a) shows an original human body key-point model of system prediction, where there are 25 key-points in total. Some key-points in the original model are omitted and only key-points numbered 0, 2, 5, 4, 8, 7, 22 and 19 are kept, thus simplifying the original model to obtain a simplified model shown in FIG. 7(b).
In this model, the zw-axis component of the point 0 is set to 1.6m; the zw-axis components of the points 1, 2, and 5 are all set to 1.3m; the zw-axis components of the points 4, 8, and 7 are set to lm; and the zw-axis components of the points 22 and 19 are set to Om because they are on the same plane as the belt. A projection effect is shown in FIG. 8 where (a) is a top-view projection and (b) is a front-view projection.
For the simplified human body key-point model, if the minimum distances dT and dF from the key-points to the dangerous region ROI respectively in the top-view and front-view projections are both less than or equal to a safe distance threshold d, and h in the front view is less than the height (1.5m) of the belt dangerous region, the system evaluates a current action as an unsafe action and gives an early warning.
5. Method for identifying a dangerous action of an underground operator The above-described method for evaluating an unsafe action based on the positional Date Recue/Date Received 2020-09-24 relationship fails to determine a specific type of the unsafe action, for example, fails to determine that the operator falls over beside the equipment or leans against or sit on the equipment. All these actions have a great safety hazard. Therefore, identification of a specific action of an underground conveyor belt operator is an issue in urgent need to be solved.
The present invention further identifies a specific type of an action in addition to evaluating a dangerous action based on the positional relationship; and sets different safe distance thresholds according to different levels of danger of the actions.
A simple DNN classifier is established, and used to classify actions according to the key-point information acquired in the above. The position information of the key-points in each picture is combined into a sample which corresponds to one action type.
A large number of marked key-point - action samples are used to train the classifier, such that the classifier has the ability to identify human actions in each single picture.
In addition, considering continuity of actions of the operator in a surveillance video, a probability judgment model for multiple successive pictures which are usually highly correlated is added based on the action identification for each single picture.
Specifically, by using five successive pictures as a unit for judgment, results of classifying actions in the five pictures one by one are returned with the action classifier, and counting is performed for different classification results separately.
Finally, a ratio of the counted number in each result to the total number in all the results is calculated, and the greatest ratio is determined as a classification result of the actions in the five pictures. A procedure of classifying the actions of underground operators by the classifier based on the DNN is shown in FIG. 9.
Unsafe actions to be identified include falling over, squatting down, and smoking.
The three action types all have varying degrees of impacts on safety of the conveyor belt operator. Therefore, different safety coefficients are set respectively for the three actions, where yi = 2.0 for falling over, y2 = 1.5 for squatting down, and y3 = 1.3 for smoking. Safe distance thresholds corresponding to the different actions may be obtained by calculation with the formula di = yi*d (i = 1, 2, 3). By combining action identification and position evaluation, the operator is pre-warned about an unsafe Date Recue/Date Received 2020-09-24 distance according to different actions with different safety coefficients, thus realizing advanced warning about dangerous actions and greatly improving reliability of the safety early-warning system. FIGs. 10 to 12 are schematic diagrams of safety judgment for the three actions respectively.
When horizontal distances d1 and dFi between each of the three dangerous actions and the belt are less than respective safety thresholds di, and further a vertical height hi from a key-point horizontally nearest to the belt to the horizontal plane is less than the height of the ROT on the belt, the system deteimines a current action as an unsafe behavior and gives an alarm.
FIG. 13 shows a specific implementation process of the present invention, where (a) is a detection diagram of a belt dangerous region and human body key-points, (b) is a top view of the human body key-points and the belt dangerous region, and (c) is a front view of the human body key-points and the belt dangerous region.
The foregoing embodiment merely describes the technical idea of the present invention, but is not intended to limit the protection scope of the present invention. Any modification made based on the technical solutions according to the technical idea provided by the present invention falls within the protection scope of the present invention.

Date Recue/Date Received 2020-09-24

Claims (8)

What is claimed is:
1. A safety monitoring and early-warning method for a man-machine interaction behavior of an operator of an underground conveyor belt, comprising the following steps:
(1) acquiring an underground real-time video by using a surveillance camera, wherein the underground real-time video comprises video frames;
(2) estimating a size of a portion of the underground conveyor belt shown in the video according to a camera calibration principle, and accordingly delineating a three-dimensional region of interest(ROI) as a belt dangerous region;
(3) detecting human body key-points of all persons appearing in the video, measuring a degree of correlation between the human body key-points by using part affinity fields, and clustering key-points belonging to each person by means of matching and optimization in a bipartite graph, thus achieving an objective of detecting humanbody key-points of each person in the video;
(4) determining x-axis and y-axis components of each of the detected human body key-points in a world coordinate system, self-defining a height component z for each of the detected human body key-points, and then combining the x-axis and y-axis components and the height component z into complete world coordinates of the human body key-point; and (5) determining, according to a positional relationship between the belt dangerous region and the detected human body key-points of each person, whether the man-machine interaction behavior is safe, thus determining whether to give an early warning.
2. The safety monitoring and early-warning method according to claim 1, wherein in step (3), using each of the video frames in the video as an input, deep features thereof are extracted to obtain a feature graph F, and then the feature graph F is input to step 1 having a first and a second convolutional neural network Date Recue/Date Received 2021-05-26 (CNN) branches; wherein the first CNN branch predicts confidence maps expressed as S1=p1(F) regarding a set of the detected human body key-points, wherein p1 indicates an inference procedure of the first CNN branch in step 1;
the second CNN branch predicts a set of part affinity fields expressed as L1=
01(F), wherein 01 indicates an inference procedure of the second CNN branch in step 1 and aims to cluster predicted human body key-points on respective limbs of each person, to obtain a set of complete information of each person's human body key-points; and afterwards, prediction results obtained by stepl based on the first and the second CNN branches are linked in series with the feature graph Fõ which are then together input to subsequent steps t, so as to obtain further prediction results, wherein the further prediction results from each of the subsequent steps t are denoted by the following formulas:
wherein t is > 2, each of the subsequent steps t comprises two CNN branches to predict confidence maps expressed as St and part affinity fields expressed as Lt respectively, and pt and 0 indicate inference procedures of the two CNN
branches in each of the subsequent steps t respectively.
3. The safety monitoring and early-warning method according to claim 2, wherein a loss function, identified as a mean squared error, is applied for the two CNN branches separately after each of the subsequent steps t, and the loss functions for the two CNN branches in each of the subsequent steps t are respectively as follows:
Date Recue/Date Received 2021-05-26 wherein fis. and fir, are respectively the loss functions for the two CNN
branches in each of the subsequent steps t; p is coordinates of a position p in a video frame to be detected;W(p) is a Boolean value, which equals 0 when there is no mark in a training dataset, or otherwise equals 1; Stj(p) indicates a confidence map of a jth human body key-point of a kth person at the position p in the subsequent step t; s*j(p) indicates a true position of the confidence map;
Ltc(p) indicates a part affinity field at the position p in the subsequent step t, and ec(p) indicates a true position of the part affinity field; a true reference for defining confidence levels of key-points at any position p in the video frame is as follows:
wherein xj,k indicates true coordinates of the jth human body key-point of the kth person in a marked training sample, and a indicates a constant for controlling a degree of dispersion of a Gaussian graph regarding confidence points; and a maximum value is taken, to obtain a confidence reference Sj(p) = maxY; (p) of the jth human body key-point of the kth person.
4. The safety monitoring and early-warning method according to claim 1, wherein each person has 9 human body key-points in total, which are each person's nose, chest, right shoulder, right hand, left shoulder, left hand, hip bone, rightfoot, and left foot.
5. The safety monitoring and early-warning method according to claim 1, wherein in step (5), front-view and top-view projections of the belt dangerous region are determined based on the belt dangerous region obtained in step (2);
for each person in the video, a minimum distance & from the human body key-points to the top-view projection of the belt dangerous region, a minimum distance dF

Date Recue/Date Received 2021-05-26 from the human body key-points to the front-view projection of the belt dangerous region, and the height h of the corresponding human body key-point are calculated; and if dT and dF are both less than or equal to a safe distance threshold d and h is less than the height of the front-view projection of the belt dangerous region, it is determined that a man-machine interaction behavior of the operator is unsafe and an alarm is given.
6. The safety monitoring and early-warning method according to claim 5, wherein in step (5), a Deep Neural Network (DNN) classifier is established and used to classify actions according to the detected information of the human body key-points; position information of the human body key-points in each of the video frames is combined into a sample which corresponds to one action type;a large number of marked human body key-point - action samples are used to train the classifier, such that the classifier identifies human actions in each of the video frames; and safe distance thresholds di corresponding to different actions are determined according to identification results of the classifier, wherein i in the safe distance thresholds di represents an ith action type.
7. The safety monitoring and early-warning method according to claim 6, wherein considering continuity of actions of the operator in the video, a probability judgment model for multiple successive video frames is added based on the action identification for each single video frame: by using M
successive video frames as a unit for judgment, returning results of classifying actions in the M successive video frames one by one with the classifier, and performing counting for different classification results separately; and finally, calculating a ratio of the counted number in each of the classification results to a total number in all the classification results, and determining the greatest ratio as a classification result of the actions in the M successive video frames.
8. The safety monitoring and early-warning method according to claim 6, wherein there are three action classification results: falling over, squatting down, Date Recue/Date Received 2021-05-26 and smoking; different safety coefficients yi are assigned to the three action types respectively, and their respective safe distance thresholds are calculated as follows: di = yi = d, wherein i = 1, 2, 3; and finally, it is determined, according to the safe distance threshold, whether man-machine interaction of the operator for a current action is safe.

Date Recue/Date Received 2021-05-26
CA3094424A 2019-06-21 2020-03-30 Safety monitoring and early-warning method for man-machine interaction behavior of underground conveyor belt operator Active CA3094424C (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
CN201910540349.7A CN110425005B (en) 2019-06-21 2019-06-21 Safety monitoring and early warning method for man-machine interaction behavior of belt transport personnel under mine
CN2019105403497 2019-06-21
PCT/CN2020/082006 WO2020253308A1 (en) 2019-06-21 2020-03-30 Human-machine interaction behavior security monitoring and forewarning method for underground belt transportation-related personnel

Publications (2)

Publication Number Publication Date
CA3094424A1 CA3094424A1 (en) 2020-12-21
CA3094424C true CA3094424C (en) 2022-02-22

Family

ID=74036685

Family Applications (1)

Application Number Title Priority Date Filing Date
CA3094424A Active CA3094424C (en) 2019-06-21 2020-03-30 Safety monitoring and early-warning method for man-machine interaction behavior of underground conveyor belt operator

Country Status (1)

Country Link
CA (1) CA3094424C (en)

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113155070B (en) * 2021-03-25 2023-03-24 深圳英飞拓科技股份有限公司 Scaffold climbing safety detection method and system
CN113111840A (en) * 2021-04-26 2021-07-13 常州典宏自动化科技有限公司 Method for early warning violation and dangerous behaviors of operators on fully mechanized coal mining face
CN113202560A (en) * 2021-06-01 2021-08-03 桂林慧谷人工智能产业技术研究院 Coal mine risk prevention and rescue decision-making system based on data mining
CN113420626A (en) * 2021-06-15 2021-09-21 福建华电可门发电有限公司 Construction site safety behavior judging method and storage device
CN113313428A (en) * 2021-06-29 2021-08-27 上海煤科信息科技有限公司 AI fault monitoring method and system of smart mine based on big data
CN113449675B (en) * 2021-07-12 2024-03-29 西安科技大学 Method for detecting crossing of coal mine personnel
CN114048439B (en) * 2021-10-09 2022-07-26 大庆恒驰电气有限公司 AI-based safety behavior analysis system and method
CN114494427B (en) * 2021-12-17 2024-09-03 山东鲁软数字科技有限公司 Method, system and terminal for detecting illegal behaviors of person with suspension arm going off station
CN115034280B (en) * 2022-03-16 2023-07-25 宁夏广天夏科技股份有限公司 System for detecting unsafe behaviors of underground personnel
CN114758457B (en) * 2022-04-19 2024-02-02 南京奥拓电子科技有限公司 Intelligent monitoring method and device for illegal operation among banknote adding
CN115412862B (en) * 2022-08-04 2024-04-30 广州市明道文化产业发展有限公司 Multi-role decentralization scenario interaction method and device based on LBS and storage medium
CN115375151B (en) * 2022-08-25 2023-07-11 合肥未来计算机技术开发有限公司 Safety scheduling method for operators in underground construction
CN115597659B (en) * 2022-09-21 2023-04-14 山东锐翊电力工程有限公司 Intelligent safety management and control method for transformer substation
CN116311534B (en) * 2023-05-17 2023-12-22 青岛润邦泽业信息技术有限公司 Intelligent worksite dangerous violation action recognition method and system based on AI
CN116895043B (en) * 2023-06-13 2024-01-26 郑州宝冶钢结构有限公司 Intelligent safety monitoring and early warning method, system and storage medium for construction site

Also Published As

Publication number Publication date
CA3094424A1 (en) 2020-12-21

Similar Documents

Publication Publication Date Title
CA3094424C (en) Safety monitoring and early-warning method for man-machine interaction behavior of underground conveyor belt operator
CN110425005B (en) Safety monitoring and early warning method for man-machine interaction behavior of belt transport personnel under mine
CN109076198B (en) Video-based object tracking occlusion detection system, method and equipment
CN105574501B (en) A kind of stream of people's video detecting analysis system
JP5603403B2 (en) Object counting method, object counting apparatus, and object counting program
US8965050B2 (en) Behavior analysis device
CA2692424C (en) System and process for detecting, tracking and counting human objects of interest
EP2608536B1 (en) Method for counting objects and apparatus using a plurality of sensors
EP3518146B1 (en) Image processing apparatus and image processing method
US9904861B2 (en) Method for detecting target objects in a surveillance region
CN110188724A (en) The method and system of safety cap positioning and color identification based on deep learning
US20120020518A1 (en) Person tracking device and person tracking program
Huang et al. Automatic moving object extraction through a real-world variable-bandwidth network for traffic monitoring systems
US20190294890A1 (en) Method and apparatus for surveillance
CN107657244B (en) Human body falling behavior detection system based on multiple cameras and detection method thereof
Boltes et al. T-junction: Experiments, trajectory collection, and analysis
JP2016099941A (en) System and program for estimating position of object
CN103735269B (en) A kind of height measurement method followed the tracks of based on video multi-target
EP3070676A1 (en) A system and a method for estimation of motion
CN104599287A (en) Object tracking method and device and object recognition method and device
CN104680145B (en) The on off state change detecting method and device of a kind of
CN110136172B (en) Detection method for wearing of underground protective equipment of miners
CN110969131A (en) Subway people flow counting method based on scene flow
CN111832450B (en) Knife holding detection method based on image recognition
CN112200828B (en) Method and device for detecting ticket evasion behavior and readable storage medium