CN117079416B - Multi-person 5D radar falling detection method and system based on artificial intelligence algorithm - Google Patents

Multi-person 5D radar falling detection method and system based on artificial intelligence algorithm Download PDF

Info

Publication number
CN117079416B
CN117079416B CN202311335592.8A CN202311335592A CN117079416B CN 117079416 B CN117079416 B CN 117079416B CN 202311335592 A CN202311335592 A CN 202311335592A CN 117079416 B CN117079416 B CN 117079416B
Authority
CN
China
Prior art keywords
target
layer
model
representing
loss
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202311335592.8A
Other languages
Chinese (zh)
Other versions
CN117079416A (en
Inventor
杨绍分
袁文忠
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Dexin Intelligent Technology Changzhou Co ltd
Original Assignee
Dexin Intelligent Technology Changzhou Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Dexin Intelligent Technology Changzhou Co ltd filed Critical Dexin Intelligent Technology Changzhou Co ltd
Priority to CN202311335592.8A priority Critical patent/CN117079416B/en
Publication of CN117079416A publication Critical patent/CN117079416A/en
Application granted granted Critical
Publication of CN117079416B publication Critical patent/CN117079416B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B21/00Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
    • G08B21/02Alarms for ensuring the safety of persons
    • G08B21/04Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons
    • G08B21/0407Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons based on behaviour analysis
    • G08B21/043Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons based on behaviour analysis detecting an emergency event, e.g. a fall
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/766Arrangements for image or video recognition or understanding using pattern recognition or machine learning using regression, e.g. by projecting features on hyperplanes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B21/00Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
    • G08B21/02Alarms for ensuring the safety of persons
    • G08B21/04Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons
    • G08B21/0407Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons based on behaviour analysis
    • G08B21/0423Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons based on behaviour analysis detecting deviation from an expected pattern of behaviour or schedule
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B21/00Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
    • G08B21/02Alarms for ensuring the safety of persons
    • G08B21/04Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons
    • G08B21/0438Sensor means for detecting
    • G08B21/0492Sensor dual technology, i.e. two or more technologies collaborate to extract unsafe condition, e.g. video tracking and RFID tracking
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B31/00Predictive alarm systems characterised by extrapolation or other computation using updated historic data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2123/00Data types
    • G06F2123/02Data types in the time domain, e.g. time-series data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/08Feature extraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/12Classification; Matching
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Mathematical Physics (AREA)
  • Multimedia (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Biomedical Technology (AREA)
  • Emergency Management (AREA)
  • Business, Economics & Management (AREA)
  • Social Psychology (AREA)
  • Gerontology & Geriatric Medicine (AREA)
  • Psychiatry (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Psychology (AREA)
  • Probability & Statistics with Applications (AREA)
  • Medical Informatics (AREA)
  • Databases & Information Systems (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides a multi-person 5D radar falling detection method and system based on an artificial intelligence algorithm, and relates to a falling detection technology, wherein the method comprises the steps of acquiring a plurality of radar echo information of a plurality of target objects in a target area based on a millimeter wave radar, and extracting a plurality of motion features corresponding to the plurality of radar echo information through a preset echo feature extraction model; acquiring video information of a plurality of target objects in the target area based on an image sensor, determining target areas corresponding to the plurality of target objects in each frame of video information based on a target detector, matching the target areas of each target object with motion characteristics, and determining comprehensive characteristics corresponding to each target object; and determining the action type of the target object through a preset action classification model according to the comprehensive characteristics corresponding to each target object, and sending out alarm information when the action type is matched with the falling action.

Description

Multi-person 5D radar falling detection method and system based on artificial intelligence algorithm
Technical Field
The invention relates to a fall detection technology, in particular to a method and a system for detecting fall of a multi-person 5D radar based on an artificial intelligence algorithm.
Background
At present, research on fall risk assessment of the elderly focuses on gait analysis and posture detection, and a large number of risk assessment models are obtained. However, most risk assessment models are constructed without daily life data in a real scene, so that the physical state of the old cannot be effectively assessed, and the risk assessment models are not effective. Furthermore, most studies use a single device to perceive the physical state of the elderly, and noise and occlusion problems are difficulties that cannot be overcome by a single perception assessment model. Meanwhile, the invasiveness of the sensing equipment brings the problems of subconscious resistance and data credibility to the old.
Disclosure of Invention
The embodiment of the invention provides a multi-person 5D radar falling detection method and system based on an artificial intelligence algorithm, which at least can solve part of problems in the prior art.
In a first aspect of an embodiment of the present invention,
the utility model provides a multi-person 5D radar fall detection method based on artificial intelligence algorithm, which comprises the following steps:
acquiring multiple radar echo information of multiple target objects in a target area based on a millimeter wave radar, and extracting multiple motion features corresponding to the multiple radar echo information through a preset echo feature extraction model, wherein the echo feature extraction model transmits the output of each layer to the next layer through the connection coefficient of the adjacent layer and combining a nonlinear compression function;
acquiring video information of a plurality of target objects in the target area based on an image sensor, determining target areas corresponding to the plurality of target objects in each frame of video information based on a target detector, matching the target areas of each target object with motion characteristics, and determining comprehensive characteristics corresponding to each target object;
and determining the action type of the target object through a preset action classification model according to the comprehensive characteristics corresponding to each target object, and sending out alarm information when the action type is matched with the falling action, wherein the action classification model is constructed based on a combined network model, and the action classification is carried out by distributing corresponding weights for different network models.
In an alternative embodiment of the present invention,
acquiring a plurality of radar echo information of a plurality of target objects in a target area based on a millimeter wave radar, and extracting a plurality of motion features corresponding to the plurality of radar echo information through a preset echo feature extraction model comprises:
multiplying the radar echo information with a weight matrix corresponding to each layer in the echo feature extraction model to obtain low-level features;
according to the low-level characteristics, combining the connection coefficients of the current layer and the adjacent next layer to determine the high-level characteristics of the adjacent next layer;
and based on the high-level features, carrying out space dimension mapping on the high-level features by combining a nonlinear compression function of the echo feature extraction model, and determining a plurality of motion features corresponding to the radar echo information.
In an alternative embodiment of the present invention,
determining a plurality of motion features corresponding to the plurality of radar echo information includes:
a plurality of motion characteristics are determined according to the following formula:
wherein I is l 、I h Representing low-level features and high-level features, respectively, C lh Representing the connection coefficient of the current layer with the next adjacent layer, W l Representing a weight matrix, b l Representing the bias parameter, P l 、Q h Respectively represent a low-level learnable matrix parameter and a high-level learnable matrix parameter, S h Representing motion characteristics, f () represents a nonlinear compression function, e, r representing nonlinearity and sensitivity, respectively.
In an alternative embodiment of the present invention,
before determining the target areas corresponding to the plurality of target objects in each frame of video information based on the target detector, the method further comprises training the target detector:
the target detector comprises a convolution layer, a candidate layer, a pooling layer and a full connection layer;
converting a training data set into a feature map through a convolution layer based on the pre-acquired training data set; generating a plurality of anchor frames with different sizes and aspect ratios in each sliding window by the candidate layer, and screening out the anchor frames with the front scores from the plurality of anchor frames as candidate frames by applying non-maximum value inhibition;
the pooling layer maps the candidate frames to the feature images with the same size as the candidate frames to carry out pooling operation and generate feature vectors;
and determining classification loss and boundary box regression loss through a full connection layer according to the feature vector, and iteratively optimizing a classification loss function and a boundary box regression loss function through a back propagation algorithm in combination with an adaptive learning rate until the classification loss and the boundary box regression loss are minimized.
In an alternative embodiment of the present invention,
determining the classification loss and the boundary box regression loss through the full connection layer according to the feature vector, and iteratively optimizing the classification loss function and the boundary box regression loss function through a back propagation algorithm in combination with the self-adaptive learning rate comprises the following steps:
wherein L represents the sum of the classification loss value and the regression loss value, S represents the learning rate, r represents the loss weight coefficient, g t Represents the model gradient at time t, θ represents the model parameters of the target detection model, m t A first moment estimation representing a time t;
L cls represents the classification loss corresponding to the classification loss function, N represents the sample number of the training data set, y i 、p i Respectively representing the actual label corresponding to the ith sample of the training data set and the label prediction probability corresponding to the ith sample of the training data set,
L reg representing regression loss corresponding to the boundary box regression loss function, T i Representing the actual regression objective corresponding to the ith sample of the training dataset, smooth () represents the smoothing loss function.
In an alternative embodiment of the present invention,
according to the comprehensive characteristics corresponding to each target object, determining the action type of the target object through a preset action classification model comprises the following steps:
the action classification model comprises a space sub-model and a time sub-model, wherein the space sub-model is constructed based on a convolutional neural network, the time sub-model is constructed based on a cyclic neural network,
taking the output of the space sub-model as the input of the time sub-model, applying global average pooling operation to the output of the time sub-model, distributing corresponding time weights to the feature vectors after the global average pooling operation through an adaptive weight distribution algorithm, and inputting the feature vectors after the global average pooling operation and the corresponding time weights into a full connection layer of the time sub-model;
and determining the probability distribution of the action category corresponding to the comprehensive characteristics through the classification activation function of the full connection layer, and taking the highest probability value in the probability distribution of each action category as the final action type.
In an alternative embodiment of the present invention,
the method for distributing the corresponding time weight to the feature vector after the global average pooling operation through the self-adaptive weight distribution algorithm comprises the following steps:
wherein S is T The time-weight is represented by a time-weight,representing the sigmod function and the ReLu function respectively,the number of feature vectors is represented,and the time weight of the feature vector after the k-1 global average pooling operation is represented, and U represents the feature vector after the global average pooling operation.
In a second aspect of an embodiment of the present invention,
provided is an artificial intelligence algorithm-based multi-person 5D radar fall detection system, comprising:
the device comprises a first unit, a second unit and a third unit, wherein the first unit is used for acquiring a plurality of radar echo information of a plurality of target objects in a target area based on a millimeter wave radar, extracting a plurality of motion characteristics corresponding to the plurality of radar echo information through a preset echo characteristic extraction model, and the echo characteristic extraction model transmits the output of each layer to the next layer through the connection coefficient of the adjacent layer and combining a nonlinear compression function;
the second unit is used for acquiring video information of a plurality of target objects in the target area based on the image sensor, determining the target area corresponding to the plurality of target objects in each frame of video information based on the target detector, matching the target area of each target object with the motion characteristics, and determining the comprehensive characteristics corresponding to each target object;
and the third unit is used for determining the action type of the target object through a preset action classification model according to the comprehensive characteristics corresponding to each target object, and sending out alarm information when the action type is matched with the falling action, wherein the action classification model is constructed based on a combined network model, and the action classification is carried out by distributing corresponding weights for different network models.
In a third aspect of an embodiment of the present invention,
there is provided an electronic device including:
a processor;
a memory for storing processor-executable instructions;
wherein the processor is configured to invoke the instructions stored in the memory to perform the method described previously.
In a fourth aspect of an embodiment of the present invention,
there is provided a computer readable storage medium having stored thereon computer program instructions which, when executed by a processor, implement the method as described above.
The beneficial effects of the embodiments of the present invention may refer to the effects corresponding to technical features in the specific embodiments, and are not described herein.
Drawings
Fig. 1 is a schematic flow chart of a multi-person 5D radar fall detection method based on an artificial intelligence algorithm according to an embodiment of the present invention;
fig. 2 is a schematic structural diagram of a multi-person 5D radar fall detection system based on an artificial intelligence algorithm according to an embodiment of the present invention.
Detailed Description
For the purpose of making the objects, technical solutions and advantages of the embodiments of the present invention more apparent, the technical solutions of the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention, and it is apparent that the described embodiments are only some embodiments of the present invention, not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
The technical scheme of the invention is described in detail below by specific examples. The following embodiments may be combined with each other, and some embodiments may not be repeated for the same or similar concepts or processes.
Fig. 1 is a flow chart of a multi-person 5D radar fall detection method based on an artificial intelligence algorithm according to an embodiment of the present invention, as shown in fig. 1, the method includes:
s101, acquiring multiple radar echo information of multiple target objects in a target area based on millimeter wave radar, and extracting multiple motion features corresponding to the multiple radar echo information through a preset echo feature extraction model, wherein the echo feature extraction model transmits the output of each layer to the next layer through the connection coefficient of the adjacent layer and combining a nonlinear compression function;
for example, when the millimeter wave radar emits a signal and reflects the signal by the target object, the received signal is radar echo information, and the information contains the position, speed and other characteristics of the target object, which is basic data for analyzing the movement of the target object.
Optionally, the signals received by the millimeter wave radar may be preprocessed, including denoising, signal amplification, and time/frequency domain conversion, to design a deep neural network structure that includes multiple layers, each layer receiving the output of the previous layer and passing the features to the next layer through a connection coefficient (which may be understood as a weight in the network) and a nonlinear compression function (such as a ReLU activation function). A nonlinear compression function is applied after the output of each layer, which may be a ReLU (Rectified Linear Unit) function, or other activation function, to introduce nonlinear properties that enable the model to learn the nonlinear relationship.
The connection coefficients are weight parameters learned during the network training process for adjusting the weights of each layer output, and are optimized during the network training process to enable the model to automatically learn and extract meaningful features. At the last layer of the network, an output layer may be designed for outputting final motion characteristics, which may include information on the speed, acceleration, etc. of the target object.
The millimeter wave radar can provide high-precision target position and speed information, and can extract fine motion characteristics of a plurality of target objects by combining an echo characteristic extraction model, and the characteristics can describe the motion state of the target objects more accurately. By extracting the motion characteristics of multiple target objects, the system can achieve multi-target tracking and analysis, which is important for monitoring the motion behavior of multiple people or multiple objects. The extracted motion features can be used in decision making by an intelligent system, for example in fall detection, where the system can automatically trigger an alarm when abnormal motion features are detected (e.g. rapid descent speed), providing emergency assistance or supervision.
In an alternative embodiment of the present invention,
acquiring a plurality of radar echo information of a plurality of target objects in a target area based on a millimeter wave radar, and extracting a plurality of motion features corresponding to the plurality of radar echo information through a preset echo feature extraction model comprises:
multiplying the radar echo information with a weight matrix corresponding to each layer in the echo feature extraction model to obtain low-level features;
according to the low-level characteristics, combining the connection coefficients of the current layer and the adjacent next layer to determine the high-level characteristics of the adjacent next layer;
and based on the high-level features, carrying out space dimension mapping on the high-level features by combining a nonlinear compression function of the echo feature extraction model, and determining a plurality of motion features corresponding to the radar echo information.
Illustratively, in a neural network, there are connections between neurons of each layer and neurons of the previous layer, which connections correspond to weights, i.e., a matrix representation of the weights. The low-level features are used to indicate that at an early level of the echo feature extraction model, the original features learned by the feature extraction model are typically directly related to the input data (radar echo information). The high-level features are used for indicating abstract features learned by the feature extraction model at the later level of the echo feature extraction model, and the features are combined and extracted from the low-level features and can better represent the motion characteristics of a target object.
Optionally, inputting radar echo information to an input layer of the network, and sequentially calculating the output of each layer through a weight matrix and a nonlinear compression function of each layer; at the early level of the neural network, obtaining low-level features by multiplying input data with a corresponding weight matrix and applying a nonlinear compression function; and each radar echo information is used as one dimension of the vector, nonlinear compression function (such as ReLU activation function) is applied to carry out nonlinear mapping on each element of the output of each layer, an activation value of the layer is obtained, the activation value is used as the input of the next layer, and forward propagation is continued until the final layer of the model is reached.
And in the later level of the neural network, combining the connection coefficients of the current layer and the adjacent next layer, and obtaining high-level characteristics through corresponding nonlinear compression functions. The method comprises the steps of calculating low-level characteristics to obtain an output vector of a current layer, and multiplying the output of the current layer by a connection coefficient matrix of an adjacent next layer to obtain an input vector of the adjacent next layer. The input vector is multiplied by the weight matrix of the next adjacent layer, and then a nonlinear compression function (e.g., reLU) is applied to obtain the output vector of the next adjacent layer. If more layers exist in the network, repeating the steps until the final layer of the network is reached.
The network is trained using appropriate loss functions (e.g., mean square error, cross entropy, etc.) and optimization algorithms (e.g., gradient descent, adam, etc.), optimizing the weight matrix and connection coefficients so that the network can learn the motion characteristics of the target object.
In the forward propagation of the neural network, the output of the last hidden layer or the penultimate hidden layer is obtained through multiple layers of computation, and the output is a high-level characteristic and represents the abstract characteristic learned by the network. And inputting the high-level features into a nonlinear compression function to obtain mapped features, wherein the mapped features are final motion features.
In an alternative embodiment of the present invention,
determining a plurality of motion features corresponding to the plurality of radar echo information includes:
a plurality of motion characteristics are determined according to the following formula:
wherein I is l 、I h Representing low-level features and high-level features, respectively, C lh Representing the connection coefficient of the current layer with the next adjacent layer, W l Representing a weight matrix, b l Representing the bias parameter, P l 、Q h Respectively represent a low-level learnable matrix parameter and a high-level learnable matrix parameter, S h Representing motion characteristics, f () represents a nonlinear compression function, e, r representing nonlinearity and sensitivity, respectively.
By multiplying the radar echo information with a weight matrix, low-level features can be extracted, which typically include basic motion information of the target, such as position and velocity, which enables the system to accurately describe the basic motion state of the target. In combination with the low-level features, the connection coefficients and the weight matrix, the system can determine the high-level features of the next layer adjacent to the low-level features, so that the system can extract high-level motion features of a plurality of targets from a plurality of radar echo information, including complex motion patterns and behaviors. By applying a nonlinear compression function, the system can introduce nonlinear factors on high-level features, which enables the system to learn more complex motion features, and nonlinear mapping increases the rich expressive power of the model on target motion behaviors.
S102, acquiring video information of a plurality of target objects in the target area based on an image sensor, determining target areas corresponding to the plurality of target objects in each frame of video information based on a target detector, matching the target areas of each target object with motion characteristics, and determining comprehensive characteristics corresponding to each target object;
illustratively, the target region information and motion characteristics of the target are combined to form a comprehensive feature vector for more fully describing the target object. Capturing a video frame sequence of a target area by using an image sensor, processing each frame of video by using a target detector, determining the position of each target object in an image to obtain a boundary box of the target area (ROI), and extracting the motion characteristics of each target object in the video sequence by using a motion analysis technology (such as an optical flow method or a frame difference-based method) including speed, direction and the like; the target area information (coordinates of the bounding box) and the motion features (speed, direction, etc.) are combined into a comprehensive feature vector, and the coordinates of the bounding box and the values of the motion features can be spliced to form a feature vector containing spatial and temporal information.
In an alternative embodiment of the present invention,
before determining the target areas corresponding to the plurality of target objects in each frame of video information based on the target detector, the method further comprises training the target detector:
the target detector comprises a convolution layer, a candidate layer, a pooling layer and a full connection layer;
converting a training data set into a feature map through a convolution layer based on the pre-acquired training data set; generating a plurality of anchor frames with different sizes and aspect ratios in each sliding window by the candidate layer, and screening out the anchor frames with the front scores from the plurality of anchor frames as candidate frames by applying non-maximum value inhibition;
the pooling layer maps the candidate frames to the feature images with the same size as the candidate frames to carry out pooling operation and generate feature vectors;
and determining classification loss and boundary box regression loss through a full connection layer according to the feature vector, and iteratively optimizing a classification loss function and a boundary box regression loss function through a back propagation algorithm in combination with an adaptive learning rate until the classification loss and the boundary box regression loss are minimized.
Illustratively, the image is converted into feature maps using a convolution layer, the feature maps containing semantic information of the image; a plurality of anchor boxes of different sizes and aspect ratios are generated at each location on the feature map as candidate boxes for target detection.
Mapping the candidate frames to the feature map, and carrying out pooling operation to generate feature vectors; classifying and carrying out bounding box regression by using the full connection layer according to the feature vector, and calculating classification loss and bounding box regression loss; using a back propagation algorithm, in combination with an adaptive learning rate (e.g., adam optimizer), the classification loss function and the bounding box regression loss function are iteratively optimized until the loss is minimized. NMS (non-maximum suppression) is applied to the predicted result, candidate frames with the front scores are screened out, and the candidate frames with larger overlap are removed, so that the final target detection result is obtained.
Optionally, multiple anchor frames of different sizes and aspect ratios are generated for each pixel point, sliding windows are used on the feature map, the generated anchor frames are applied to each window position, each candidate frame is mapped onto the feature map, and the candidate frames of different sizes are mapped into feature vectors of the same size through an ROI pooling operation.
The input is a Feature Map (Feature Map) and coordinate information (typically, upper left and lower right coordinates) from candidate boxes, and each candidate box is mapped to the Feature Map. The corresponding region is defined on the feature map according to the size of the candidate frame, and a method is generally adopted in which the region is divided into grids with fixed sizes, and then the feature value in each grid is subjected to a pooling operation (such as maximum pooling) to obtain a value. For different size candidate boxes, mapping them onto feature maps of the same size, a fixed output grid size is typically used to maintain consistency of the output features, and after pooling, each candidate box generates a feature vector of fixed length that will be used for subsequent classification and regression tasks.
In an alternative embodiment of the present invention,
determining the classification loss and the boundary box regression loss through the full connection layer according to the feature vector, and iteratively optimizing the classification loss function and the boundary box regression loss function through a back propagation algorithm in combination with the self-adaptive learning rate comprises the following steps:
wherein L represents the sum of the classification loss value and the regression loss value, S represents the learning rate, r represents the loss weight coefficient, g t Represents the model gradient at time t, θ represents the model parameters of the target detection model, m t A first moment estimation representing a time t;
L cls represents the classification loss corresponding to the classification loss function, N represents the sample number of the training data set, y i 、p i Respectively representing the actual label corresponding to the ith sample of the training data set and the label prediction probability corresponding to the ith sample of the training data set,
L reg representing regression loss corresponding to the boundary box regression loss function, T i Representing the actual regression objective corresponding to the ith sample of the training dataset, smooth () represents the smoothing loss function.
The target detector obtained through training can accurately position a plurality of target objects in each frame of video information and identify the categories of the target objects; by means of the anchor frames with multiple sizes and aspect ratios generated by the candidate layers, effective detection of targets with different sizes and shapes is achieved, and robustness of target detection is improved; the anchor frames with the front scores are screened out by applying a non-maximum suppression (NMS) algorithm, redundant detection frames are removed, each target is ensured to be detected only once, and the detection efficiency is improved; the pooling layer maps the candidate boxes to feature maps of a fixed size and generates feature vectors through a pooling operation. The method retains the characteristic information of the targets, and maps targets with different sizes into the characteristic vectors with the same size, so that the subsequent processing is facilitated. The classification loss and the bounding box regression loss are calculated through the full connection layer, and the loss functions are iteratively optimized by combining a back propagation algorithm and an adaptive learning rate, so that the classification and the bounding box prediction are more accurate; the whole flow realizes high-efficiency target detection, can rapidly and accurately position a plurality of target objects in a video frame, and provides a reliable target area for subsequent falling detection.
S103, determining the action type of the target object through a preset action classification model according to the comprehensive characteristics corresponding to each target object, and sending out alarm information when the action type is matched with the falling action.
The action classification model is constructed based on a combined network model, and action classification is performed by distributing corresponding weights for different network models. The action classification model comprises a space sub-model and a time sub-model, wherein the space sub-model is constructed based on a convolutional neural network, the time sub-model is constructed based on a cyclic neural network, the space sub-model of the embodiment of the application can comprise the convolutional neural network model, and the time sub-model can comprise the cyclic neural network model.
In an alternative embodiment of the present invention,
according to the comprehensive characteristics corresponding to each target object, determining the action type of the target object through a preset action classification model comprises the following steps:
the action classification model comprises a space sub-model and a time sub-model, wherein the space sub-model is constructed based on a convolutional neural network, the time sub-model is constructed based on a cyclic neural network,
taking the output of the space sub-model as the input of the time sub-model, applying global average pooling operation to the output of the time sub-model, distributing corresponding time weights to the feature vectors after the global average pooling operation through an adaptive weight distribution algorithm, and inputting the feature vectors after the global average pooling operation and the corresponding time weights into a full connection layer of the time sub-model;
and determining the probability distribution of the action category corresponding to the comprehensive characteristics through the classification activation function of the full connection layer, and taking the highest probability value in the probability distribution of each action category as the final action type.
Illustratively, the output of the spatial sub-model is taken as the input of the temporal sub-model, which may be achieved by directly connecting the output of the spatial sub-model to the input layer of the temporal sub-model; the output of the time submodel (usually a time sequence feature sequence) is subjected to global average pooling operation, and the time sequence features are compressed into a feature vector with fixed length; and applying an adaptive weight distribution algorithm to the feature vectors subjected to the global averaging pooling operation, wherein the algorithm can distribute a weight for each time step according to the importance of the features, and the weight can be learned according to an optimization algorithm such as gradient descent and the like so as to ensure that the weight is adaptively adjusted in the training process.
The feature vector and the time weight after the global averaging pooling operation are connected into a larger feature vector, and then the feature vector is input into the full connection layer of the time submodel. Linear transformation and nonlinear activation are performed at the fully connected layer to capture complex relationships between features, and the probability distribution of the corresponding action class is obtained through a classification activation function (e.g., softmax function) of the fully connected layer.
The integrated feature vector is input into a fully connected layer, the number of neurons of which is equal to the number of action classes, and then the output of the fully connected layer is converted into a probability distribution of action classes using a class activation function (e.g., softmax function). From the probability distribution of each action category, the final action type with the highest probability value is selected.
Through the fully connected layer, the network can learn the complex relation in the comprehensive characteristics, the quantity of neurons of the layer is matched with the quantity of action categories, and each neuron is ensured to correspond to one action category. The Softmax function converts the output of the fully connected layer into a probability distribution, ensuring that the probability values for all action classes are between 0 and 1 and the sum is 1. From the probability distribution of the Softmax output, the action category with the highest probability value is selected as the final prediction result, because the Softmax function converts the output into the probability distribution, and the most probable action type can be determined by selecting the category with the highest probability.
In an alternative embodiment of the present invention,
the method for distributing the corresponding time weight to the feature vector after the global average pooling operation through the self-adaptive weight distribution algorithm comprises the following steps:
wherein S is T The time-weight is represented by a time-weight,representing the sigmod function and the ReLu function respectively,the number of feature vectors is represented,and the time weight of the feature vector after the k-1 global average pooling operation is represented, and U represents the feature vector after the global average pooling operation.
Through the space sub-model and the time sub-model, the system can extract rich space and time sequence characteristics from a target area of a target object; through the self-adaptive weight distribution algorithm, the system can distribute different weights for different parts of the time sequence characteristics, so that key moments of actions can be focused more, and the accuracy of action recognition is improved. Through the processing of the full connection layer, the comprehensive features can be mapped to different action categories, and the features are mapped to probability distributions of the action categories through a classification activation function (usually a Softmax function). Because convolutional neural networks and recurrent neural networks are used, a balance can be achieved between real-time performance and stability, convolutional Neural Networks (CNNs) can generally process spatial features with high efficiency, while Recurrent Neural Networks (RNNs) are suitable for processing temporal features. Through reasonable structural design, good performance can be obtained in real-time and precision.
Fig. 2 is a schematic structural diagram of a multi-person 5D radar fall detection system based on an artificial intelligence algorithm according to an embodiment of the present invention, as shown in fig. 2, the system includes:
the device comprises a first unit, a second unit and a third unit, wherein the first unit is used for acquiring a plurality of radar echo information of a plurality of target objects in a target area based on a millimeter wave radar, extracting a plurality of motion characteristics corresponding to the plurality of radar echo information through a preset echo characteristic extraction model, and the echo characteristic extraction model transmits the output of each layer to the next layer through the connection coefficient of the adjacent layer and combining a nonlinear compression function;
the second unit is used for acquiring video information of a plurality of target objects in the target area based on the image sensor, determining the target area corresponding to the plurality of target objects in each frame of video information based on the target detector, matching the target area of each target object with the motion characteristics, and determining the comprehensive characteristics corresponding to each target object;
and the third unit is used for determining the action type of the target object through a preset action classification model according to the comprehensive characteristics corresponding to each target object, and sending out alarm information when the action type is matched with the falling action, wherein the action classification model is constructed based on a combined network model, and the action classification is carried out by distributing corresponding weights for different network models.
In a third aspect of an embodiment of the present invention,
there is provided an electronic device including:
a processor;
a memory for storing processor-executable instructions;
wherein the processor is configured to invoke the instructions stored in the memory to perform the method described previously.
In a fourth aspect of an embodiment of the present invention,
there is provided a computer readable storage medium having stored thereon computer program instructions which, when executed by a processor, implement the method as described above.
The present invention may be a method, apparatus, system, and/or computer program product. The computer program product may include a computer readable storage medium having computer readable program instructions embodied thereon for performing various aspects of the present invention.
Finally, it should be noted that: the above embodiments are only for illustrating the technical solution of the present invention, and not for limiting the same; although the invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical scheme described in the foregoing embodiments can be modified or some or all of the technical features thereof can be replaced by equivalents; such modifications and substitutions do not depart from the spirit of the invention.

Claims (6)

1. The method for detecting the falling of the multi-person 5D radar based on the artificial intelligence algorithm is characterized by comprising the following steps of:
acquiring multiple radar echo information of multiple target objects in a target area based on a millimeter wave radar, and extracting multiple motion features corresponding to the multiple radar echo information through a preset echo feature extraction model, wherein the echo feature extraction model transmits the output of each layer to the next layer through the connection coefficient of the adjacent layer and combining a nonlinear compression function;
acquiring video information of a plurality of target objects in the target area based on an image sensor, determining target areas corresponding to the plurality of target objects in each frame of video information based on a target detector, matching the target areas of each target object with motion characteristics, and determining comprehensive characteristics corresponding to each target object;
determining the action type of each target object through a preset action classification model according to the comprehensive characteristics corresponding to each target object, and sending out alarm information when the action type is matched with the falling action, wherein the action classification model is constructed based on a combined network model, and performs action classification by distributing corresponding weights for different network models;
acquiring a plurality of radar echo information of a plurality of target objects in a target area based on a millimeter wave radar, and extracting a plurality of motion features corresponding to the plurality of radar echo information through a preset echo feature extraction model comprises:
multiplying the radar echo information with a weight matrix corresponding to each layer in the echo feature extraction model to obtain low-level features;
according to the low-level characteristics, combining the connection coefficients of the current layer and the adjacent next layer to determine the high-level characteristics of the adjacent next layer;
based on the high-level features, carrying out space dimension mapping on the high-level features by combining a nonlinear compression function of the echo feature extraction model, and determining a plurality of motion features corresponding to the radar echo information;
determining a plurality of motion features corresponding to the plurality of radar echo information includes:
a plurality of motion characteristics are determined according to the following formula:
wherein,I l I h representing a low-level feature and a high-level feature respectively,C lh representing the connection coefficients of the current layer with the next adjacent layer,W l the weight matrix is represented by a matrix of weights,b l the bias parameter is indicated as such,P l Q h respectively represent a low-level learnable matrix parameter and a high-level learnable matrix parameter,S h the characteristics of the movement are represented by the characteristics of the movement,f()representing a non-linear compression function,errespectively representing nonlinearity and sensitivity;
before determining the target areas corresponding to the plurality of target objects in each frame of video information based on the target detector, the method further comprises training the target detector:
the target detector comprises a convolution layer, a candidate layer, a pooling layer and a full connection layer;
converting a training data set into a feature map through a convolution layer based on the pre-acquired training data set; generating a plurality of anchor frames with different sizes and aspect ratios in each sliding window by the candidate layer, and screening out the anchor frames with the front scores from the plurality of anchor frames as candidate frames by applying non-maximum value inhibition;
the pooling layer maps the candidate frames to the feature images with the same size as the candidate frames to carry out pooling operation and generate feature vectors;
determining classification loss and boundary box regression loss through a full connection layer according to the feature vector, and iteratively optimizing a classification loss function and a boundary box regression loss function through a back propagation algorithm in combination with an adaptive learning rate until the classification loss and the boundary box regression loss are minimized;
determining the classification loss and the boundary box regression loss through the full connection layer according to the feature vector, and iteratively optimizing the classification loss function and the boundary box regression loss function through a back propagation algorithm in combination with the self-adaptive learning rate comprises the following steps:
wherein,La sum of the loss values representing the classification loss value and the regression loss value,Sthe learning rate is indicated as being indicative of the learning rate,rthe loss weight coefficient is represented as a function of the loss weight coefficient,g t representation oftThe model gradient of the moment in time,θmodel parameters representing the object detection model,m t representation oftEstimating a first moment of time;
L cls representing the classification loss corresponding to the classification loss function,Nrepresenting the number of samples of the training data set,y i p i respectively represent the first of training data setsiActual tags corresponding to the individual samples and the first training data setiThe label prediction probabilities corresponding to the individual samples,
L reg representing the regression loss corresponding to the boundary box regression loss function,T i representing the first training data setiThe actual regression target corresponding to the individual samples,smooth()representing a smoothing loss function.
2. The method of claim 1, wherein determining the action type of the target object by the preset action classification model according to the comprehensive characteristics corresponding to each target object comprises:
the action classification model comprises a space sub-model and a time sub-model, wherein the space sub-model is constructed based on a convolutional neural network, the time sub-model is constructed based on a cyclic neural network,
taking the output of the space sub-model as the input of the time sub-model, applying global average pooling operation to the output of the time sub-model, distributing corresponding time weights to the feature vectors after the global average pooling operation through an adaptive weight distribution algorithm, and inputting the feature vectors after the global average pooling operation and the corresponding time weights into a full connection layer of the time sub-model;
and determining the probability distribution of the action category corresponding to the comprehensive characteristics through the classification activation function of the full connection layer, and taking the highest probability value in the probability distribution of each action category as the final action type.
3. The method of claim 2, wherein assigning corresponding temporal weights to the feature vectors after the global averaging pooling operation by an adaptive weight assignment algorithm comprises:
wherein,S T the time-weight is represented by a time-weight,、/>representing the sigmod function and the ReLu function, respectively, k representing the number of eigenvectors,/->Represent the firstk-1The temporal weights of the feature vectors after the global averaging pooling operation,Urepresenting the feature vector after the global averaging pooling operation.
4. An artificial intelligence algorithm-based multi-person 5D radar fall detection system, comprising:
the device comprises a first unit, a second unit and a third unit, wherein the first unit is used for acquiring a plurality of radar echo information of a plurality of target objects in a target area based on a millimeter wave radar, extracting a plurality of motion characteristics corresponding to the plurality of radar echo information through a preset echo characteristic extraction model, and the echo characteristic extraction model transmits the output of each layer to the next layer through the connection coefficient of the adjacent layer and combining a nonlinear compression function;
the second unit is used for acquiring video information of a plurality of target objects in the target area based on the image sensor, determining the target area corresponding to the plurality of target objects in each frame of video information based on the target detector, matching the target area of each target object with the motion characteristics, and determining the comprehensive characteristics corresponding to each target object;
a third unit, configured to determine, according to the comprehensive characteristics corresponding to each target object, an action type of the target object through a preset action classification model, and send out alarm information when the action type is matched with a falling action, where the action classification model is constructed based on a combined network model, and performs action classification by allocating corresponding weights to different network models;
acquiring a plurality of radar echo information of a plurality of target objects in a target area based on a millimeter wave radar, and extracting a plurality of motion features corresponding to the plurality of radar echo information through a preset echo feature extraction model comprises:
multiplying the radar echo information with a weight matrix corresponding to each layer in the echo feature extraction model to obtain low-level features;
according to the low-level characteristics, combining the connection coefficients of the current layer and the adjacent next layer to determine the high-level characteristics of the adjacent next layer;
based on the high-level features, carrying out space dimension mapping on the high-level features by combining a nonlinear compression function of the echo feature extraction model, and determining a plurality of motion features corresponding to the radar echo information;
determining a plurality of motion features corresponding to the plurality of radar echo information includes:
a plurality of motion characteristics are determined according to the following formula:
wherein,I l I h representing a low-level feature and a high-level feature respectively,C lh representing the connection coefficients of the current layer with the next adjacent layer,W l the weight matrix is represented by a matrix of weights,b l the bias parameter is indicated as such,P l Q h respectively represent a low-level learnable matrix parameter and a high-level learnable matrix parameter,S h the characteristics of the movement are represented by the characteristics of the movement,f()representing a non-linear compression function,errespectively representing nonlinearity and sensitivity;
before determining the target areas corresponding to the plurality of target objects in each frame of video information based on the target detector, the system further comprises training the target detector:
the target detector comprises a convolution layer, a candidate layer, a pooling layer and a full connection layer;
converting a training data set into a feature map through a convolution layer based on the pre-acquired training data set; generating a plurality of anchor frames with different sizes and aspect ratios in each sliding window by the candidate layer, and screening out the anchor frames with the front scores from the plurality of anchor frames as candidate frames by applying non-maximum value inhibition;
the pooling layer maps the candidate frames to the feature images with the same size as the candidate frames to carry out pooling operation and generate feature vectors;
determining classification loss and boundary box regression loss through a full connection layer according to the feature vector, and iteratively optimizing a classification loss function and a boundary box regression loss function through a back propagation algorithm in combination with an adaptive learning rate until the classification loss and the boundary box regression loss are minimized;
determining the classification loss and the boundary box regression loss through the full connection layer according to the feature vector, and iteratively optimizing the classification loss function and the boundary box regression loss function through a back propagation algorithm in combination with the self-adaptive learning rate comprises the following steps:
wherein,La sum of the loss values representing the classification loss value and the regression loss value,Sthe learning rate is indicated as being indicative of the learning rate,rthe loss weight coefficient is represented as a function of the loss weight coefficient,g t representation oftThe model gradient of the moment in time,θmodel parameters representing the object detection model,m t representation oftEstimating a first moment of time;
L cls representing the classification loss corresponding to the classification loss function,Nrepresenting the number of samples of the training data set,y i p i respectively represent the first of training data setsiActual tags corresponding to the individual samples and the first training data setiThe label prediction probabilities corresponding to the individual samples,
L reg representing the regression loss corresponding to the boundary box regression loss function,T i representing the first training data setiThe actual regression target corresponding to the individual samples,smooth()representing a smoothing loss function.
5. An electronic device, comprising:
a processor;
a memory for storing processor-executable instructions;
wherein the processor is configured to invoke the instructions stored in the memory to perform the method of any of claims 1 to 3.
6. A computer readable storage medium having stored thereon computer program instructions, which when executed by a processor, implement the method of any of claims 1 to 3.
CN202311335592.8A 2023-10-16 2023-10-16 Multi-person 5D radar falling detection method and system based on artificial intelligence algorithm Active CN117079416B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311335592.8A CN117079416B (en) 2023-10-16 2023-10-16 Multi-person 5D radar falling detection method and system based on artificial intelligence algorithm

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311335592.8A CN117079416B (en) 2023-10-16 2023-10-16 Multi-person 5D radar falling detection method and system based on artificial intelligence algorithm

Publications (2)

Publication Number Publication Date
CN117079416A CN117079416A (en) 2023-11-17
CN117079416B true CN117079416B (en) 2023-12-26

Family

ID=88713737

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311335592.8A Active CN117079416B (en) 2023-10-16 2023-10-16 Multi-person 5D radar falling detection method and system based on artificial intelligence algorithm

Country Status (1)

Country Link
CN (1) CN117079416B (en)

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109447048A (en) * 2018-12-25 2019-03-08 苏州闪驰数控系统集成有限公司 A kind of artificial intelligence early warning system
CN110532966A (en) * 2019-08-30 2019-12-03 深兰科技(上海)有限公司 A kind of method and apparatus carrying out tumble identification based on disaggregated model
CN112633101A (en) * 2020-12-14 2021-04-09 深兰人工智能(深圳)有限公司 Obstacle speed detection method and device
CN113807314A (en) * 2021-10-08 2021-12-17 江苏云禾峰智能科技有限公司 Millimeter wave radar video fusion method based on micro-Doppler effect
CN113903147A (en) * 2021-09-30 2022-01-07 湖南时变通讯科技有限公司 Radar-based human body posture distinguishing method, device, equipment and medium
CN115209798A (en) * 2019-12-23 2022-10-18 沃伊亚影像有限公司 Fall detection system and method
CN115291184A (en) * 2022-10-08 2022-11-04 四川启睿克科技有限公司 Attitude monitoring method combining millimeter wave radar and deep learning
WO2023274254A1 (en) * 2021-06-29 2023-01-05 上海高德威智能交通系统有限公司 Object detection method, apparatus and system, electronic device, and storage medium
CN116027324A (en) * 2023-03-24 2023-04-28 德心智能科技(常州)有限公司 Fall detection method and device based on millimeter wave radar and millimeter wave radar equipment
WO2023133007A1 (en) * 2022-01-04 2023-07-13 Qualcomm Incorporated Machine learning based object detection using radar information
CN116682105A (en) * 2023-05-24 2023-09-01 桂林电子科技大学 Millimeter wave radar and visual feature attention fusion target detection method

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109581361A (en) * 2018-11-22 2019-04-05 九牧厨卫股份有限公司 A kind of detection method, detection device, terminal and detection system
KR102580688B1 (en) * 2022-10-26 2023-09-20 주식회사 에너자이 A training method of a fall detection model based on a radar data

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109447048A (en) * 2018-12-25 2019-03-08 苏州闪驰数控系统集成有限公司 A kind of artificial intelligence early warning system
CN110532966A (en) * 2019-08-30 2019-12-03 深兰科技(上海)有限公司 A kind of method and apparatus carrying out tumble identification based on disaggregated model
CN115209798A (en) * 2019-12-23 2022-10-18 沃伊亚影像有限公司 Fall detection system and method
CN112633101A (en) * 2020-12-14 2021-04-09 深兰人工智能(深圳)有限公司 Obstacle speed detection method and device
WO2023274254A1 (en) * 2021-06-29 2023-01-05 上海高德威智能交通系统有限公司 Object detection method, apparatus and system, electronic device, and storage medium
CN113903147A (en) * 2021-09-30 2022-01-07 湖南时变通讯科技有限公司 Radar-based human body posture distinguishing method, device, equipment and medium
CN113807314A (en) * 2021-10-08 2021-12-17 江苏云禾峰智能科技有限公司 Millimeter wave radar video fusion method based on micro-Doppler effect
WO2023133007A1 (en) * 2022-01-04 2023-07-13 Qualcomm Incorporated Machine learning based object detection using radar information
CN115291184A (en) * 2022-10-08 2022-11-04 四川启睿克科技有限公司 Attitude monitoring method combining millimeter wave radar and deep learning
CN116027324A (en) * 2023-03-24 2023-04-28 德心智能科技(常州)有限公司 Fall detection method and device based on millimeter wave radar and millimeter wave radar equipment
CN116682105A (en) * 2023-05-24 2023-09-01 桂林电子科技大学 Millimeter wave radar and visual feature attention fusion target detection method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
基于毫米波雷达的人体行为识别算法研究;戴鹏程;《中国优秀硕士学位论文全文数据库信息科技辑》;正文39-53页 *
基于深度学习的低频雷达人体典型动作分类研究;陈思贤;《中国优秀硕士学位论文全文数据库信息科技辑》;正文37-55页 *

Also Published As

Publication number Publication date
CN117079416A (en) 2023-11-17

Similar Documents

Publication Publication Date Title
CN110135319B (en) Abnormal behavior detection method and system
CN111027487B (en) Behavior recognition system, method, medium and equipment based on multi-convolution kernel residual error network
CN108921051B (en) Pedestrian attribute identification network and technology based on cyclic neural network attention model
CN110569901B (en) Channel selection-based countermeasure elimination weak supervision target detection method
JP6517681B2 (en) Image pattern learning apparatus, method and program
US20230048405A1 (en) Neural network optimization method and apparatus
JP6892606B2 (en) Positioning device, position identification method and computer program
CN111368634B (en) Human head detection method, system and storage medium based on neural network
EP4187213A1 (en) Rgb-d-based ai system for inferring weight of livestock, and computer-readable medium having recorded thereon computer program for providing same
US20230316720A1 (en) Anomaly detection apparatus, anomaly detection method, and program
CN115620141A (en) Target detection method and device based on weighted deformable convolution
CN116879910B (en) Laser scanning distance measuring device and method thereof
CN113065379B (en) Image detection method and device integrating image quality and electronic equipment
CN117079416B (en) Multi-person 5D radar falling detection method and system based on artificial intelligence algorithm
CN116960954A (en) Short-term wind power prediction method and device, storage medium and computer equipment
CN108596068B (en) Method and device for recognizing actions
CN113344168B (en) Short-term berth prediction method and system
CN114898202A (en) Underwater video target scale space discriminant tracking system and method based on multi-model fusion
CN114049684A (en) Human body sitting posture identification method and device, electronic equipment and storage medium
CN113705328A (en) Depression detection method and system based on facial feature points and facial movement units
CN117037287B (en) Behavior recognition method, system and device based on 3D impulse neural network
CN117689880B (en) Method and system for target recognition in biomedical images based on machine learning
CN117523428B (en) Ground target detection method and device based on aircraft platform
CN117315380B (en) Deep learning-based pneumonia CT image classification method and system
CN117668701B (en) AI artificial intelligence machine learning system and method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant