CN116907510A - Intelligent motion recognition method based on Internet of things technology - Google Patents

Intelligent motion recognition method based on Internet of things technology Download PDF

Info

Publication number
CN116907510A
CN116907510A CN202310886760.6A CN202310886760A CN116907510A CN 116907510 A CN116907510 A CN 116907510A CN 202310886760 A CN202310886760 A CN 202310886760A CN 116907510 A CN116907510 A CN 116907510A
Authority
CN
China
Prior art keywords
feature set
training
internet
target
unit time
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202310886760.6A
Other languages
Chinese (zh)
Other versions
CN116907510B (en
Inventor
杨勇
刘彩文
余成柱
叶兴盛
张平
林尧
李文友
邵振华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fujian University Of Science And Technology
Fujian Yaoshun Development Co ltd
Minsheng Group Urban Construction Development Co ltd
Fujian Haoli Construction Engineering Co ltd
Minjiang University
Original Assignee
Fujian University Of Science And Technology
Fujian Yaoshun Development Co ltd
Minsheng Group Urban Construction Development Co ltd
Fujian Haoli Construction Engineering Co ltd
Minjiang University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fujian University Of Science And Technology, Fujian Yaoshun Development Co ltd, Minsheng Group Urban Construction Development Co ltd, Fujian Haoli Construction Engineering Co ltd, Minjiang University filed Critical Fujian University Of Science And Technology
Priority to CN202310886760.6A priority Critical patent/CN116907510B/en
Publication of CN116907510A publication Critical patent/CN116907510A/en
Application granted granted Critical
Publication of CN116907510B publication Critical patent/CN116907510B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/20Instruments for performing navigational calculations
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/005Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 with correlation of navigation data from several sources, e.g. map or contour matching
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S19/00Satellite radio beacon positioning systems; Determining position, velocity or attitude using signals transmitted by such systems
    • G01S19/38Determining a navigation solution using signals transmitted by a satellite radio beacon positioning system
    • G01S19/39Determining a navigation solution using signals transmitted by a satellite radio beacon positioning system the satellite radio beacon positioning system transmitting time-stamped messages, e.g. GPS [Global Positioning System], GLONASS [Global Orbiting Navigation Satellite System] or GALILEO
    • G01S19/42Determining position
    • G01S19/48Determining position by combining or switching between position solutions derived from the satellite radio beacon positioning system and position solutions derived from a further system
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/10Pre-processing; Data cleansing
    • G06F18/15Statistical pre-processing, e.g. techniques for normalisation or restoring missing data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2131Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on a transform domain processing, e.g. wavelet transform
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • G06N3/0442Recurrent networks, e.g. Hopfield networks characterised by memory or gating, e.g. long short-term memory [LSTM] or gated recurrent units [GRU]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/12Protocols specially adapted for proprietary or special-purpose networking environments, e.g. medical networks, sensor networks, networks in vehicles or remote metering networks
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D30/00Reducing energy consumption in communication networks
    • Y02D30/70Reducing energy consumption in communication networks in wireless communication networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Remote Sensing (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Automation & Control Theory (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Evolutionary Biology (AREA)
  • Signal Processing (AREA)
  • Probability & Statistics with Applications (AREA)
  • Medical Informatics (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention relates to the field of data processing, in particular to an intelligent motion recognition method based on the technology of the Internet of things, which comprises the following steps: acquiring various original information of a target in each unit time through the Internet of things, extracting an initial feature set through the various original information, splicing the corresponding initial feature set to obtain a first feature set, obtaining a strengthening mapping feature set through a self-attention mechanism, splicing to obtain a second feature set, configuring a long-term and short-term memory network, inputting the second feature set for training to obtain a pre-training model, and further training to obtain a final model. According to the method, more comprehensive information is collected through different types of Internet of things equipment, the self-attention mechanism is adopted to obtain the enhanced mapping feature set, the long-period memory network training model is used, and after the final model is obtained, the prediction precision of the target motion trail is improved.

Description

Intelligent motion recognition method based on Internet of things technology
Technical Field
The invention relates to the technical field of motion recognition, in particular to an intelligent motion recognition method based on the internet of things technology.
Background
In an automated warehouse or factory environment, autonomous navigational robots, drones, or vehicles are changing the manner in which logistics and production lines operate. These autonomous navigational units need to perform tasks in complex and dynamic environments, including moving to a designated destination on a predetermined path and avoiding other moving obstacle targets en route. In this case, the ability to predict the target motion profile in real time becomes critical.
Current solutions rely mainly on sensors of the autonomous navigation unit itself to scan the surrounding environment, such as using cameras, lidar, infrared sensors or ultrasonic sensors. The sensors can collect information about the surrounding environment, help autonomous navigational units identify the trajectory of the target and make predictions, and then plan new paths. The autonomous navigation unit adjusts the motion trail of the autonomous navigation unit according to the information, avoids potential obstacles, and completes tasks. Patent document publication No. CN106054889B discloses a robot autonomous obstacle avoidance method and device, patent document publication No. CN110597269B discloses a vehicle autonomous obstacle avoidance method and a vehicle autonomous obstacle avoidance system, and both methods rely on own sensors for obstacle recognition and avoidance.
However, such a single position sensor-dependent solution may suffer from a problem of low prediction accuracy.
Disclosure of Invention
Therefore, the invention provides an intelligent motion recognition method based on the internet of things technology, which solves the problem of low accuracy in recognition of target motion in the prior art.
In order to achieve the above object, an aspect of the present invention provides an intelligent motion recognition method based on the internet of things technology, the method comprising:
acquiring first original information of a target in each unit time through radio frequency identification equipment in the Internet of things, acquiring second original information of the target in each unit time through a global positioning system in the Internet of things, acquiring third original information of the target in each unit time through an infrared motion sensor in the Internet of things, and acquiring fourth original information of the target in each unit time through a laser scanner in the Internet of things;
extracting a first initial feature set based on the first original information, extracting a second initial feature set based on the second original information, extracting a third initial feature set based on the third original information, and extracting a fourth initial feature set based on the fourth original information;
the first initial feature set, the second initial feature set and the third initial feature set are spliced to obtain a first feature set, an enhanced mapping feature set of a fourth initial feature set for the first feature set is obtained through a self-attention mechanism, and the enhanced mapping feature set is spliced with the first feature set to obtain a second feature set;
configuring a long-term and short-term memory network;
inputting the second feature set into the long-short-period memory network for training to obtain a pre-training model capable of predicting a target motion track, inputting the second feature set obtained in a first unit time into the pre-training model to obtain a predicted track of the target in a second unit time, comparing the predicted track with an actual track to calculate a prediction error after the second unit time is over, taking the prediction error as a third feature set, and inputting the third feature set and the second feature set into the pre-training model for training to obtain a final model.
Further, after the final model is obtained, inputting the second feature set obtained in the first unit time into the final model, so that a predicted track of the target in the second unit time can be obtained;
in the second unit time, if the prediction error does not appear, inputting a second feature set obtained in the second unit time into a final model to obtain a predicted track of the target in the third unit time;
and if the prediction error occurs, inputting the second feature set and the third feature set obtained in the second unit time into a final model to obtain a predicted track of the target in the third unit time.
Further, extracting the first initial feature set based on the first original information includes unicoding the first original information to extract the first initial feature set;
extracting a second initial feature set based on the second original information includes using a min-max normalization process on the second original information to extract a second initial feature set;
extracting a third initial feature set based on the third initial information includes converting the third initial information from a time domain to a frequency domain using fourier transform, and then extracting a main frequency component in the frequency domain as the third initial feature set;
extracting the fourth initial feature set based on the fourth original information includes performing a Z-score normalization process on the fourth original information to extract the fourth initial feature set.
Further, obtaining the first feature set includes: and parallelly splicing the values of the first initial feature set, the second initial feature set and the third initial feature set at the same time to form a matrix, so as to obtain a first feature set.
Further, the enhanced mapping feature set is obtained by calculating feature vector scores of a query vector, a key vector and a value vector through a self-attention mechanism of a transducer model.
Further, the obtaining of the query vector, the key vector, and the value vector includes: obtaining a query vector by multiplying a weight matrix by the first feature set, obtaining a key vector and a value vector by multiplying a weight matrix by the fourth initial feature set, wherein the weight matrix is randomly initialized based on a calculation flow of a transducer model.
Further, configuring the long-term memory network includes: adopting a multi-layer long-short-term memory network structure, and performing preliminary processing on the second feature set by the first layer network to generate intermediate state features;
the second layer network processes the intermediate state characteristics generated by the first layer to obtain high-level characteristics;
the third layer network converts the output of the second layer into the motion state prediction of the target;
during the training process, each layer of network updates network parameters through a time back propagation algorithm to minimize a loss function that measures the difference between the predicted target motion state and the true motion state.
Further, obtaining a pre-training model capable of predicting a target motion trajectory includes:
and when the second feature set is input into the long-short-term memory network for training, applying an early stop and regularization over-fitting prevention technology, wherein the output of the pre-training model is a set of coordinates, and obtaining a predicted track of the target after fitting.
Further, comparing the predicted trajectory with the actual trajectory to calculate a prediction error includes: and calculating the difference between the predicted track and the actual track at each moment in unit time, summing the squares of each difference, and finally dividing the sum by the total moment number to obtain a mean square error, wherein the mean square error is the predicted error.
Further, inputting the third feature set and the second feature set into the pre-training model for training, and obtaining a final model includes: and splicing the third feature set and the second feature set according to a time sequence to obtain a fusion feature set, inputting the fusion feature set into the pre-training model for training, dividing the fusion feature set into a training set and a verification set in the training process, performing parameter optimization by using the loss function and a random gradient descent strategy, dividing data into a plurality of batches by adopting a batch training mode, and training by using only one batch of data each time.
Compared with the prior art, the invention has the beneficial effects that a plurality of different types of internet of things equipment are comprehensively used for collecting more comprehensive information, including radio frequency identification equipment, a global positioning system, an infrared motion sensor and a laser scanner. These devices can provide richer, more accurate data than a single position sensor, thereby improving the accuracy of target motion trajectory prediction.
In particular, the invention adopts a self-attention mechanism to strengthen the mapping feature set, thereby further improving the utilization efficiency of the data. The self-attention mechanism enables the system to better understand and analyze correlations and complex patterns in the data, thereby enhancing the performance of the predictive model.
In particular, the invention can significantly improve the motion prediction capability of autonomous navigation units in complex and dynamic environments. Through the use of long and short term memory networks, the invention can learn and memorize past data, thereby more accurately predicting future tracks.
In particular, the invention can update the prediction result in real time, and after each unit time is finished, the prediction error of the last unit time is introduced as a new feature set to train, so that the prediction result can be timely adjusted under the condition that the parameters of the model are unchanged, and the model adapts to the dynamic change of the environment.
In summary, the method and the device can adapt to complex and dynamic environments while improving the accuracy of predicting the target motion trail, and provide more accurate and reliable prediction for the motion of the autonomous navigation unit in logistics and production lines, thereby effectively improving the task execution capability.
Drawings
Fig. 1 is a flowchart of an intelligent motion recognition method based on the internet of things technology according to an embodiment of the present invention.
Detailed Description
In order that the objects and advantages of the invention will become more apparent, the invention will be further described with reference to the following examples; it should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
Preferred embodiments of the present invention are described below with reference to the accompanying drawings. It should be understood by those skilled in the art that these embodiments are merely for explaining the technical principles of the present invention, and are not intended to limit the scope of the present invention.
It should be noted that, in the description of the present invention, terms such as "upper," "lower," "left," "right," "inner," "outer," and the like indicate directions or positional relationships based on the directions or positional relationships shown in the drawings, which are merely for convenience of description, and do not indicate or imply that the apparatus or elements must have a specific orientation, be constructed and operated in a specific orientation, and thus should not be construed as limiting the present invention.
Furthermore, it should be noted that, in the description of the present invention, unless explicitly specified and limited otherwise, the terms "mounted," "connected," and "connected" are to be construed broadly, and may be either fixedly connected, detachably connected, or integrally connected, for example; can be mechanically or electrically connected; can be directly connected or indirectly connected through an intermediate medium, and can be communication between two elements. The specific meaning of the above terms in the present invention can be understood by those skilled in the art according to the specific circumstances.
Referring to fig. 1, an intelligent motion recognition method based on the internet of things technology provided by an embodiment of the present invention includes:
step S100: acquiring first original information of a target in each unit time through radio frequency identification equipment in the Internet of things, acquiring second original information of the target in each unit time through a global positioning system in the Internet of things, acquiring third original information of the target in each unit time through an infrared motion sensor in the Internet of things, and acquiring fourth original information of the target in each unit time through a laser scanner in the Internet of things;
step S200: extracting a first initial feature set based on the first original information, extracting a second initial feature set based on the second original information, extracting a third initial feature set based on the third original information, and extracting a fourth initial feature set based on the fourth original information;
step S300: the first initial feature set, the second initial feature set and the third initial feature set are spliced to obtain a first feature set, an enhanced mapping feature set of a fourth initial feature set for the first feature set is obtained through a self-attention mechanism, and the enhanced mapping feature set is spliced with the first feature set to obtain a second feature set;
step S400: configuring a long-term and short-term memory network;
step S500: inputting the second feature set into the long-short-period memory network for training to obtain a pre-training model capable of predicting a target motion track, inputting the second feature set obtained in a first unit time into the pre-training model to obtain a predicted track of the target in a second unit time, comparing the predicted track with an actual track to calculate a prediction error after the second unit time is over, taking the prediction error as a third feature set, and inputting the third feature set and the second feature set into the pre-training model for training to obtain a final model.
Specifically, in step S100, four kinds of internet of things devices, namely, a Radio Frequency Identification Device (RFID), a Global Positioning System (GPS), an infrared motion sensor, and a laser scanner are relied upon. These devices are all key components to achieve target motion recognition. The RFID can identify and track the target in real time in the environment of the Internet of things, and provides the identity and state information of the target. GPS can provide accurate geographic location and speed of movement of objects. The infrared motion sensor is capable of capturing motion and behavioral characteristics of a target, such as walking, running, jumping, and the like. The laser scanner is then able to provide high precision three-dimensional information of the environment and the shape of the target. The four devices are connected through a wireless network and work cooperatively to acquire the comprehensive information of the target.
After step S100, we will obtain a series of raw data related to the motion of the object, including the identification information, position information, motion characteristics and three-dimensional shape information of the object. This information is critical to accurately predicting the motion trajectory of the target, which will be used to generate feature sets that will be used as inputs to the deep learning model for training and prediction. Step S100 is the first step of the whole intelligent motion recognition method, and is also the most basic data collection step, and the quality of the data collection step directly affects the effect of the subsequent steps. For the whole invention, the step provides the basic data required by the whole intelligent motion recognition method, and is the basis for realizing effective motion recognition.
Specifically, in step S200, based on four types of original information, we need to perform feature extraction for each type of information. Feature extraction generally depends on the nature of the data and the target task.
Specifically, extracting the first initial feature set based on the first original information includes performing one-time encoding on the first original information to extract the first initial feature set;
extracting a second initial feature set based on the second original information includes using a min-max normalization process on the second original information to extract a second initial feature set;
extracting a third initial feature set based on the third initial information includes converting the third initial information from a time domain to a frequency domain using fourier transform, and then extracting a main frequency component in the frequency domain as the third initial feature set;
extracting the fourth initial feature set based on the fourth original information includes performing a Z-score normalization process on the fourth original information to extract the fourth initial feature set.
Specifically, step S200 involves some feature extraction and preprocessing techniques. The first original information is acquired through radio frequency identification equipment, and specifically is a unique identifier of a target; the second original information is obtained through a global positioning system, and specifically is geographic coordinates of a target; the third initial information is obtained through an infrared motion sensor, and specifically is the position coordinates and the motion state of the target; the fourth raw information is acquired by a laser scanner, in particular the distance or depth information of the object. For the first original information, features are extracted by means of One-hot Encoding (One-hot Encoding), which is used to process category type data, which is converted into numerical type data capable of being input into a model by Encoding. For the second raw information, the data is processed by means of a Min-max normalization (Min-Max Normalization), which can map the data to a specific range and reduce the influence of outliers on the model. For the third raw information, the frequency domain information is extracted using fourier transform (Fourier Transform), and the dynamic characteristics of the data can be analyzed and understood from the frequency domain perspective. For the fourth original information, Z score normalization (Z-Score Normalization) is carried out, and the method can normalize data to a standard normal distribution with a mean value of 0 and a variance of 1, so that optimization of an algorithm and stability of a model are facilitated.
Specifically, after step S200 is completed, we obtain four initial feature sets obtained by performing various processes and feature extraction on four kinds of original information, where the feature sets are data after preprocessing and feature engineering, and are more suitable for being input into a model for training and prediction. In the whole process, step S200 is in a data preprocessing and feature extraction stage, which is to convert the original data into feature data that can be input into the model, and is the basis and premise of the whole model training. For the whole invention, the step ensures the quality of model input data, provides possibility for training and optimizing the model afterwards, and is beneficial to improving the accuracy of model prediction.
Specifically, in step S300, feature merging and self-attention mechanisms are involved. First, the first, second and third initial feature sets are spliced to obtain a first feature set, which is mainly feature fusion, and features from various sources are combined together to obtain a more comprehensive feature representation. Next, a fourth initial feature set is processed by a Self-attention mechanism (Self-Attention Mechanism) that can weight each feature entered, extracting relatively important feature information. This mechanism has the capability of capturing relationships between features and is particularly effective for sequential data. The enhanced mapping feature set obtained through the self-attention mechanism is spliced with the first feature set to form a second feature set.
Specifically, obtaining the first feature set includes: and parallelly splicing the values of the first initial feature set, the second initial feature set and the third initial feature set at the same time to form a matrix, so as to obtain a first feature set.
In particular, the above mainly relates to stitching of feature vectors. This is a common feature processing technique used to combine features from different sources together to form a more comprehensive characterization. For example, at a certain time, the first initial feature set is [0.2,0.3], the second initial feature set is [0.1,0.6], and the third initial feature set is [0.4,0.7], then at that time, the result of stitching the three feature sets is [ [0.2,0.1,0.4], [0.3,0.6,0.7] ]. Thus, we obtain a feature vector containing all the initial feature information.
Specifically, a two-dimensional matrix is obtained after stitching, wherein each row contains all the features at a certain point in time. The splicing in the mode ensures that the influence of time continuity is fully considered in the characteristic processing stage, and more information is provided for subsequent model training. This step is in the feature integration stage throughout the process, i.e., the combination of the initial feature sets from the previous step.
Specifically, the enhancement map feature set is obtained by performing feature vector score calculation of a query vector, a key vector and a value vector through a self-attention mechanism of a transducer model.
In particular, the self-attention mechanism is a core component of the transducer model. It obtains the enhanced mapping feature set by computing scores for the query vector, the key vector, and the value vector.
Specifically, the obtaining of the query vector, the key vector, and the value vector includes: obtaining a query vector by multiplying a weight matrix by the first feature set, obtaining a key vector and a value vector by multiplying a weight matrix by the fourth initial feature set, wherein the weight matrix is randomly initialized based on a calculation flow of a transducer model.
In particular, the process of acquisition of query vectors, key vectors, and value vectors is accomplished by multiplying a weight matrix by a feature set. As a specific example, we have a first feature set that is the vector [ [0.2,0.1,0.4], [0.3,0.6,0.7] ], while we have randomly initialized weight matrices w_q, w_k, and w_v. These weight matrices are used to generate query vectors, key vectors, and value vectors. By multiplying the first feature set with the weight matrix W q, a query vector can be obtained. The key vector and the value vector are also multiplied by a fourth initial feature set (e.g., a 1 x 4 vector 0.6,0.3,0.2,0.9) using the corresponding weight matrix.
Specifically, by computing the dot product of the query vector and the key vector, and then performing a softmax function conversion, a set of weights are obtained, which are used to weight sum the value vectors, resulting in an enhanced mapping feature set.
Specifically, the enhanced mapping feature set has stronger representativeness, effectively extracts important information and weakens unimportant information, and can better represent the characteristics of the target object. The step is in the feature processing stage of the whole operation process, and is to further process the input features, extract more effective information and prepare for subsequent model training.
Specifically, a more descriptive and differentiated second feature set is obtained after completion of step S300. The method is a feature set enhanced through feature fusion and a self-attention mechanism, contains comprehensive information of original data, and enhances the identification and grabbing of key features. In the whole flow, the step still belongs to a feature preprocessing stage, which is a further expansion of feature extraction, and the expression capacity of the features is improved through different technical means. The method has the significance that the learning effect of the subsequent model can be improved through the reinforcement and fusion of the characteristics, and the prediction accuracy is improved, so that the motion trail of the target is predicted better.
Specifically, in step S400, a long-short-term memory network is configured to initialize the pre-training model.
Specifically, configuring the long-term memory network includes: adopting a multi-layer long-short-term memory network structure, and performing preliminary processing on the second feature set by the first layer network to generate intermediate state features;
the second layer network processes the intermediate state characteristics generated by the first layer to obtain high-level characteristics;
the third layer network converts the output of the second layer into the motion state prediction of the target;
during the training process, each layer of network updates network parameters through a time back propagation algorithm to minimize a loss function that measures the difference between the predicted target motion state and the true motion state.
Specifically, in step S400, the process involves a multi-layered long and short term memory network (LSTM) structure. In this configuration, the first layer network performs a preliminary process on the second feature set generated in step S300 to generate intermediate state features. The second level network then processes these intermediate state features again, generating higher level features. The third level network then converts these high level features into a motion state prediction of the target. In the whole training process, the parameters of each layer of network are updated through a time back propagation algorithm so as to minimize a loss function, and the loss function measures the difference between the motion state of the predicted target and the actual motion state.
Specifically, in step S500, the second feature set is input into the long-short term memory network for training, to obtain a pre-training model capable of predicting the target motion trajectory, and in a simple example, it is assumed that there is a second feature set, which is a 1×4 vector [0.2,0.3,0.7,0.1], and after the first layer network processing, we obtain a new intermediate state feature vector, for example [0.4,0.5,0.2,0.1]. The second tier network then processes it into a higher level feature vector, e.g. [0.6,0.3,0.1,0.1]. Finally, the third layer network converts it to predicted target motion states, e.g., [0.7,0.8], [0.6,0.9], [0.5,0.8].
Specifically, the obtaining of the pre-training model capable of predicting the target motion trajectory includes: and when the second feature set is input into the long-short-term memory network for training, applying an early stop and regularization over-fitting prevention technology, wherein the output of the pre-training model is a set of coordinates, and obtaining a predicted track of the target after fitting.
Specifically, the technical features involved in step S500 are mainly the training process of the pre-training model and the generation of the predicted trajectory. In the training process, the second feature set is input into a long-term and short-term memory network for training, and meanwhile, an early stop and regularized overfitting prevention technology is applied to avoid overlearning of the model on the training set, so that the generalization capability of the model on unseen samples is damaged. The output of the pre-trained model is a set of coordinates that can be used to fit the predicted trajectory of the target. For a specific example, assuming the second feature set is a 1 x 3 vector [0.2,0.3,0.5], after training through the LSTM network we may get a set of output coordinates such as [0.7,0.8], [0.6,0.9], [0.5,0.8]. These coordinate points may be linked together to form a trajectory, which is the predicted trajectory of the object.
Specifically, after the pre-training model is obtained, the model can output a set of coordinates according to the input feature set, and the set of coordinates can form a predicted track of the target after fitting. The pre-training model can preliminarily predict the motion trail of the target in the future. Preparation is made for further obtaining the final model.
Specifically, comparing the predicted trajectory with the actual trajectory to calculate a prediction error includes: and calculating the difference between the predicted track and the actual track at each moment in unit time, summing the squares of each difference, and finally dividing the sum by the total moment number to obtain a mean square error, wherein the mean square error is the predicted error.
In particular, the process of calculating the prediction error mainly involves a comparison between the predicted trajectory and the actual trajectory, and a calculation of the mean square error. The predicted trajectory is a set of coordinates output by the pre-training model, and the actual trajectory is the motion trajectory of the target object in the actual environment. By comparing the two trajectories, the difference between the predicted trajectory and the actual trajectory at each time can be calculated. And summing each difference value after squaring, dividing the sum by the total time quantity to obtain a result which is a mean square error, wherein the value can be used for measuring the prediction accuracy of the prediction model. Assuming that three times are set, the difference between the predicted and actual trajectories is 0.1,0.2,0.1, respectively, the mean square error is ((0.1≡2+0.2≡2+0.1≡2)/3) =0.02.
Specifically, after the prediction error is calculated, a specific value of the prediction error is obtained, and this value reflects the prediction accuracy of the prediction model. This result will provide important feedback about the performance of the model, which is beneficial to tuning and optimizing the model, improving the prediction accuracy of the model. This is an important step in model training, and a third feature set is obtained, based on which the pre-trained model is further trained, enabling optimization of the model.
Specifically, inputting the third feature set and the second feature set into the pre-training model for training, and obtaining a final model includes: and splicing the third feature set and the second feature set according to a time sequence to obtain a fusion feature set, inputting the fusion feature set into the pre-training model for training, dividing the fusion feature set into a training set and a verification set in the training process, performing parameter optimization by using the loss function and a random gradient descent strategy, dividing data into a plurality of batches by adopting a batch training mode, and training by using only one batch of data each time.
Specifically, the method mainly relates to the technical characteristics of feature set splicing, training of a pre-trained model, parameter optimization by using a loss function and a random gradient descent strategy, and the like. First, the third feature set and the second feature set are spliced according to the time sequence, and a fusion feature set is obtained. This fused feature set is then input into a pre-training model for training. In the training process, the fusion feature set is divided into a training set and a verification set, and then parameter optimization is performed by using a loss function and a random gradient descent strategy.
Specifically, the third feature set, that is, the prediction error, is not used as a label, but is a part of the feature, and the final model does not adjust its parameters according to the third feature set in practical application, but remains stable.
Specifically, the final model will be obtained after this step is completed. This model is trained and optimized based on the fused feature set, so that it can better utilize the information in the third feature set and the second feature set for prediction. This step is in the model training and optimization stage during the whole operation process, and its objective is to improve the prediction performance of the model through training and optimization. Implementation of this step may improve the accuracy of the predictive model for the entire invention, thereby improving the user experience. The method has the advantages that the method can utilize various characteristic information to predict, and the accuracy of prediction is improved.
Specifically, the final output stage of the whole operation process marks the establishment of the whole prediction model, and key technical support is provided for realizing safer and more accurate prediction.
Specifically, after the final model is obtained, inputting the second feature set obtained in the first unit time into the final model, so that a predicted track of a target in the second unit time can be obtained;
in the second unit time, if the prediction error does not appear, inputting a second feature set obtained in the second unit time into a final model to obtain a predicted track of the target in the third unit time;
and if the prediction error occurs, inputting the second feature set and the third feature set obtained in the second unit time into a final model to obtain a predicted track of the target in the third unit time.
Specifically, after the final model is obtained, the predicted trajectory of the target in different unit time may be acquired using different feature sets as inputs according to the presence or absence of the prediction error. This step mainly involves the application of a predictive model and the determination of prediction errors. And under the condition that no prediction error occurs, only the second characteristic set input model of the corresponding unit time is needed to predict. And under the condition that the prediction error occurs, the second feature set and the third feature set of the corresponding unit time are input into the model together for prediction. For example, assuming that the second feature set of the first unit time is [0.1,0.2,0.3], the predicted trajectory of the second unit time may be [0.7,0.8], [0.6,0.9], and [0.5,0.8] after being input to the model without a prediction error. If the prediction of the second unit time is in error, the second feature set [0.15,0.25,0.35] of the second unit time and the third feature set [0.02] need to be input into the model together, and the predicted track of the third unit time can be obtained as [0.4,0.3], [0.5,0.4], [0.6,0.5].
In particular, we will get predicted trajectories of the target at different units of time. The accuracy of this predicted trajectory is affected by the prediction error, which, when it occurs, can be further improved by using the second feature set and the third feature set as inputs at the same time. This step is in the model application stage during the whole operation process, and aims to adjust the input data according to the actual situation so as to optimize the prediction result. For the whole invention, this step helps to realize more accurate track prediction, thereby improving the practicability and practical value of the prediction model.
Thus far, the technical solution of the present invention has been described in connection with the preferred embodiments shown in the drawings, but it is easily understood by those skilled in the art that the scope of protection of the present invention is not limited to these specific embodiments. Equivalent modifications and substitutions for related technical features may be made by those skilled in the art without departing from the principles of the present invention, and such modifications and substitutions will be within the scope of the present invention.
The foregoing description is only of the preferred embodiments of the invention and is not intended to limit the invention; various modifications and variations of the present invention will be apparent to those skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (10)

1. An intelligent motion recognition method based on the internet of things technology is characterized by comprising the following steps:
acquiring first original information of a target in each unit time through radio frequency identification equipment in the Internet of things, acquiring second original information of the target in each unit time through a global positioning system in the Internet of things, acquiring third original information of the target in each unit time through an infrared motion sensor in the Internet of things, and acquiring fourth original information of the target in each unit time through a laser scanner in the Internet of things;
extracting a first initial feature set based on the first original information, extracting a second initial feature set based on the second original information, extracting a third initial feature set based on the third original information, and extracting a fourth initial feature set based on the fourth original information;
the first initial feature set, the second initial feature set and the third initial feature set are spliced to obtain a first feature set, an enhanced mapping feature set of a fourth initial feature set for the first feature set is obtained through a self-attention mechanism, and the enhanced mapping feature set is spliced with the first feature set to obtain a second feature set;
configuring a long-term and short-term memory network;
inputting the second feature set into the long-short-period memory network for training to obtain a pre-training model capable of predicting a target motion track, inputting the second feature set obtained in a first unit time into the pre-training model to obtain a predicted track of the target in a second unit time, comparing the predicted track with an actual track to calculate a prediction error after the second unit time is over, taking the prediction error as a third feature set, and inputting the third feature set and the second feature set into the pre-training model for training to obtain a final model.
2. The intelligent motion recognition method based on the internet of things technology according to claim 1, further comprising: after the final model is obtained, inputting the second feature set obtained in the first unit time into the final model, so that a predicted track of a target in the second unit time can be obtained;
if the prediction error does not appear in the second unit time, inputting a second feature set obtained in the second unit time into a final model to obtain a predicted track of the target in the third unit time;
if the prediction error occurs, the second feature set and the third feature set obtained in the second unit time are input into a final model to obtain a predicted track of the target in the third unit time.
3. The intelligent motion recognition method based on the internet of things technology according to claim 2, wherein extracting a first initial feature set based on the first original information comprises:
performing one-time thermal encoding on the first original information to extract a first initial feature set;
extracting a second initial feature set based on the second original information includes using a min-max normalization process on the second original information to extract a second initial feature set;
extracting a third initial feature set based on the third initial information includes converting the third initial information from a time domain to a frequency domain using fourier transform, and then extracting a main frequency component in the frequency domain as the third initial feature set;
extracting the fourth initial feature set based on the fourth original information includes performing a Z-score normalization process on the fourth original information to extract the fourth initial feature set.
4. The method for intelligent motion recognition based on internet of things according to claim 3, wherein obtaining the first feature set comprises: and parallelly splicing the values of the first initial feature set, the second initial feature set and the third initial feature set at the same time to form a matrix, so as to obtain a first feature set.
5. The method for intelligent motion recognition based on the internet of things according to claim 4, wherein the enhanced mapping feature set is obtained by performing feature vector score calculation of query vectors, key vectors and value vectors through a self-attention mechanism of a transducer model.
6. The internet of things-based intelligent motion recognition method of claim 5, wherein the obtaining of the query vector, the key vector, and the value vector comprises: obtaining a query vector by multiplying a weight matrix by the first feature set, obtaining a key vector and a value vector by multiplying a weight matrix by the fourth initial feature set, wherein the weight matrix is randomly initialized based on the calculation flow of the transducer model.
7. The intelligent motion recognition method based on the internet of things technology according to claim 6, wherein configuring the long-term memory network comprises: adopting a multi-layer long-short-term memory network structure, and performing preliminary processing on the second feature set by the first layer network to generate intermediate state features;
the second layer network processes the intermediate state characteristics generated by the first layer to obtain high-level characteristics;
the third layer network converts the output of the second layer into the motion state prediction of the target;
during the training process, each layer of network updates network parameters through a time back propagation algorithm to minimize a loss function that measures the difference between the predicted target motion state and the true motion state.
8. The method for intelligent motion recognition based on the internet of things of claim 7, wherein obtaining a pre-training model capable of predicting a target motion trajectory comprises:
and when the second feature set is input into the long-short-term memory network for training, applying an early stop and regularization over-fitting prevention technology, wherein the output of the pre-training model is a set of coordinates, and obtaining a predicted track of the target after fitting.
9. The intelligent motion recognition method based on the internet of things technology according to claim 8, wherein comparing the predicted trajectory with an actual trajectory to calculate a prediction error comprises: and calculating the difference between the predicted track and the actual track at each moment in unit time, summing the squares of each difference, and finally dividing the sum by the total moment number to obtain a mean square error, wherein the mean square error is the predicted error.
10. The intelligent motion recognition method based on the internet of things technology according to claim 9, wherein inputting the third feature set and the second feature set into the pre-training model for training, and obtaining a final model comprises: and splicing the third feature set and the second feature set according to a time sequence to obtain a fusion feature set, inputting the fusion feature set into the pre-training model for training, dividing the fusion feature set into a training set and a verification set in the training process, performing parameter optimization by using the loss function and a random gradient descent strategy, dividing data into a plurality of batches by adopting a batch training mode, and training by using only one batch of data each time.
CN202310886760.6A 2023-07-19 2023-07-19 Intelligent motion recognition method based on Internet of things technology Active CN116907510B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310886760.6A CN116907510B (en) 2023-07-19 2023-07-19 Intelligent motion recognition method based on Internet of things technology

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310886760.6A CN116907510B (en) 2023-07-19 2023-07-19 Intelligent motion recognition method based on Internet of things technology

Publications (2)

Publication Number Publication Date
CN116907510A true CN116907510A (en) 2023-10-20
CN116907510B CN116907510B (en) 2024-08-09

Family

ID=88364266

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310886760.6A Active CN116907510B (en) 2023-07-19 2023-07-19 Intelligent motion recognition method based on Internet of things technology

Country Status (1)

Country Link
CN (1) CN116907510B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117841751A (en) * 2024-03-08 2024-04-09 沧州智慧城科技有限公司 Charging gun system and charging method based on intention recognition

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200309541A1 (en) * 2019-03-28 2020-10-01 Nexar Ltd. Localization and mapping methods using vast imagery and sensory data collected from land and air vehicles
CN113204718A (en) * 2021-04-22 2021-08-03 武汉大学 Vehicle track destination prediction method considering space-time semantics and driving state
US20220011122A1 (en) * 2020-07-09 2022-01-13 Beijing Tusen Weilai Technology Co., Ltd. Trajectory prediction method and device
CN114021660A (en) * 2021-11-10 2022-02-08 云从科技集团股份有限公司 Training method, device, medium and equipment for feature mining model
CN114065884A (en) * 2021-10-20 2022-02-18 国网浙江省电力有限公司杭州供电公司 Infrared electronic tag position prediction method combined with local attention mechanism
CN114529799A (en) * 2022-01-06 2022-05-24 浙江工业大学 Aircraft multi-target tracking method based on improved YOLOV5 algorithm
CN115048870A (en) * 2022-07-02 2022-09-13 南京航空航天大学 Target track identification method based on residual error network and attention mechanism
CN115376099A (en) * 2022-07-21 2022-11-22 长安大学 Track prediction method based on target local terminal and multi-head attention mechanism
CN115907154A (en) * 2022-11-22 2023-04-04 中冶南方都市环保工程技术股份有限公司 Industrial internet prediction method and system based on frequency domain and long-term and short-term feature fusion

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200309541A1 (en) * 2019-03-28 2020-10-01 Nexar Ltd. Localization and mapping methods using vast imagery and sensory data collected from land and air vehicles
US20220011122A1 (en) * 2020-07-09 2022-01-13 Beijing Tusen Weilai Technology Co., Ltd. Trajectory prediction method and device
CN113204718A (en) * 2021-04-22 2021-08-03 武汉大学 Vehicle track destination prediction method considering space-time semantics and driving state
CN114065884A (en) * 2021-10-20 2022-02-18 国网浙江省电力有限公司杭州供电公司 Infrared electronic tag position prediction method combined with local attention mechanism
CN114021660A (en) * 2021-11-10 2022-02-08 云从科技集团股份有限公司 Training method, device, medium and equipment for feature mining model
CN114529799A (en) * 2022-01-06 2022-05-24 浙江工业大学 Aircraft multi-target tracking method based on improved YOLOV5 algorithm
CN115048870A (en) * 2022-07-02 2022-09-13 南京航空航天大学 Target track identification method based on residual error network and attention mechanism
CN115376099A (en) * 2022-07-21 2022-11-22 长安大学 Track prediction method based on target local terminal and multi-head attention mechanism
CN115907154A (en) * 2022-11-22 2023-04-04 中冶南方都市环保工程技术股份有限公司 Industrial internet prediction method and system based on frequency domain and long-term and short-term feature fusion

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117841751A (en) * 2024-03-08 2024-04-09 沧州智慧城科技有限公司 Charging gun system and charging method based on intention recognition
CN117841751B (en) * 2024-03-08 2024-05-17 沧州智慧城科技有限公司 Charging gun system and charging method based on intention recognition

Also Published As

Publication number Publication date
CN116907510B (en) 2024-08-09

Similar Documents

Publication Publication Date Title
CN114384920B (en) Dynamic obstacle avoidance method based on real-time construction of local grid map
Yuan et al. A novel GRU-RNN network model for dynamic path planning of mobile robot
Zhao et al. A spatial-temporal attention model for human trajectory prediction.
Grigorescu et al. Neurotrajectory: A neuroevolutionary approach to local state trajectory learning for autonomous vehicles
Chen et al. Robot navigation with map-based deep reinforcement learning
CN113205466A (en) Incomplete point cloud completion method based on hidden space topological structure constraint
CN106714110A (en) Auto building method and system of Wi-Fi position fingerprint map
CN112148008B (en) Real-time unmanned aerial vehicle path prediction method based on deep reinforcement learning
KR102288933B1 (en) METHOD, APPARATUS AND SYSTEM FOR BUILDING OF precision road map
CN116907510B (en) Intelligent motion recognition method based on Internet of things technology
CN114708435B (en) Obstacle size prediction and uncertainty analysis method based on semantic segmentation
CN112857370A (en) Robot map-free navigation method based on time sequence information modeling
Xue et al. Fuzzy controller for autonomous vehicle based on rough sets
CN113219506A (en) Positioning method for multimode fusion seamless switching
Prokop et al. Neuro-heuristic pallet detection for automated guided vehicle navigation
CN114386599B (en) Method and device for training trajectory prediction model and trajectory planning
Lei et al. Digital twin‐based multi‐objective autonomous vehicle navigation approach as applied in infrastructure construction
CN110926470B (en) AGV navigation control method and system
Daass et al. Design of multi-sensor fusion architectures based on the covariance intersection algorithm—estimating calculation burdens
Gao et al. A prediction method of localizability based on deep learning
CN116429112A (en) Multi-robot co-location method and device, equipment and storage medium
Khan et al. Learning sample-efficient target reaching for mobile robots
CN115900725A (en) Path planning device, electronic device, storage medium and related method
CN115373383A (en) Autonomous obstacle avoidance method and device for garbage recovery unmanned boat and related equipment
Wang et al. Research on SLAM road sign observation based on particle filter

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant