WO2021184468A1 - 行为识别方法、装置、设备及介质 - Google Patents
行为识别方法、装置、设备及介质 Download PDFInfo
- Publication number
- WO2021184468A1 WO2021184468A1 PCT/CN2020/084694 CN2020084694W WO2021184468A1 WO 2021184468 A1 WO2021184468 A1 WO 2021184468A1 CN 2020084694 W CN2020084694 W CN 2020084694W WO 2021184468 A1 WO2021184468 A1 WO 2021184468A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- behavior
- data
- serialized
- behavior recognition
- result
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 57
- 238000007781 pre-processing Methods 0.000 claims abstract description 10
- 230000006399 behavior Effects 0.000 claims description 398
- 238000012549 training Methods 0.000 claims description 34
- 230000011218 segmentation Effects 0.000 claims description 14
- 230000007246 mechanism Effects 0.000 claims description 12
- 230000003945 visual behavior Effects 0.000 claims description 12
- 238000004590 computer program Methods 0.000 claims description 7
- 238000013507 mapping Methods 0.000 claims description 4
- 241001465754 Metazoa Species 0.000 description 77
- 239000000523 sample Substances 0.000 description 47
- 238000002372 labelling Methods 0.000 description 15
- 239000003814 drug Substances 0.000 description 14
- 230000015654 memory Effects 0.000 description 14
- 241000699670 Mus sp. Species 0.000 description 13
- 229940079593 drug Drugs 0.000 description 13
- 230000003542 behavioural effect Effects 0.000 description 11
- 238000010586 diagram Methods 0.000 description 10
- 238000012800 visualization Methods 0.000 description 10
- 241000699666 Mus <mouse, genus> Species 0.000 description 9
- 238000001514 detection method Methods 0.000 description 8
- 230000008569 process Effects 0.000 description 8
- 238000002474 experimental method Methods 0.000 description 6
- 230000003287 optical effect Effects 0.000 description 6
- 208000019901 Anxiety disease Diseases 0.000 description 5
- 238000013079 data visualisation Methods 0.000 description 4
- 230000000694 effects Effects 0.000 description 4
- 238000003058 natural language processing Methods 0.000 description 4
- 238000003062 neural network model Methods 0.000 description 4
- 238000012346 open field test Methods 0.000 description 4
- 238000012545 processing Methods 0.000 description 4
- 230000000306 recurrent effect Effects 0.000 description 4
- 230000036506 anxiety Effects 0.000 description 3
- 150000001875 compounds Chemical class 0.000 description 3
- 239000011159 matrix material Substances 0.000 description 3
- 238000005457 optimization Methods 0.000 description 3
- 230000002093 peripheral effect Effects 0.000 description 3
- 238000012827 research and development Methods 0.000 description 3
- 230000009154 spontaneous behavior Effects 0.000 description 3
- 238000012360 testing method Methods 0.000 description 3
- 238000012546 transfer Methods 0.000 description 3
- 230000001755 vocal effect Effects 0.000 description 3
- 230000006400 anxiety behaviour Effects 0.000 description 2
- 238000013528 artificial neural network Methods 0.000 description 2
- 238000007405 data analysis Methods 0.000 description 2
- 230000003247 decreasing effect Effects 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 230000018109 developmental process Effects 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000000644 propagated effect Effects 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 230000000630 rising effect Effects 0.000 description 2
- 238000005070 sampling Methods 0.000 description 2
- 230000002123 temporal effect Effects 0.000 description 2
- 208000024827 Alzheimer disease Diseases 0.000 description 1
- 206010002869 Anxiety symptoms Diseases 0.000 description 1
- 206010003805 Autism Diseases 0.000 description 1
- 208000020706 Autistic disease Diseases 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000002457 bidirectional effect Effects 0.000 description 1
- 230000036772 blood pressure Effects 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 125000004122 cyclic group Chemical group 0.000 description 1
- 238000013480 data collection Methods 0.000 description 1
- 238000013499 data model Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 238000001647 drug administration Methods 0.000 description 1
- 238000009509 drug development Methods 0.000 description 1
- 230000021061 grooming behavior Effects 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 239000003550 marker Substances 0.000 description 1
- 238000002483 medication Methods 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000006403 short-term memory Effects 0.000 description 1
- 238000004904 shortening Methods 0.000 description 1
- 238000007619 statistical method Methods 0.000 description 1
- 238000000528 statistical test Methods 0.000 description 1
- 208000024891 symptom Diseases 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/217—Validation; Performance evaluation; Active pattern learning techniques
Definitions
- the embodiments of the present application relate to the technical field of behavior recognition, for example, to a behavior recognition method, device, equipment, and medium.
- EPM Elevated Plus Maze
- OFT Open Field Test
- OFT devices generally have a square box with a camera on top. This square field can be divided into a central area and a peripheral area. If the mice move more in the central area, the anxiety level of the mice is relatively lower.
- anxiety symptoms are manifested in various forms, and there are also different types of medications and methods of medication for different symptoms of patients.
- the embodiments of the present application provide a behavior recognition method, device, equipment, and medium, so as to improve the accuracy of behavior recognition, and further improve the efficiency and accuracy of detection results determined based on behavior recognition.
- an embodiment of the present application provides a behavior recognition method, including:
- an embodiment of the present application also provides a behavior recognition device, including:
- the serialized data acquisition module is configured to acquire original behavior data, preprocess the original behavior data, and obtain serialized behavior data;
- the output result obtaining module is configured to input the serialized behavior data into the pre-trained behavior recognition model to obtain the output result of the behavior recognition model;
- the recognition result output module is configured to generate a behavior recognition result according to the output result, and output the behavior recognition result.
- an embodiment of the present application also provides a computer device, and the device includes:
- One or more processors are One or more processors;
- Storage device for storing one or more programs
- the one or more processors implement the behavior recognition method as provided in any embodiment of the present application.
- an embodiment of the present application also provides a computer-readable storage medium, and the computer-readable storage medium stores a computer program.
- the computer program is executed by a processor, the implementation of the Method of behavior recognition.
- FIG. 1a is a flowchart of a behavior recognition method provided by Embodiment 1 of the present application.
- FIG. 1b is a schematic diagram of a glimpse diagram of animal behavior provided in Example 1 of the present application.
- FIG. 2 is a flowchart of a behavior recognition method provided by Embodiment 2 of the present application.
- Fig. 3a is a schematic structural diagram of a columnized animal behavior recognition and classification system provided in the third embodiment of the present application.
- FIG. 3b is a flowchart of a behavior recognition method provided by Embodiment 3 of the present application.
- FIG. 4 is a schematic structural diagram of a behavior recognition device provided by Embodiment 4 of the present application.
- FIG. 5 is a schematic structural diagram of a computer device provided by Embodiment 5 of the present application.
- FIG. 1a is a flowchart of a behavior recognition method provided in Embodiment 1 of the present application. This embodiment can be applied to the situation when the original behavior data is used for behavior recognition, for example, when it is used to identify the animal behavior data.
- the method may be executed by a behavior recognition device, and the behavior recognition device may be implemented in a software and/or hardware manner.
- the behavior recognition device may be configured in a computer device. As shown in Figure 1a, the method includes:
- the original behavior data may be behavior data collected through different collection methods.
- the original behavior data may include image data, video data, physiological signal data, etc., where the physiological signal data may be data such as heart rate data, blood pressure data, and brain electrical data.
- the original behavior data may include multi-modal signal acquisition data.
- the animal's original behavior data is a time series, that is, the mode state in the animal behavior data changes with time. Therefore, it is necessary to consider the dynamics of time when distinguishing different behavior patterns of animals.
- preprocessing the multi-modal original behavior data into serialized behavior data may be: fusing the multi-modal original behavior data, and then segmenting the fused data to obtain multiple pieces of serialized behavior data .
- the original behavior data includes multi-modal signal acquisition data
- the step of preprocessing the original behavior data to obtain serialized behavior data includes: processing each signal according to the timestamp in the signal acquisition data.
- the signal acquisition data of each mode is aligned to obtain the alignment behavior data;
- the alignment behavior data is segmented using the set segmentation algorithm to obtain the segmentation behavior data;
- the segmentation behavior data is serialized and mapped to obtain the serialized behavior data.
- the multi-modal signal acquisition data may include image data, video data, and various physiological signal data.
- the obtained alignment behavior data can be segmented to obtain multiple segments of segmentation behavior data, and then the segmentation behavior data can be serialized and mapped to obtain Multiple pieces of serialized behavior data.
- aligning the signal acquisition data of each mode since the acquisition frequency of the signal acquisition data of each mode may be different, the signal acquisition data of each mode cannot be aligned at each time, and it is necessary to align the signal acquisition data of each mode.
- Part of the signal acquisition data is resampled to make the sampling frequency of the signal acquisition data of each mode consistent after re-sampling, to ensure that the signal acquisition data of each mode can be aligned at each moment, that is, to ensure that different equipment is used for acquisition
- the received signal collection data needs to ensure that there is corresponding data at the same time. For example, the animal’s heart rate should be increased at the beginning of the recording of the animal’s running, and the rising edge of the running speed curve must be matched to the animal’s running speed to a certain extent. The rising edge of the ECG signal curve is aligned.
- a dynamic time warping (Dynamic Time Warping, DTW) method may be used to segment behavior data with similar patterns.
- the DTW method can ensure the continuity of the time series in the time dimension, and find the most similar patterns in the data through the optimization search method, abstract the differences between the patterns into the optimal distance, and make the same through the difference in the data distance
- the data of the patterns are clustered together, and the data of different patterns are separated to ensure that the similarity within the category is high (small distance), and the similarity between the categories is low (large distance).
- DTW Dynamic Time Warping
- the optional mapping character table can choose the American Standard Code for Information Interchange (ASCII).
- ASCII is the most common information exchange standard and is a computer coding system based on the Latin alphabet. It is mainly used to display modern English and other Western European languages have defined a total of 128 characters so far. Some of these 128 characters are mapped to similar data to form a series of serialized animal behavior languages to obtain multiple pieces of serialized behavior data.
- the output result of the behavior recognition model may be each behavior and the probability corresponding to each behavior.
- the behavior output by the behavior recognition model may be a single behavior or a complex behavior.
- the behavior output by the behavior recognition model may be walking or It may be walking and probes.
- the output result may be directly output as the behavior recognition result, and the output result may be counted to generate a visual behavior statistics result, and the generated visual behavior statistics may be added for output.
- the output results can be counted, and the visual behavior statistical results can be generated for output. The output of the visual behavior statistical results can enable the inspector to understand animal behaviors more vividly and intuitively.
- the step of generating a behavior recognition result according to the output result includes: determining characteristic information corresponding to each behavior according to the output result, generating a visual behavior statistics result based on the characteristic information corresponding to each behavior, and combining The visual behavior statistics result is used as the behavior recognition result.
- the visualized behavior statistical results may include the visualization of the statistical results of serialized animal behavior data and the visualization of the results of animal behavior recognition and classification.
- a certain statistical method is used to describe the law of each behavior in the output result, and the statistical result of the visual serialized animal behavior data is obtained.
- the proportion of a certain type of special behavior in all behaviors, or after performing different operations on the detection target the difference before and after the detection target behavior, for example, after using different drugs on the mouse, the difference between the mouse before and after different behaviors difference.
- the difference in behavior will not only be reflected in the number or time of different behaviors, but also in the state transition between different behaviors. For example, in mice of different genetically modified strains, there is no significant difference in overall behavior, that is, there is no difference in the number and time of each behavior, but the transfer patterns between the different inherent behaviors are quite different.
- strain 1 mouse and strain 2 mice both probe 10 times and sniff 5 times, but it is possible that strain 1 mouse and strain 2 mouse probe and sniff The transfer method between probes is different, and this is statistically significant.
- Animal behavior recognition and classification result visualization not only need to visualize the global animal behavior in the time dimension, but also need to visualize the local, that is, a single behavior.
- the overall visualization of animal behaviors uses behavior maps, and the local visualization of animal behaviors uses glimpses.
- the visualized behavior statistical results include behavioral glimpses and/or behavioral maps.
- Fig. 1b is a schematic diagram of a glimpse of animal behavior provided in Example 1 of the present application.
- the glimpse of animal behavior is composed of upper and lower parts.
- the upper part is the behavioral glimpse graph
- the lower part is the different behavior recognition probabilities of the behavior recognition model.
- the glimpse map samples a certain behavior in the time period at a certain time interval, and after removing the background, it is horizontally arranged and spliced to efficiently display the behavior sequence.
- the recognition probability directly describes the recognition results of the current behavior of the model.
- the model recognizes that the mouse has a 70% probability of a Twisting behavior, a 25% probability of an Observing behavior, and about 5%
- the probability is the grooming behavior. Obviously, the model is more accurate in the recognition of the current behavior, and the glimpse map can accurately correspond to the recognition probability.
- the abscissa is time, and the ordinate is different behavior types. Different colors can be used to represent different behaviors, and the transparency of the color represents the probability of the model recognizing the behavior. In each column of the behavior map, multiple color bars are allowed. This data display form can effectively represent high-dimensional animal behavior data. For describing different behavior patterns that occur at the same time, and the probability of occurrence of each different behavior Compared with the traditional behavioral graph without transparency description, it contains a larger amount of behavioral information.
- the original behavior data is obtained, and the original behavior data is preprocessed to obtain serialized behavior data; the serialized behavior data is input into a pre-trained behavior recognition model to obtain the output result of the behavior recognition model; according to the output The result generates behavior recognition results, and outputs the behavior recognition results.
- serialized behavior data for behavior recognition the information in the original behavior data is fully utilized, the accuracy of behavior recognition is improved, and the determination based on behavior recognition is improved. The efficiency and accuracy of the test results.
- Fig. 2 is a flowchart of a behavior recognition method provided in the second embodiment of the present application.
- the method includes:
- the sample serialized data may be serialized behavior data obtained after preprocessing the sample behavior data. It can be understood that the detection target to which the sample behavior data belongs and the detection target to which the original behavior data belongs are the same type of detection target. Exemplarily, if the original behavior data is the behavior data of mice, the sample behavior data should also be the behavior data of mice.
- the label corresponding to the sample serialized data is realized by manual labeling.
- the sample behavior data can be preprocessed to obtain the sample serialized data, and the sample serialized data and sample behavior data can be played in chronological order.
- the experimenter can observe the sample behavior data to determine the behaviors that need to be marked, and then serialize the data in the sample. Label it in, and get the label corresponding to the sample serialized data.
- the manner of obtaining sample serialized data from sample behavior data can refer to the manner of obtaining serialized behavior data from original behavior data in the foregoing embodiment, which will not be repeated here.
- the label corresponding to the sample serialized data includes at least one of a single label, a multi-label, and a language description label.
- the manually marked behavior tags may include multiple forms, such as single tags, multiple tags, language description tags, refined tags, and so on.
- a single label is the most traditional form of labeling different behaviors, which uses a single word to describe a sequence of different behaviors as a category label; multi-label is to mark different behaviors as multiple labels.
- Multi-labeling takes into account the high-dimensional properties of animal behavior, that is, at the same time or within the same period of time, there will be situations where the behaviors occur at the same time (for example, a mouse may be sniffing while walking), and multiple words need to be used to describe The behaviors of animals at the same time; verbal description tags mark different behaviors in the form of verbal description.
- the spontaneous behavior of animals occupies a high proportion of all behaviors, and these spontaneous behaviors often cannot be defined with a simple word or a few descriptive nouns. For example, a mouse stops and raises its head while walking. The right front scratched his right ear.
- S220 Use the training sample data to train the pre-built behavior recognition model to obtain the trained behavior recognition model.
- the pre-built behavior recognition model can be
- the pre-built behavior recognition model can use common models in Natural Language Processing (NLP) tasks, such as sequence-to-sequence (seq2seq) network based on attention mechanism, Bidirectional Encoder Representations from Transformers (BERT) models, etc.
- NLP Natural Language Processing
- the seq2seq model is a commonly used codec model in the NLP field. It consists of an Encoder part and a Decoder part. Natural language sequences have temporal dynamics. Therefore, to encode natural language sequences in the time dimension, you need to use Recurrent Neural Network (RNN). In the seq2seq model, use Long Short Term Memory (LSTM) to input sequences. Encoding, using RNN to decode the features learned from the input sequence, LSTM has an excellent effect in solving the long-term dependence of the time sequence.
- RNN Recurrent Neural Network
- the seq2seq model that introduces the attention mechanism is used as the behavior recognition model, and the attention module is used for the decoding of semantic features, replacing the RNN in the traditional seq2seq model.
- the attention mechanism When humans deal with natural language processing tasks such as Chinese-English translation tasks, they will selectively pay attention to the keywords in a sentence. This mechanism is called the attention mechanism.
- the attention mechanism In the model, by increasing the attention weight of keywords and decreasing the attention weight of non-keywords, an attention mechanism similar to that of humans can be obtained. In the process of animal behavior identification and classification, similar conclusions also exist.
- the training performance in the process of training the behavior recognition model, can be visualized, so that the inspector can understand the training level of the behavior recognition model.
- the training loss Training Loss
- the recognition accuracy Precision
- recall recall
- Confusion Matrix confusion matrix
- the training loss directly describes the optimization of the model. The smaller the training loss, the better the optimization effect of the model.
- the decreasing law of training loss can play a certain guiding role.
- training loss decreases with time, indicating that the model is still optimizing learning; training loss increases with time, indicating that the model has not learned useful data laws; training loss oscillating, indicating that the current model has reached the best performance, and you want to continue
- the parameters need to be adjusted to increase the recognition effect of the model.
- the recognition accuracy and recall rate directly reflect the recognition effect of the current test data.
- the accuracy rate describes the correct proportion of the model in a certain category of all data, that is, the ability to identify and classify this type of data in all categories; the recall rate describes the proportion of the model that is correctly judged in a certain category of data, That is, the degree of discrimination of the model for this type of data.
- the confusion matrix is the most basic indicator of the machine learning model, which directly describes the degree of correspondence between the label of the data and the prediction result of the model.
- S250 Generate a behavior recognition result according to the output result, and output the behavior recognition result.
- the embodiment of the application generates training sample data according to the sample serialized data and the label corresponding to the sample serialized data by obtaining the sample serialized data and the label corresponding to the sample serialized data; and uses the training sample data to train the pre-built behavior recognition model , Obtain a well-trained behavior recognition model.
- the behavior recognition model makes full use of the temporal characteristics of the behavior data when learning behavior characteristics, and improves the behavior recognition model Accuracy of behavior recognition.
- the behavior recognition method can be executed by a serialized animal behavior recognition and classification system.
- the system first serializes the data acquired by the animal behavior collection device, then performs serialized animal behavior data labeling, manually labeling the data tags, and finally uses the labeled serialized animal behavior data for training
- the seq2seq cyclic neural network model with the attention mechanism is introduced to obtain the animal behavior sequence corresponding to the behavior label.
- the system is suitable for different animal behavior recognition and classification tasks and big data analysis of animal behaviors, and automatically obtains the behaviors of experimental attention and the inherent transfer mode of behaviors, and improves the efficiency of animal behavior data analysis.
- Fig. 3a is a schematic structural diagram of a serialized animal behavior recognition and classification system provided in the third embodiment of the present application.
- the serialized animal behavior recognition and classification system includes: an animal behavior data serialization unit, serialization There are six parts: data labeling unit, seq2seq model training unit, behavior sequence recognition and classification unit, data visualization unit, and control host.
- the animal behavior data serialization unit includes a data acquisition module and a data serialization module.
- the data acquisition module acquires image, video and physiological signal data of animal behaviors.
- the data serialization module is responsible for discretizing the time data acquired by the data acquisition module. Clustering and coding generates serialized data of animal behavior.
- the serialized data labeling unit includes a data playing module and a data labeling module.
- the data playing module maps the clustered data in the data serialization module to the original data space for playing and visualization for observation by experimenters.
- the experimenters used the data labeling module to manually label the specific animal behavior sequence according to the pattern of the observed behavior data (image, video, and physiological signal).
- the seq2seq model training unit includes a sequence data preprocessing module and a seq2seq model training module.
- the sequence data preprocessing module performs different preprocessing of the data according to different data labeling forms.
- the seq2seq model training module obtains the preprocessed data training and introduces attention Mechanism of seq2seq recurrent neural network model.
- the behavior sequence recognition and classification unit includes the seq2seq model recognition and classification module and the recognition data segmentation and labeling module.
- the seq2seq model recognition and classification module inputs the serialized animal behavior data that needs to be recognized into the seq2seq model to realize the automatic recognition and classification of animal behaviors, and the recognition data segmentation
- the marking module obtains the identified and classified data tags, maps the original animal behavior data and the serialized data tags to the same time dimension, and at the same time divides similar data into the same folder according to the set rules.
- the data visualization unit includes a data statistics module and a data drawing module.
- the data statistics module performs statistics on the segmented data according to set rules, and explores the laws and differences in animal behavior data.
- the data drawing module draws animal behavior data and statistical results.
- the control host is the basis for the operation of the entire algorithm. It supports the collection of animal behavior data, the storage and recall of a large amount of animal behavior data, provides hardware computing power for data serialization, and provides seq2seq model training and animal behavior recognition and classification Provides a parallel computing graphics processing unit, which speeds up the training and verification of the model.
- the control host also provides experimenters with an interactive interface that can use this method. The high-performance control host saves experimenters a lot of time for adjusting model parameters and statistical test data, ensuring the efficiency of data model operation, and shortening the experiment cycle.
- Fig. 3b is a flowchart of a behavior recognition method provided in the third embodiment of the present application. As shown in Figure 3b, the method includes:
- S310 Use the control host to collect and process animal behavior data to obtain serialized data.
- the collection of animal behavior data can be divided into two ways. One is to connect the sensor that collects animal behavior to the control host, and the animal behavior data is collected through the data collection module in the control host; the other is to collect the offline device.
- the past animal behavior data is loaded into the system by connecting to the hard disk. After collecting animal behavior data, serialize the loaded data.
- the serialization process is divided into three steps: multi-modal data alignment, animal behavior data segmentation, and animal behavior data serialization mapping. Among them, a more detailed solution for serializing data can be found in the foregoing embodiment, which will not be repeated here.
- the serialized data and animal raw data can be played in chronological order by the data playback module in the serialized data marking unit, and the experimenter can observe the behaviors that need to be marked, and use the data marking module to set a specific behavior sequence Manually mark the label.
- Manually labeled behavior labels can include multiple forms: a single label form for different behaviors, a multi-label form for different behaviors, a verbal description form for different behaviors, and a fine-labeled form for focused behaviors.
- the final recognition and classification results are also different due to different data labeling formats.
- the data is labeled with a single label for different behaviors, and the result of recognition and classification is a single label for different behaviors.
- the results of recognition and classification may have multiple label forms, or a single label may appear.
- serialization of behavioral data it is equivalent to compressing the data. The redundant information of the data will be removed while the main features are retained as much as possible, and the noise of the data will be reduced.
- manual marking because there is a certain compound ratio of compound behaviors, for example, a mouse probes at the same time while walking, the possible walking component accounts for 90%, the probe behavior accounts for 5%, and other small behaviors.
- the markers here can only mark the more obvious behaviors they see.
- the data marker will mark the current behavior based on their subjective judgment, and this subjectivity also has a time dynamic, and it is impossible to accurately judge the proportion of the behavior. It is possible to mark as walking, probe, or compound behaviors such as walking, probe, tail-wagging, etc.
- the model will extract the most obvious and credible results in the data as the output, which is often reflected in the expectations of the data. Therefore, in this example, it is possible that 90% of the data is a single walk, 5% of the data is walking and probes, and the remaining 5% may be walking, probes, and sniffing. Behavior that is difficult to judge with the naked eye.
- use different behavior language description forms to label the data.
- the results of recognition and classification are basically the same as those described in the previous seq2seq model training unit.
- the animal behavior language is translated into the language defined by the data tagger, and the animal behavior sequence is obtained. describe.
- the data visualization unit includes: visualization of neural network training performance results, visualization of statistical results of serialized animal behavior data, and visualization of animal behavior recognition and classification results.
- visualization of neural network training performance results visualization of statistical results of serialized animal behavior data
- visualization of animal behavior recognition and classification results visualization of animal behavior recognition and classification results.
- This embodiment provides a universal animal behavior recognition and classification system, which serializes the data acquired by the animal behavior collection device to fuse multi-modal data while retaining the continuity of behavior data in the time dimension; use The seq2seq recurrent neural network model that introduces the attention mechanism extracts high-dimensional semantic features in behavior sequence data, and decodes the features into the artificially labeled behavior label space, effectively extracting low-dimensional information in high-dimensional behavior data while retaining animals High-dimensional structure of behavior.
- artificially marked behavior labels can contain various forms, which greatly enrich the description indicators of animal behavior, provide more reference data for drug research and development, and improve the efficiency and accuracy of drug efficacy testing.
- FIG. 4 is a schematic structural diagram of a behavior recognition device provided by Embodiment 4 of the present application.
- the behavior recognition device can be implemented in software and/or hardware.
- the behavior recognition device can be configured in a computer device.
- the device includes a serialized data acquisition module 410, an output result acquisition module 420, and a recognition result output module 430, where:
- the serialized data acquisition module 410 is configured to acquire original behavior data, preprocess the original behavior data, and obtain serialized behavior data;
- the output result obtaining module 420 is configured to input the serialized behavior data into the pre-trained behavior recognition model to obtain the output result of the behavior recognition model;
- the recognition result output module 430 is configured to generate a behavior recognition result according to the output result, and output the behavior recognition result.
- the original behavior data is obtained by the serialized data acquisition module, and the original behavior data is preprocessed to obtain the serialized behavior data; the output result acquisition module inputs the serialized behavior data into the pre-trained behavior recognition model to obtain The output result of the behavior recognition model; the recognition result output module generates behavior recognition results based on the output results, and outputs the behavior recognition results.
- serialized behavior data for behavior recognition it makes full use of the information in the original behavior data to improve behavior
- the accuracy of recognition further improves the efficiency and accuracy of detection results determined based on behavior recognition.
- the original behavior data includes multi-modal signal acquisition data
- the serialized data acquisition module 410 can be used for:
- the segmentation behavior data is serialized and mapped to obtain the serialized behavior data.
- the device further includes a model training module for:
- the pre-built behavior recognition model is a sequence-to-sequence network based on the attention mechanism.
- the label corresponding to the sample serialized data includes at least one of a single label, a multi-label, and a language description label.
- the recognition result output module 430 may be used to:
- the characteristic information corresponding to each behavior is determined, and the visual behavior statistics result is generated based on the characteristic information corresponding to each behavior, and the visual behavior statistics result is used as the behavior recognition result.
- the visual behavior statistics result includes a behavior glimpse graph and/or a behavior graph.
- the behavior recognition device provided in the embodiment of the present application can execute the behavior recognition method provided in any embodiment of the present application, and has functional modules and beneficial effects corresponding to the execution method.
- FIG. 5 is a schematic structural diagram of a computer device provided by Embodiment 5 of the present application.
- Figure 5 shows a block diagram of an exemplary computer device 512 suitable for implementing embodiments of the present application.
- the computer device 512 shown in FIG. 5 is only an example.
- the computer device 512 is represented in the form of a general-purpose computing device.
- the components of the computer device 512 may include: one or more processors 516, a system memory 528, and a bus 518 connecting different system components (including the system memory 528 and the processor 516).
- the bus 518 represents one or more of several types of bus structures, including a memory bus or a memory controller, a peripheral bus, a graphics acceleration port, a processor 516, or a local bus using any bus structure among multiple bus structures.
- these architectures can include industry standard architecture (ISA) bus, microchannel architecture (MAC) bus, enhanced ISA bus, Video Electronics Standards Association (VESA) local bus, and Peripheral Component Interconnect (PCI) bus.
- ISA industry standard architecture
- MAC microchannel architecture
- VESA Video Electronics Standards Association
- PCI Peripheral Component Interconnect
- Computer device 512 typically includes a variety of computer system readable media. These media may be any available media that can be accessed by the computer device 512, including volatile and nonvolatile media, removable and non-removable media.
- the system memory 528 may include computer system readable media in the form of volatile memory, such as random access memory (RAM) 530 and/or cache memory 532.
- the computer device 512 may include other removable/non-removable, volatile/non-volatile computer system storage media.
- the storage device 534 may be used to read and write non-removable, non-volatile magnetic media (not shown in FIG. 5, usually referred to as a "hard drive").
- a disk drive for reading and writing to removable non-volatile disks such as "floppy disks”
- a removable non-volatile disk such as CD-ROM, DVD-ROM
- other optical media read and write optical disc drives.
- each drive may be connected to the bus 518 through one or more data medium interfaces.
- the memory 528 may include at least one program product, and the program product has a set of (for example, at least one) program modules, and these program modules are configured to perform the functions of the embodiments of the present application.
- a program/utility tool 540 having a set of (at least one) program module 542 may be stored in, for example, the memory 528.
- Such program module 542 may include an operating system, one or more application programs, other program modules, and program data. Each of the examples or some combination may include the realization of a network environment.
- the program module 542 generally executes the functions and/or methods in the embodiments described in this application.
- the computer device 512 can also communicate with one or more external devices 514 (such as a keyboard, pointing device, display 524, etc.), and can also communicate with one or more devices that enable a user to interact with the computer device 512, and/or communicate with Any device (such as a network card, modem, etc.) that enables the computer device 512 to communicate with one or more other computing devices. Such communication can be performed through an input/output (I/O) interface 522.
- the computer device 512 may also communicate with one or more networks (for example, a local area network (LAN), a wide area network (WAN), and/or a public network, such as the Internet) through the network adapter 520.
- networks for example, a local area network (LAN), a wide area network (WAN), and/or a public network, such as the Internet
- the network adapter 520 communicates with other modules of the computer device 512 through the bus 518. It should be understood that although not shown in the figure, other hardware and/or software modules may be used in conjunction with the computer device 512, which may include: microcode, device drivers, redundant processing units, external disk drive arrays, RAID systems, tape drives, and data Backup storage system, etc.
- the processor 516 executes various functional applications and data processing by running programs stored in the system memory 528, for example, to implement the behavior recognition method provided in the embodiment of the present application, the method includes:
- processor may also implement the technical solution of the behavior recognition method provided in any embodiment of the present application.
- the sixth embodiment of the present application also provides a computer-readable storage medium on which a computer program is stored.
- the behavior recognition method provided in the embodiment of the present application is implemented, and the method may include:
- the computer program stored thereon can operate in the above method, and can also perform related operations of the behavior recognition method provided in any embodiment of the present application.
- the computer storage medium of the embodiment of the present application may adopt any combination of one or more computer-readable media.
- the computer-readable medium may be a computer-readable signal medium or a computer-readable storage medium.
- the computer-readable storage medium may be, for example, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or a combination of any of the above.
- Examples of computer-readable storage media may include: electrical connections with one or more wires, portable computer disks, hard disks, random access memory (RAM), read-only memory (ROM), erasable Type programmable read only memory (EPROM or flash memory), optical fiber, portable compact disk read only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
- the computer-readable storage medium can be any tangible medium that contains or stores a program, and the program can be used by or in combination with an instruction execution system, apparatus, or device.
- the computer-readable signal medium may include a data signal propagated in baseband or as a part of a carrier wave, and computer-readable program code is carried therein. This propagated data signal can take many forms, and can include electromagnetic signals, optical signals, or any suitable combination of the foregoing.
- the computer-readable signal medium may also be any computer-readable medium other than the computer-readable storage medium.
- the computer-readable medium may send, propagate, or transmit the program for use by or in combination with the instruction execution system, apparatus, or device .
- the program code contained on the computer-readable medium can be transmitted by any suitable medium, which may include wireless, wire, optical cable, RF, etc., or any suitable combination of the above.
- the computer program code used to perform the operations of this application can be written in one or more programming languages or a combination thereof.
- the programming languages include object-oriented programming languages—such as Java, Smalltalk, C++, and also conventional Procedural programming language-such as "C" language or similar programming language.
- the program code can be executed entirely on the user's computer, partly on the user's computer, executed as an independent software package, partly on the user's computer and partly executed on a remote computer, or entirely executed on the remote computer or server.
- the remote computer can be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or it can be connected to an external computer (for example, using an Internet service provider to pass Internet connection).
- LAN local area network
- WAN wide area network
- Internet service provider for example, using an Internet service provider to pass Internet connection.
Landscapes
- Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (10)
- 一种行为识别方法,包括:获取原始行为数据,对所述原始行为数据进行预处理,得到序列化行为数据;将所述序列化行为数据输入至预先训练好的行为识别模型中,获得所述行为识别模型的输出结果;根据所述输出结果生成行为识别结果,并将所述行为识别结果进行输出。
- 根据权利要求1所述的方法,其中,所述原始行为数据包括多模态的信号采集数据;所述对所述原始行为数据进行预处理,得到序列化行为数据的步骤,包括:根据所述信号采集数据中的时间戳对每个模态的信号采集数据进行数据对齐,得到对齐行为数据;使用设定分割算法对所述对齐行为数据进行分割,得到分割行为数据;将所述分割行为数据进行序列化映射,得到所述序列化行为数据。
- 根据权利要求1所述的方法,其中,在所述将所述序列化行为数据输入至预先训练好的行为识别模型中的步骤之前,所述方法还包括:获取样本序列化数据以及所述样本序列化数据对应的标签,根据所述样本序列化数据以及所述样本序列化数据对应的标签生成训练样本数据;使用所述训练样本数据对预先构建的行为识别模型进行训练,得到训练好的行为识别模型。
- 根据权利要求3所述的方法,其中,所述预先构建的行为识别模型为基于注意力机制的序列到序列网络。
- 根据权利要求3所述的方法,其中,所述样本序列化数据对应的标签包括单一标签、多标签以及语言描述标签中的至少一种。
- 根据权利要求1所述的方法,其中,所述根据所述输出结果生成行为识别结果的步骤,包括:根据所述输出结果确定每个行为对应的特征信息,基于每个行为对应的特征信息生成可视化行为统计结果,并将所述可视化行为统计结果作为所述行为识别结果。
- 根据权利要求6所述的方法,其中,所述可视化行为统计结果包括行为掠影图和行为图谱中至少之一。
- 一种行为识别装置,包括:序列化数据获取模块,被配置为获取原始行为数据,对所述原始行为数据进行预处理,得到序列化行为数据;输出结果获取模块,被配置为将所述序列化行为数据输入至预先训练好的行为识别模型中,获得所述行为识别模型的输出结果;识别结果输出模块,被配置为根据所述输出结果生成行为识别结果,并将所述行为识别结果进行输出。
- 一种计算机设备,所述设备包括:一个或多个处理器;存储装置,用于存储一个或多个程序;当所述一个或多个程序被所述一个或多个处理器执行,使得所述一个或多个处理器实现如权利要求1-7中任一所述的行为识别方法。
- 一种计算机可读存储介质,所述计算机可读存储介质上存储有计算机程序,所述计算机程序被处理器执行时实现如权利要求1-7中任一所述的行为识别方法。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010190264.3A CN111325289A (zh) | 2020-03-18 | 2020-03-18 | 一种行为识别方法、装置、设备及介质 |
CN202010190264.3 | 2020-03-18 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021184468A1 true WO2021184468A1 (zh) | 2021-09-23 |
Family
ID=71173397
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/084694 WO2021184468A1 (zh) | 2020-03-18 | 2020-04-14 | 行为识别方法、装置、设备及介质 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN111325289A (zh) |
WO (1) | WO2021184468A1 (zh) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114461468A (zh) * | 2022-01-21 | 2022-05-10 | 电子科技大学 | 一种基于人工神经网络的微处理器应用场景识别方法 |
CN114817635A (zh) * | 2022-04-15 | 2022-07-29 | 阿里巴巴达摩院(杭州)科技有限公司 | 行为识别方法、行为识别装置、视频检测方法及存储介质 |
CN115115967A (zh) * | 2022-05-13 | 2022-09-27 | 清华大学 | 模式生物的视频动作分析方法、装置、设备及介质 |
CN115204753A (zh) * | 2022-09-14 | 2022-10-18 | 深圳市深信信息技术有限公司 | 一种智慧农贸场所行为监测方法、系统及可读存储介质 |
CN116011633A (zh) * | 2022-12-23 | 2023-04-25 | 浙江苍南仪表集团股份有限公司 | 区域燃气用量预测方法、系统、设备及物联网云平台 |
CN116030272A (zh) * | 2023-03-30 | 2023-04-28 | 之江实验室 | 一种基于信息抽取的目标检测方法、系统和装置 |
CN116467500A (zh) * | 2023-06-15 | 2023-07-21 | 阿里巴巴(中国)有限公司 | 数据关系识别、自动问答、查询语句生成方法 |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111883252A (zh) * | 2020-07-29 | 2020-11-03 | 济南浪潮高新科技投资发展有限公司 | 一种婴儿自闭症辅助诊断方法、装置、设备和存储介质 |
CN112057079B (zh) * | 2020-08-07 | 2022-07-29 | 中国科学院深圳先进技术研究院 | 一种基于状态与图谱的行为量化方法和终端 |
CN113705296A (zh) * | 2021-03-11 | 2021-11-26 | 腾讯科技(深圳)有限公司 | 生理电信号分类处理方法、装置、计算机设备和存储介质 |
CN113255597B (zh) * | 2021-06-29 | 2021-09-28 | 南京视察者智能科技有限公司 | 一种基于transformer的行为分析方法、装置及其终端设备 |
CN113989728A (zh) * | 2021-12-06 | 2022-01-28 | 北京航空航天大学 | 一种动物行为分析方法、装置及电子设备 |
CN114140682B (zh) * | 2021-12-14 | 2024-09-06 | 湖南师范大学 | 一种基于增强连接时序分类网络的步态识别方法 |
CN114241376A (zh) * | 2021-12-15 | 2022-03-25 | 深圳先进技术研究院 | 行为识别模型训练和行为识别方法、装置、系统及介质 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8463721B2 (en) * | 2010-08-05 | 2013-06-11 | Toyota Motor Engineering & Manufacturing North America, Inc. | Systems and methods for recognizing events |
CN110276259A (zh) * | 2019-05-21 | 2019-09-24 | 平安科技(深圳)有限公司 | 唇语识别方法、装置、计算机设备及存储介质 |
US20190294881A1 (en) * | 2018-03-22 | 2019-09-26 | Viisights Solutions Ltd. | Behavior recognition |
US10503967B2 (en) * | 2014-11-21 | 2019-12-10 | The Regents Of The University Of California | Fast behavior and abnormality detection |
CN110765939A (zh) * | 2019-10-22 | 2020-02-07 | Oppo广东移动通信有限公司 | 身份识别方法、装置、移动终端及存储介质 |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106056043B (zh) * | 2016-05-19 | 2019-07-30 | 中国科学院自动化研究所 | 基于迁移学习的动物行为识别方法和装置 |
CN108764176A (zh) * | 2018-05-31 | 2018-11-06 | 郑州云海信息技术有限公司 | 一种动作序列识别方法、系统及设备和存储介质 |
CN110689041A (zh) * | 2019-08-20 | 2020-01-14 | 陈羽旻 | 一种多目标行为动作识别预测方法、电子设备及存储介质 |
-
2020
- 2020-03-18 CN CN202010190264.3A patent/CN111325289A/zh active Pending
- 2020-04-14 WO PCT/CN2020/084694 patent/WO2021184468A1/zh active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8463721B2 (en) * | 2010-08-05 | 2013-06-11 | Toyota Motor Engineering & Manufacturing North America, Inc. | Systems and methods for recognizing events |
US10503967B2 (en) * | 2014-11-21 | 2019-12-10 | The Regents Of The University Of California | Fast behavior and abnormality detection |
US20190294881A1 (en) * | 2018-03-22 | 2019-09-26 | Viisights Solutions Ltd. | Behavior recognition |
CN110276259A (zh) * | 2019-05-21 | 2019-09-24 | 平安科技(深圳)有限公司 | 唇语识别方法、装置、计算机设备及存储介质 |
CN110765939A (zh) * | 2019-10-22 | 2020-02-07 | Oppo广东移动通信有限公司 | 身份识别方法、装置、移动终端及存储介质 |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114461468A (zh) * | 2022-01-21 | 2022-05-10 | 电子科技大学 | 一种基于人工神经网络的微处理器应用场景识别方法 |
CN114817635A (zh) * | 2022-04-15 | 2022-07-29 | 阿里巴巴达摩院(杭州)科技有限公司 | 行为识别方法、行为识别装置、视频检测方法及存储介质 |
CN115115967A (zh) * | 2022-05-13 | 2022-09-27 | 清华大学 | 模式生物的视频动作分析方法、装置、设备及介质 |
CN115204753A (zh) * | 2022-09-14 | 2022-10-18 | 深圳市深信信息技术有限公司 | 一种智慧农贸场所行为监测方法、系统及可读存储介质 |
CN116011633A (zh) * | 2022-12-23 | 2023-04-25 | 浙江苍南仪表集团股份有限公司 | 区域燃气用量预测方法、系统、设备及物联网云平台 |
CN116011633B (zh) * | 2022-12-23 | 2023-08-18 | 浙江苍南仪表集团股份有限公司 | 区域燃气用量预测方法、系统、设备及物联网云平台 |
CN116030272A (zh) * | 2023-03-30 | 2023-04-28 | 之江实验室 | 一种基于信息抽取的目标检测方法、系统和装置 |
CN116467500A (zh) * | 2023-06-15 | 2023-07-21 | 阿里巴巴(中国)有限公司 | 数据关系识别、自动问答、查询语句生成方法 |
CN116467500B (zh) * | 2023-06-15 | 2023-11-03 | 阿里巴巴(中国)有限公司 | 数据关系识别、自动问答、查询语句生成方法 |
Also Published As
Publication number | Publication date |
---|---|
CN111325289A (zh) | 2020-06-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021184468A1 (zh) | 行为识别方法、装置、设备及介质 | |
CN108831559B (zh) | 一种中文电子病历文本分析方法与系统 | |
CN112001177B (zh) | 融合深度学习与规则的电子病历命名实体识别方法及系统 | |
CN111540468B (zh) | 一种诊断原因可视化的icd自动编码方法与系统 | |
CN110705293A (zh) | 基于预训练语言模型的电子病历文本命名实体识别方法 | |
Zhou et al. | Facial depression recognition by deep joint label distribution and metric learning | |
CN107076567A (zh) | 多语言图像问答 | |
CN111524578B (zh) | 一种基于电子心理沙盘的心理评估装置、方法及系统 | |
Li et al. | Ffa-ir: Towards an explainable and reliable medical report generation benchmark | |
CN111145903B (zh) | 获取眩晕症问诊文本的方法、装置、电子设备及问诊系统 | |
US20210241906A1 (en) | Machine-aided dialog system and medical condition inquiry apparatus and method | |
CN112599213B (zh) | 一种分类编码确定方法、装置、设备及存储介质 | |
CN111832298A (zh) | 病历的质检方法、装置、设备以及存储介质 | |
CN112860842A (zh) | 病历标注方法、装置及存储介质 | |
CN115545021A (zh) | 一种基于深度学习的临床术语识别方法与装置 | |
CN112749277A (zh) | 医学数据的处理方法、装置及存储介质 | |
Nasser et al. | A review on depression detection and diagnoses based on visual facial cues | |
CN106951917A (zh) | 一种淋巴瘤病理类型的智能分类系统和方法 | |
Shi et al. | Understanding patient query with weak supervision from doctor response | |
CN112614562A (zh) | 基于电子病历的模型训练方法、装置、设备及存储介质 | |
CN117894439A (zh) | 一种基于人工智能的导诊方法、系统、电子设备及介质 | |
CN117422074A (zh) | 一种临床信息文本标准化的方法、装置、设备及介质 | |
CN110060749B (zh) | 基于sev-sdg-cnn的电子病历智能诊断方法 | |
CN111898528A (zh) | 数据处理方法、装置、计算机可读介质及电子设备 | |
CN111627566A (zh) | 适应症信息处理方法与装置、存储介质、电子设备 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20925705 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20925705 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20925705 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 10.07.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20925705 Country of ref document: EP Kind code of ref document: A1 |