CN108764176A - A kind of action sequence recognition methods, system and equipment and storage medium - Google Patents

A kind of action sequence recognition methods, system and equipment and storage medium Download PDF

Info

Publication number
CN108764176A
CN108764176A CN201810550608.XA CN201810550608A CN108764176A CN 108764176 A CN108764176 A CN 108764176A CN 201810550608 A CN201810550608 A CN 201810550608A CN 108764176 A CN108764176 A CN 108764176A
Authority
CN
China
Prior art keywords
action sequence
sample
neural networks
lstm neural
identified
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
CN201810550608.XA
Other languages
Chinese (zh)
Inventor
刘栩辰
董刚
赵雅倩
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhengzhou Yunhai Information Technology Co Ltd
Original Assignee
Zhengzhou Yunhai Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhengzhou Yunhai Information Technology Co Ltd filed Critical Zhengzhou Yunhai Information Technology Co Ltd
Priority to CN201810550608.XA priority Critical patent/CN108764176A/en
Publication of CN108764176A publication Critical patent/CN108764176A/en
Withdrawn legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/50Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

This application discloses a kind of action sequence recognition methods, system and equipment and computer readable storage medium, this method to include:Raw image data is obtained, and pretreatment operation is carried out to the raw image data and obtains sample to be identified;Feature extraction is carried out to each frame image in the sample to be identified, obtains feature vector;Described eigenvector is inputted in the LSTM neural networks that training is completed, obtains action sequence recognition result.Action sequence recognition methods provided by the present application, pretreatment operation is carried out to raw image data first, obtain sample to be identified, ensure that each frame image in sample to be identified is all of the same size and discrimination, feature extraction is carried out to each frame image again, and in the LSTM neural networks for the feature vector input training completion for obtaining extraction, to obtain action sequence recognition result.Raw image data is handled by feature extraction, then is combined with LSTM neural networks, the discrimination of action sequence is improved.

Description

A kind of action sequence recognition methods, system and equipment and storage medium
Technical field
This application involves technical field of image processing, more specifically to a kind of action sequence recognition methods, system and Equipment and a kind of computer readable storage medium.
Background technology
In recent years, the research of human action identification is paid high attention to by industrial quarters, in video monitoring, game and machine There is important application in the fields such as people.However efficiently action recognition algorithm is very challenging:First, different mobile speed Degree leads to the fluctuation of the same action in time;Secondly, it is many action have similitude, such as it is high throwing and wave;Most Afterwards, different people also results in the difficulty of identification in the difference of height, figure etc..Action sequence identification side in the prior art Method generally directly using image data as the input of deep learning network, however directly uses raw video picture data meeting Influence the discrimination of final action sequence.
Therefore, how to improve the discrimination of action sequence is those skilled in the art's problem to be solved.
Invention content
The application's is designed to provide a kind of action sequence recognition methods, system and equipment and a kind of computer-readable deposits Storage media improves the discrimination of action sequence.
To achieve the above object, this application provides a kind of action sequence recognition methods, including:
Raw image data is obtained, and pretreatment operation is carried out to the raw image data and obtains sample to be identified;
Feature extraction is carried out to each frame image in the sample to be identified, obtains feature vector;
Described eigenvector is inputted in the LSTM neural networks that training is completed, obtains action sequence recognition result.
Wherein, it is described obtain action sequence recognition result after, further include:
The average recognition rate of the sample to be identified is calculated according to the discrimination of each frame image.
Wherein, the pretreatment operation includes any one of turning operation, down-sampling operation or cutting operation or several Combination.
Wherein, before described eigenvector being inputted in the LSTM neural networks that training is completed, further include:
Training sample is obtained, and feature extraction is carried out to each frame image in the training sample, obtains training characteristics Vector;
The training feature vector is inputted in LSTM neural networks, and adjusts the key parameter of the LSTM neural networks Until the discrimination of LSTM neural networks output reaches preset value, to obtain the LSTM neural networks of training completion.
Wherein, the key parameter of the LSTM neural networks is adjusted, including:
The key parameter of the LSTM neural networks is adjusted using cross validation method and pair-wise algorithms.
Wherein, the propagated forward algorithm of the LSTM neural networks includes two level derivatives of the cell to the time.
Wherein, feature extraction is carried out to each frame image in the sample to be identified, obtains feature vector, including:
S321:Target image is determined from the sample to be identified, and obtains the previous frame image of target image and latter Frame image;
S322:The target image, the previous frame image and a later frame image are formed into the target image pair The candidate image answered, and the point cloud by characteristic value in the candidate image more than feature preset value is determined as key point;
S323:The HOPC features of all key points are extracted, and the target image is determined according to all HOPC features Feature vector;
S321-S323 is repeated until all images in the sample to be identified are all extracted and completed.
To achieve the above object, this application provides a kind of action sequence identifying systems, including:
Acquisition module for obtaining raw image data, and carries out pretreatment operation to the raw image data and obtains Sample to be identified;
Extraction module obtains feature vector for carrying out feature extraction to each frame image in the sample to be identified;
Input module obtains action sequence for inputting described eigenvector in the LSTM neural networks that training is completed Recognition result.
To achieve the above object, this application provides a kind of action sequence identification equipments, including:
Memory, for storing computer program;
Processor is realized when for executing the computer program such as the step of above-mentioned action sequence recognition methods.
To achieve the above object, this application provides a kind of computer readable storage medium, the computer-readable storages It is stored with computer program on medium, such as above-mentioned action sequence recognition methods is realized when the computer program is executed by processor The step of.
By above scheme it is found that a kind of action sequence recognition methods provided by the present application, including:Obtain original image number According to, and pretreatment operation is carried out to the raw image data and obtains sample to be identified;To each in the sample to be identified Frame image carries out feature extraction, obtains feature vector;Described eigenvector is inputted in the LSTM neural networks that training is completed, is obtained To action sequence recognition result.
Action sequence recognition methods provided by the present application carries out pretreatment operation to raw image data first, is waited for It identifies sample, ensures that each frame image in sample to be identified is all of the same size and discrimination, then to each frame image In the LSTM neural networks that the feature vector input training for carrying out feature extraction, and extraction being obtained is completed, to be acted Recognition sequence result.Raw image data is handled by feature extraction, then is combined with LSTM neural networks, action sequence is improved Discrimination.Disclosed herein as well is a kind of action sequence identifying system and equipment and a kind of computer readable storage mediums, together Sample can realize above-mentioned technique effect.
Description of the drawings
In order to illustrate the technical solutions in the embodiments of the present application or in the prior art more clearly, to embodiment or will show below There is attached drawing needed in technology description to be briefly described, it should be apparent that, the accompanying drawings in the following description is only this Some embodiments of application for those of ordinary skill in the art without creative efforts, can be with Obtain other attached drawings according to these attached drawings.
Fig. 1 is a kind of flow chart of action sequence recognition methods disclosed in the embodiment of the present application;
Fig. 2 is a kind of structure chart of LSTM neural networks;
Fig. 3 is the flow chart of another action sequence recognition methods disclosed in the embodiment of the present application;
Fig. 4 is a kind of structure chart of action sequence identifying system disclosed in the embodiment of the present application;
Fig. 5 is a kind of structure chart of action sequence identification equipment disclosed in the embodiment of the present application;
Fig. 6 is the structure chart of another action sequence identification equipment disclosed in the embodiment of the present application.
Specific implementation mode
Below in conjunction with the attached drawing in the embodiment of the present application, technical solutions in the embodiments of the present application carries out clear, complete Site preparation describes, it is clear that described embodiments are only a part of embodiments of the present application, instead of all the embodiments.It is based on Embodiment in the application, those of ordinary skill in the art are obtained every other without creative efforts Embodiment shall fall in the protection scope of this application.
The embodiment of the present application discloses a kind of action sequence recognition methods, improves the discrimination of action sequence.
Referring to Fig. 1, a kind of flow chart of action sequence recognition methods disclosed in the embodiment of the present application, as shown in Figure 1, packet It includes:
S101:Raw image data is obtained, and pretreatment operation is carried out to the raw image data and obtains sample to be identified This;
In specific implementation, it needs to carry out pretreatment operation to the raw image data after obtaining raw image data, i.e., Sample to be identified is obtained after enhancing operation, the present embodiment is not defined specific pretreatment operation, those skilled in the art It can flexibly be selected according to actual conditions.As a preferred implementation manner, pretreatment operation herein include turning operation, under Any one of sampling operation or cutting operation or several combinations.Wherein, down-sampling operation is i.e. between a sample sequence It is primary every the sampling of several sample values.
S102:Feature extraction is carried out to each frame image in the sample to be identified, obtains feature vector;
In specific implementation, in order to improve the action sequence discrimination of neural network output, in needing to the sample identified Each frame image carries out feature extraction, obtains feature vector.The present embodiment does not equally limit the specific features of extraction and feature carries The concrete mode taken, those skilled in the art can flexibly select according to actual conditions.For example, each frame image can be extracted HOPC (Chinese names:The histogram of principal component, full name in English:Histogram of Principal Component) feature, Specific feature extracting method will describe in detail in next embodiment.
S103:Described eigenvector is inputted in the LSTM neural networks that training is completed, obtains action sequence recognition result.
In specific implementation, the feature vector of each frame image is inputted into the LSTM (Chinese names that training is completed:Shot and long term Memory network, full name in English:Network-Long Short Term Memory Network) in neural network, to be moved Make sequence recognition result.
As a preferred implementation manner, as shown in Fig. 2, the propagated forward algorithm of LSTM neural networks herein includes Two level derivatives of the cell to the time.It can derive that corresponding back-propagating is calculated by propagated forward algorithm those skilled in the art Method, since there are cell in propagated forward algorithm and Back Propagation Algorithm to the two level derivative of time, can be fine preserve it is dynamic The temporal information for making sequence avoids the time misalignment of recognition result.
It is understood that this step gives tacit consent to the trained completion of LSTM neural networks, the instruction of above-mentioned LSTM neural networks Practicing process is specially:Training sample is obtained, and feature extraction is carried out to each frame image in the training sample, is trained Feature vector;The training feature vector is inputted in LSTM neural networks, and adjusts the crucial ginseng of the LSTM neural networks Number is until the discrimination of LSTM neural networks output reaches preset value, to obtain the LSTM neural networks of training completion.
It should be noted that each frame image in above-mentioned training sample passes through pretreatment operation, i.e., turning operation, under Sampling operation and cutting operation etc..The feature of training sample extraction and the mode of feature extraction and above-described sample to be identified The feature of extraction is identical with the mode of feature extraction.The training feature vector extracted is inputted in LSTM neural networks, is adjusted The key parameter of LSTM neural networks is until the discrimination of LSTM neural networks output reaches preset value.As a kind of preferred Embodiment can utilize cross validation method and pair-wise algorithms to adjust key parameter.Referring herein to key parameter can With include Batch_video (the video number inputted every time), Batch_frame (video frame number that each video includes), Epoch (numbers of the primary all data of training), learning rate or learning rate decaying etc..Herein not to the initial of above-mentioned key parameter Value is specifically limited, and those skilled in the art can be flexibly arranged according to actual conditions, for example, Batch_video=6, Batch_frame=24, epoch=5000~8000, learning rate Learning_rate=0.1, learning rate decaying lr_decay =0.1/1000 time.
It is understood that after above-mentioned LSTM neural networks are completed in training, test sample can also be utilized to training The LSTM neural networks of completion are tested, specifically, the average recognition rate of all picture frames in test sample is calculated, to obtain The action recognition accuracy rate of the LSTM neural networks.
It is understood that obtaining action sequence recognition result in this step not only including the action sequence at identification, also It may include discrimination, that is, calculate the discrimination of each frame image, and sample to be identified is calculated according to the discrimination of each frame image This average recognition rate.
Action sequence recognition methods provided by the embodiments of the present application carries out pretreatment operation to raw image data first, Sample to be identified is obtained, ensures that each frame image in sample to be identified is all of the same size and discrimination, then to each Frame image carries out feature extraction, and in the LSTM neural networks for the feature vector input training completion that extraction is obtained, to To action sequence recognition result.Raw image data is handled by feature extraction, then is combined with LSTM neural networks, is improved dynamic Make the discrimination of sequence.
The embodiment of the present application discloses a kind of action sequence recognition methods, and relative to a upper embodiment, the present embodiment is to skill Art scheme has made further instruction and optimization.Specifically:
Referring to Fig. 3, the flow chart of another kind action sequence recognition methods provided by the embodiments of the present application, as shown in figure 3, packet It includes:
S301:Raw image data is obtained, and pretreatment operation is carried out to the raw image data and obtains sample to be identified This;
S321:Target image is determined from the sample to be identified, and obtains the previous frame image of target image and latter Frame image;
S322:The target image, the previous frame image and a later frame image are formed into the target image pair The candidate image answered, and the point cloud by characteristic value in the candidate image more than feature preset value is determined as key point;
S323:The HOPC features of all key points are extracted, and the target image is determined according to all HOPC features Feature vector;
In the present embodiment, to the sample identified in the HOPC features of each frame image extract.First to be identified Target image is chosen in all picture frames in sample, and it is to wait to merge target image with its previous frame image and a later frame image Image is selected, and determines the key point of candidate image, key point herein is chosen from the point cloud in candidate image, selected characteristic value More than feature preset value point cloud as key point, feature preset value those skilled in the art herein can be according to feature extraction The precision needed is configured.
It is understood that in specific implementation, can also determine the point cloud of all picture frames first, remerge target figure As the point cloud with its previous frame image and a later frame image, key point finally is chosen from cloud, the HOPC for extracting key point is special Sign.It is, of course, also possible to determine the key point of all picture frames first according to preset rules, target image and its former frame are remerged The key point of image and a later frame image is finally extracted the HOPC features of all key points, is not specifically limited herein.
For example, by taking a-th cloud of F frames as an example, merge F-1, F, the point cloud of F+1 frames;Using a as the centre of sphere, r is radius, Ask the covariance matrix and its feature vector of a;By in eigenvector projection to 20 face bodies, the HOPC features of a are finally obtained;Frame F HOPC feature vectors be to connect the HOPC features of all key points.
S324:Judge whether all images in the sample to be identified complete by extraction, if so, into S303, If it is not, then entering S321;
S303:Described eigenvector is inputted in the LSTM neural networks that training is completed, obtains action sequence recognition result.
A kind of action sequence identifying system provided by the embodiments of the present application is introduced below, it is described below a kind of dynamic Making sequence identifying system can be cross-referenced with a kind of above-described action sequence recognition methods.
Referring to Fig. 4, a kind of structure chart of action sequence identifying system provided by the embodiments of the present application, as shown in figure 4, packet It includes:
Acquisition module 401 for obtaining raw image data, and carries out pretreatment operation to the raw image data and obtains To sample to be identified;
Extraction module 402, for in the sample to be identified each frame image carry out feature extraction, obtain feature to Amount;
Input module 403 obtains action sequence for inputting described eigenvector in the LSTM neural networks that training is completed Row recognition result.
Action sequence identifying system provided by the embodiments of the present application carries out pretreatment operation to raw image data first, Sample to be identified is obtained, ensures that each frame image in sample to be identified is all of the same size and discrimination, then to each Frame image carries out feature extraction, and in the LSTM neural networks for the feature vector input training completion that extraction is obtained, to To action sequence recognition result.Raw image data is handled by feature extraction, then is combined with LSTM neural networks, is improved dynamic Make the discrimination of sequence.
On the basis of the above embodiments, further include as a preferred implementation manner,:
Computing module, the average identification for calculating the sample to be identified according to the discrimination of each frame image Rate.
On the basis of the above embodiments, as a preferred implementation manner, the pretreatment operation include turning operation, Any one of down-sampling operation or cutting operation or several combinations.
On the basis of the above embodiments, further include as a preferred implementation manner,:
Training sample processing module is carried out for obtaining training sample, and to each frame image in the training sample Feature extraction obtains training feature vector;
Training module for inputting the training feature vector in LSTM neural networks, and adjusts the LSTM nerves The key parameter of network is until the discrimination of LSTM neural networks output reaches preset value, to obtain the LSTM of training completion Neural network.
On the basis of the above embodiments, the training module is specially by the instruction as a preferred implementation manner, Practice in feature vector input LSTM neural networks, and the LSTM god is adjusted using cross validation method and pair-wise algorithms Key parameter through network is until the discrimination of LSTM neural networks output reaches preset value, to obtain training completion The module of LSTM neural networks.
On the basis of the above embodiments, the propagated forward of the LSTM neural networks as a preferred implementation manner, Algorithm includes two level derivatives of the cell to the time.
On the basis of the above embodiments, the extraction module 402 includes as a preferred implementation manner,:
Determination unit for determining target image from the sample to be identified, and obtains the former frame figure of target image Picture and a later frame image;
Screening unit, for the target image, the previous frame image and a later frame image to be formed the mesh The corresponding candidate image of logo image, and the point cloud by characteristic value in the candidate image more than feature preset value is determined as key Point;
Extraction unit, the HOPC features for extracting all key points, and the mesh is determined according to all HOPC features The feature vector of logo image;The workflow of the determination unit is restarted up to all images in the sample to be identified All extraction is completed.
Present invention also provides a kind of action sequence identification equipments, referring to Fig. 5, a kind of action provided by the embodiments of the present application The structure chart of recognition sequence equipment, as shown in figure 5, including:
Memory 100, for storing computer program;
Processor 200, may be implemented the step of above-described embodiment is provided when for executing the computer program.
Specifically, memory 100 includes non-volatile memory medium, built-in storage.The non-volatile memory medium stores Have operating system and computer-readable instruction, the built-in storage be non-volatile memory medium in operating system and computer can The operation of reading instruction provides environment.Processor 200 provides calculating and control ability for action sequence identification equipment, is deposited described in execution When the computer program preserved in reservoir 100, following steps may be implemented:Raw image data is obtained, and to the original graph Sample to be identified is obtained as data carry out pretreatment operation;Feature is carried out to each frame image in the sample to be identified to carry It takes, obtains feature vector;Described eigenvector is inputted in the LSTM neural networks that training is completed, obtains action sequence identification knot Fruit.
The embodiment of the present application carries out pretreatment operation to raw image data first, obtains sample to be identified, ensures to wait knowing Each frame image very in this is all of the same size and discrimination, then carries out feature extraction to each frame image, and will It extracts in the LSTM neural networks that obtained feature vector input training is completed, to obtain action sequence recognition result.Pass through Feature extraction handles raw image data, then is combined with LSTM neural networks, and the discrimination of action sequence is improved.
Preferably, it when the processor 200 executes the computer subprogram preserved in the memory 100, may be implemented Following steps:The average recognition rate of the sample to be identified is calculated according to the discrimination of each frame image.
Preferably, it when the processor 200 executes the computer subprogram preserved in the memory 100, may be implemented Following steps:Training sample is obtained, and feature extraction is carried out to each frame image in the training sample, obtains training characteristics Vector;
Preferably, it when the processor 200 executes the computer subprogram preserved in the memory 100, may be implemented Following steps:The training feature vector is inputted in LSTM neural networks, and adjusts the crucial ginseng of the LSTM neural networks Number is until the discrimination of LSTM neural networks output reaches preset value, to obtain the LSTM neural networks of training completion.
Preferably, it when the processor 200 executes the computer subprogram preserved in the memory 100, may be implemented Following steps:The key parameter of the LSTM neural networks is adjusted using cross validation method and pair-wise algorithms.
Preferably, it when the processor 200 executes the computer subprogram preserved in the memory 100, may be implemented Following steps:Target image is determined from the sample to be identified, and obtains the previous frame image and a later frame figure of target image Picture;The target image, the previous frame image and a later frame image are formed into the corresponding candidate figure of the target image Picture, and the point cloud by characteristic value in the candidate image more than feature preset value is determined as key point;Extract all key points HOPC features, and determine according to all HOPC features the feature vector of the target image;It repeats the above steps until described All images in sample to be identified all complete by extraction.
On the basis of the above embodiments, preferably, referring to Fig. 6, the action sequence identification equipment is also Including:
Input interface 300 is connected with processor 200, computer program, parameter and instruction for obtaining external importing, It is preserved into memory 100 through the control of processor 200.The input interface 300 can be connected with input unit, and it is manual to receive user The parameter of input or instruction.The input unit can be the touch layer covered on display screen, can also be to be arranged in terminal enclosure Button, trace ball or Trackpad, can also be keyboard, Trackpad or mouse etc..Specifically, in the present embodiment, Ke Yitong It crosses input interface 300 and inputs LSTM neural network models etc..
Display unit 400 is connected with processor 200, the data sent for video-stream processor 200.The display unit 400 Can be display screen, liquid crystal display or the electric ink display screen etc. in PC machine.It, can be with specifically, in the present embodiment The action sequence recognition result etc. of sample to be identified is shown by display unit 400.
The network port 500 is connected with processor 200, for being communicatively coupled with external each terminal device.The communication link The communication technology used by connecing can be cable communicating technology or wireless communication technique, and such as mobile high definition chained technology (MHL) leads to It is blue with universal serial bus (USB), high-definition media interface (HDMI), adopting wireless fidelity technology (WiFi), Bluetooth Communication Technology, low-power consumption The tooth communication technology, the communication technology etc. based on IEEE802.11s.Specifically, in the present embodiment, the network port can be passed through 500 import LSTM neural network models etc. to processor 200.
Video collector 600 is connected with processor 200, for obtaining video data, then sends video data to place It manages device 200 and carries out Data Analysis Services, handling result can be sent to display unit 400 and shown by subsequent processor 200, Or be transmitted to processor 100 and preserved, or preset data receiver end can be sent to by the network port 500 End.Specifically, in the present embodiment, sample, training sample and test sample to be identified etc. can be obtained with video collector 600.
Present invention also provides a kind of computer readable storage medium, which may include:USB flash disk, mobile hard disk, Read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic The various media that can store program code such as dish or CD.Computer program, the calculating are stored on the storage medium Machine program realizes following steps when being executed by processor:Raw image data is obtained, and the raw image data is carried out pre- Processing operation obtains sample to be identified;To in the sample to be identified each frame image carry out feature extraction, obtain feature to Amount;Described eigenvector is inputted in the LSTM neural networks that training is completed, obtains action sequence recognition result.
The embodiment of the present application carries out pretreatment operation to raw image data first, obtains sample to be identified, ensures to wait knowing Each frame image very in this is all of the same size and discrimination, then carries out feature extraction to each frame image, and will It extracts in the LSTM neural networks that obtained feature vector input training is completed, to obtain action sequence recognition result.Pass through Feature extraction handles raw image data, then is combined with LSTM neural networks, and the discrimination of action sequence is improved.
Preferably, when the computer subprogram stored in the computer readable storage medium is executed by processor, specifically Following steps may be implemented:The average recognition rate of the sample to be identified is calculated according to the discrimination of each frame image.
Preferably, when the computer subprogram stored in the computer readable storage medium is executed by processor, specifically Following steps may be implemented:Training sample is obtained, and feature extraction is carried out to each frame image in the training sample, is obtained Training feature vector;
Preferably, when the computer subprogram stored in the computer readable storage medium is executed by processor, specifically Following steps may be implemented:The training feature vector is inputted in LSTM neural networks, and adjusts the LSTM neural networks Key parameter until the LSTM neural networks output discrimination reach preset value, with obtain training completion LSTM nerve Network.
Preferably, when the computer subprogram stored in the computer readable storage medium is executed by processor, specifically Following steps may be implemented:The crucial ginseng of the LSTM neural networks is adjusted using cross validation method and pair-wise algorithms Number.
Preferably, when the computer subprogram stored in the computer readable storage medium is executed by processor, specifically Following steps may be implemented:From the sample to be identified determine target image, and obtain target image previous frame image and A later frame image;The target image, the previous frame image and a later frame image are formed the target image to correspond to Candidate image, and by characteristic value in the candidate image be more than feature preset value point cloud be determined as key point;Extraction is all The HOPC features of key point, and determine according to all HOPC features the feature vector of the target image;It repeats the above steps Until all images in the sample to be identified all complete by extraction.
Each embodiment is described by the way of progressive in this specification, the highlights of each of the examples are with other The difference of embodiment, just to refer each other for identical similar portion between each embodiment.
The foregoing description of the disclosed embodiments enables professional and technical personnel in the field to realize or use the application. Various modifications to these embodiments will be apparent to those skilled in the art, as defined herein General Principle can in other embodiments be realized in the case where not departing from spirit herein or range.Therefore, the application It is not intended to be limited to the embodiments shown herein, and is to fit to and the principles and novel features disclosed herein phase one The widest range caused.
Each embodiment is described by the way of progressive in specification, the highlights of each of the examples are with other realities Apply the difference of example, just to refer each other for identical similar portion between each embodiment.For system disclosed in embodiment Speech, since it is corresponded to the methods disclosed in the examples, so description is fairly simple, related place is referring to method part illustration ?.It should be pointed out that for those skilled in the art, under the premise of not departing from the application principle, also Can to the application, some improvement and modification can also be carried out, these improvement and modification also fall into the application scope of the claims It is interior.
It should also be noted that, in the present specification, relational terms such as first and second and the like be used merely to by One entity or operation are distinguished with another entity or operation, without necessarily requiring or implying these entities or operation Between there are any actual relationship or orders.Moreover, the terms "include", "comprise" or its any other variant meaning Covering non-exclusive inclusion, so that the process, method, article or equipment including a series of elements includes not only that A little elements, but also include other elements that are not explicitly listed, or further include for this process, method, article or The intrinsic element of equipment.In the absence of more restrictions, the element limited by sentence "including a ...", is not arranged Except there is also other identical elements in the process, method, article or apparatus that includes the element.

Claims (10)

1. a kind of action sequence recognition methods, which is characterized in that including:
Raw image data is obtained, and pretreatment operation is carried out to the raw image data and obtains sample to be identified;
Feature extraction is carried out to each frame image in the sample to be identified, obtains feature vector;
Described eigenvector is inputted in the LSTM neural networks that training is completed, obtains action sequence recognition result.
2. action sequence recognition methods according to claim 1, which is characterized in that it is described obtain action sequence recognition result it Afterwards, further include:
The average recognition rate of the sample to be identified is calculated according to the discrimination of each frame image.
3. action sequence recognition methods according to claim 1, which is characterized in that the pretreatment operation includes that overturning is grasped Any one of work, down-sampling operation or cutting operation or several combinations.
4. action sequence recognition methods according to claim 1, which is characterized in that described eigenvector is inputted training and is completed LSTM neural networks in before, further include:
Training sample is obtained, and feature extraction is carried out to each frame image in the training sample, obtains training feature vector;
By the training feature vector input LSTM neural networks in, and adjust the LSTM neural networks key parameter until The discrimination of the LSTM neural networks output reaches preset value, to obtain the LSTM neural networks of training completion.
5. action sequence recognition methods according to claim 4, which is characterized in that adjust the key of the LSTM neural networks Parameter, including:
The key parameter of the LSTM neural networks is adjusted using cross validation method and pair-wise algorithms.
6. action sequence recognition methods according to claim 4, which is characterized in that the propagated forward of the LSTM neural networks Algorithm includes two level derivatives of the cell to the time.
7. according to any one of the claim 1-6 action sequence recognition methods, which is characterized in that in the sample to be identified Each frame image carry out feature extraction, obtain feature vector, including:
S321:Target image is determined from the sample to be identified, and obtains the previous frame image and a later frame figure of target image Picture;
S322:It is corresponding that the target image, the previous frame image and a later frame image are formed into the target image Candidate image, and the point cloud by characteristic value in the candidate image more than feature preset value is determined as key point;
S323:The HOPC features of all key points are extracted, and determine the feature of the target image according to all HOPC features Vector;
S321-S323 is repeated until all images in the sample to be identified are all extracted and completed.
8. a kind of action sequence identifying system, which is characterized in that including:
Acquisition module for obtaining raw image data, and carries out pretreatment operation to the raw image data and obtains waiting knowing Very originally;
Extraction module obtains feature vector for carrying out feature extraction to each frame image in the sample to be identified;
Input module obtains action sequence identification for inputting described eigenvector in the LSTM neural networks that training is completed As a result.
9. a kind of action sequence identification equipment, which is characterized in that including:
Memory, for storing computer program;
Processor realizes the action sequence identification side as described in any one of claim 1 to 7 when for executing the computer program The step of method.
10. a kind of computer readable storage medium, which is characterized in that be stored with computer on the computer readable storage medium Program realizes the action sequence recognition methods as described in any one of claim 1 to 7 when the computer program is executed by processor The step of.
CN201810550608.XA 2018-05-31 2018-05-31 A kind of action sequence recognition methods, system and equipment and storage medium Withdrawn CN108764176A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810550608.XA CN108764176A (en) 2018-05-31 2018-05-31 A kind of action sequence recognition methods, system and equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810550608.XA CN108764176A (en) 2018-05-31 2018-05-31 A kind of action sequence recognition methods, system and equipment and storage medium

Publications (1)

Publication Number Publication Date
CN108764176A true CN108764176A (en) 2018-11-06

Family

ID=64001506

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810550608.XA Withdrawn CN108764176A (en) 2018-05-31 2018-05-31 A kind of action sequence recognition methods, system and equipment and storage medium

Country Status (1)

Country Link
CN (1) CN108764176A (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109902729A (en) * 2019-02-18 2019-06-18 清华大学 Behavior prediction method and device based on sequence state evolution
CN110070052A (en) * 2019-04-24 2019-07-30 广东工业大学 A kind of robot control method based on mankind's demonstration video, device and equipment
CN110070029A (en) * 2019-04-17 2019-07-30 北京易达图灵科技有限公司 A kind of gait recognition method and device
CN110765967A (en) * 2019-10-30 2020-02-07 腾讯科技(深圳)有限公司 Action recognition method based on artificial intelligence and related device
CN111325289A (en) * 2020-03-18 2020-06-23 中国科学院深圳先进技术研究院 Behavior recognition method, device, equipment and medium
CN111444895A (en) * 2020-05-08 2020-07-24 商汤集团有限公司 Video processing method and device, electronic equipment and storage medium
CN111783650A (en) * 2020-06-30 2020-10-16 北京百度网讯科技有限公司 Model training method, action recognition method, device, equipment and storage medium
WO2020252923A1 (en) * 2019-06-18 2020-12-24 平安科技(深圳)有限公司 Sample data processing method and apparatus, computer apparatus, and storage medium
CN112580577A (en) * 2020-12-28 2021-03-30 出门问问(苏州)信息科技有限公司 Training method and device for generating speaker image based on face key points

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104504365A (en) * 2014-11-24 2015-04-08 闻泰通讯股份有限公司 System and method for smiling face recognition in video sequence
CN105787963A (en) * 2016-02-26 2016-07-20 浪潮软件股份有限公司 Video target tracking method and device
CN107451552A (en) * 2017-07-25 2017-12-08 北京联合大学 A kind of gesture identification method based on 3D CNN and convolution LSTM

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104504365A (en) * 2014-11-24 2015-04-08 闻泰通讯股份有限公司 System and method for smiling face recognition in video sequence
CN105787963A (en) * 2016-02-26 2016-07-20 浪潮软件股份有限公司 Video target tracking method and device
CN107451552A (en) * 2017-07-25 2017-12-08 北京联合大学 A kind of gesture identification method based on 3D CNN and convolution LSTM

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
HOSSEIN RAHMANI 等,: "Histogram of Oriented Principal Components for Cross-View Action Recognition", 《IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE》 *
VIVEK VEERIAH 等,: "Differential Recurrent Neural Networks for Action Recognition", 《2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV)》 *
于成龙,: "基于视频的人体行为识别关键技术研究", 《中国博士学位论文全文数据库信息科技辑》 *

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109902729A (en) * 2019-02-18 2019-06-18 清华大学 Behavior prediction method and device based on sequence state evolution
CN110070029A (en) * 2019-04-17 2019-07-30 北京易达图灵科技有限公司 A kind of gait recognition method and device
CN110070052A (en) * 2019-04-24 2019-07-30 广东工业大学 A kind of robot control method based on mankind's demonstration video, device and equipment
WO2020252923A1 (en) * 2019-06-18 2020-12-24 平安科技(深圳)有限公司 Sample data processing method and apparatus, computer apparatus, and storage medium
CN110765967A (en) * 2019-10-30 2020-02-07 腾讯科技(深圳)有限公司 Action recognition method based on artificial intelligence and related device
CN110765967B (en) * 2019-10-30 2022-04-22 腾讯科技(深圳)有限公司 Action recognition method based on artificial intelligence and related device
CN111325289A (en) * 2020-03-18 2020-06-23 中国科学院深圳先进技术研究院 Behavior recognition method, device, equipment and medium
CN111444895A (en) * 2020-05-08 2020-07-24 商汤集团有限公司 Video processing method and device, electronic equipment and storage medium
CN111444895B (en) * 2020-05-08 2024-04-19 商汤集团有限公司 Video processing method, device, electronic equipment and storage medium
CN111783650A (en) * 2020-06-30 2020-10-16 北京百度网讯科技有限公司 Model training method, action recognition method, device, equipment and storage medium
CN112580577A (en) * 2020-12-28 2021-03-30 出门问问(苏州)信息科技有限公司 Training method and device for generating speaker image based on face key points
CN112580577B (en) * 2020-12-28 2023-06-30 出门问问(苏州)信息科技有限公司 Training method and device for generating speaker image based on facial key points

Similar Documents

Publication Publication Date Title
CN108764176A (en) A kind of action sequence recognition methods, system and equipment and storage medium
CN108830235B (en) Method and apparatus for generating information
CN108154105B (en) Underwater biological detection and identification method and device, server and terminal equipment
CN108229489A (en) Crucial point prediction, network training, image processing method, device and electronic equipment
CN110909630B (en) Abnormal game video detection method and device
US10719693B2 (en) Method and apparatus for outputting information of object relationship
CN110362494B (en) Method for displaying microservice state information, model training method and related device
CN110175502A (en) A kind of backbone Cobb angle measuring method, device, readable storage medium storing program for executing and terminal device
CN108229478A (en) Image, semantic segmentation and training method and device, electronic equipment, storage medium and program
CN110930296B (en) Image processing method, device, equipment and storage medium
AU2011254040B2 (en) Method, apparatus and system for determining a saliency map for an input image
JP2021532434A (en) Face feature extraction model Training method, face feature extraction method, device, equipment and storage medium
CN108229262B (en) Pornographic video detection method and device
US11669990B2 (en) Object area measurement method, electronic device and storage medium
KR101955919B1 (en) Method and program for providing tht region-of-interest in image by deep-learing algorithm
WO2020181706A1 (en) Plant species identification method and apparatus
CN110197183A (en) A kind of method, apparatus and computer equipment of Image Blind denoising
CN105979283A (en) Video transcoding method and device
CN113191478A (en) Training method, device and system of neural network model
CN114037003A (en) Question-answer model training method and device and electronic equipment
CN108921138B (en) Method and apparatus for generating information
CN109241930B (en) Method and apparatus for processing eyebrow image
CN111062914B (en) Method, apparatus, electronic device and computer readable medium for acquiring facial image
CN116912187A (en) Image generation model training and image generation method, device, equipment and medium
CN110490065B (en) Face recognition method and device, storage medium and computer equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WW01 Invention patent application withdrawn after publication
WW01 Invention patent application withdrawn after publication

Application publication date: 20181106