WO2021143103A1 - 视频数据处理方法、装置、设备及计算机可读存储介质 - Google Patents
视频数据处理方法、装置、设备及计算机可读存储介质 Download PDFInfo
- Publication number
- WO2021143103A1 WO2021143103A1 PCT/CN2020/105904 CN2020105904W WO2021143103A1 WO 2021143103 A1 WO2021143103 A1 WO 2021143103A1 CN 2020105904 W CN2020105904 W CN 2020105904W WO 2021143103 A1 WO2021143103 A1 WO 2021143103A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- walking
- person
- target
- target person
- information
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
- G06V40/23—Recognition of whole body movements, e.g. for sport training
- G06V40/25—Recognition of walking or running movements, e.g. gait recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/41—Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/30—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for calculating health indices; for individual health risk assessment
Definitions
- This application relates to the technical field of data analysis, and in particular to a video data processing method, device, equipment, and computer-readable storage medium.
- a video data processing method includes the following steps:
- the walking information includes at least one of a walking state, a walking duration, a walking distance, and a walking speed;
- the walking ability index of the target person is determined.
- a video data processing device includes:
- the extraction module is used to obtain video data, and extract a plurality of person images from the video data through a preset target detection network, wherein the video data includes the walking situation information of the target person;
- the detection module is configured to detect each of the person images through a preset bone key point detection network to obtain multiple bone key points in each of the person images;
- the first determining module is used to determine the walking information of the target person according to the multiple bone key points in each of the person images, wherein the walking information includes walking status, walking time, walking distance, and walking speed. At least one of
- the second determining module is used to determine the walking ability index of the target person according to the walking information.
- a computer device including a processor, a memory, and a computer program stored on the memory and executable by the processor, wherein when the computer program is executed by the processor, the following steps are implemented :
- the walking information of the target person includes at least one of a walking state, a walking duration, a walking distance, and a walking speed;
- the walking ability index of the target person is determined.
- the computer program is executed by a processor, the following steps are implemented:
- the walking information includes at least one of a walking state, a walking duration, a walking distance, and a walking speed;
- the walking ability index of the target person is determined.
- FIG. 1 is a schematic flowchart of a video data processing method provided by an embodiment of this application
- Figure 2 is a schematic diagram of a rectangular frame in an embodiment of the application.
- FIG. 3 is a schematic flowchart of sub-steps of the video data processing method in FIG. 1;
- FIG. 4 is a schematic diagram of a scene in which the video data processing method provided by this embodiment is implemented.
- FIG. 5 is a schematic flowchart of another video data processing method provided by an embodiment of the application.
- FIG. 6 is a schematic block diagram of a video data processing device provided by an embodiment of this application.
- FIG. 7 is a schematic block diagram of sub-modules of the video data processing device in FIG. 5;
- FIG. 8 is a schematic block diagram of another video data processing device provided by an embodiment of this application.
- FIG. 9 is a schematic block diagram of the structure of a computer device related to an embodiment of the application.
- the embodiments of the present application provide a video data processing method, device, equipment, and computer-readable storage medium.
- the video data processing method can be applied to an electronic device or a server.
- the electronic device can be a mobile phone, a tablet computer, a notebook computer, a desktop computer, a personal digital assistant, a wearable device, etc.;
- the server can be a single server, It can also be a server cluster composed of multiple servers.
- the following takes the video data processing method applied to the server as an example for explanation.
- FIG. 1 is a schematic flowchart of a video data processing method according to an embodiment of the application.
- the video data processing method includes steps S101 to S104.
- Step S101 Obtain video data, and extract multiple person images from the video data through a preset target detection network, where the video data includes walking condition information of the target person.
- the user can record the walking process of the target person through an electronic device, so as to obtain the video data including the walking situation information of the target person.
- the user can directly connect to the electronic device through the server, or transmit the video data recorded by the electronic device to the server, so that the server can evaluate the walking ability of the target person in the video data.
- the target person is a person whose walking ability is to be evaluated.
- the electronic equipment includes, but is not limited to, electronic equipment with video recording functions such as video recorders, video cameras, and digital cameras.
- the user records the walking process of the target person through the electronic device, and stores the recorded video data in the database.
- the server detects the walking ability evaluation request triggered by the user, the server will respond according to the walking ability evaluation request.
- the video identifier of extracts the video data from the database, and the video identifier uniquely corresponds to the video data.
- the electronic device sends the recorded video data directly to the server, and the server receives it in real time or regularly.
- the video data can also be included in the walking ability evaluation request.
- the server performs real-time evaluation and analysis on the received video data, and Output the result of this evaluation analysis.
- the above-mentioned electronic devices include but are not limited to mobile phones, tablet computers, notebook computers, etc.
- the database may be a local database or a cloud database.
- the electronic device when recording the walking process of the target person through the electronic device, displays a video recording page, and the video recording page displays reminder information, and the reminder information is used to remind the user that the target person photographed is in the video recording page.
- the user should pay attention to adjusting the shooting angle and shooting distance, so that the target person in the video data obtained by shooting remains intact.
- the electronic device detects the recording instruction, it records the walking process of the target person based on the recording instruction, and obtains video data including the walking condition information of the target person.
- the server After the server obtains the video data, it extracts multiple person images from the video data through a preset target detection network.
- the video data is composed of several frames of images.
- the several frames of images include the walking situation information of the target person.
- the walking situation information records the walking process of the target person.
- the person image includes the image of the person extracted from the several frame images.
- the preset target detection network can automatically detect the person in each frame of the video data, and at least one person image can be extracted from each frame of the image of the person through the preset target detection network.
- the above-mentioned preset target detection network is an improved target detection network.
- the improved target detection network is obtained by performing migration learning on the traditional target detection network.
- the model sample data set of the improved target detection network for migration learning is marked with
- the data set of people with a large amount of data optionally, is obtained from the MSCOCO database labeled with a data set of people with a large amount of data.
- the process of migration learning for the target detection network is: in the model design stage, the convolutional layer parameters of the traditional target detection network are retained, and the model parameters corresponding to the model layers other than the convolutional layer are used as the model parameters to be trained ; Through the acquired model sample data set, iterative training is performed on the model parameters corresponding to the model layer other than the convolutional layer, until the model converges, and the model converges to obtain an improved target detection network. Since the improved target detection network is trained based on the data set of annotated people, the improved target detection network only includes the ability to detect people, and can accurately and quickly detect the position of the person in the image, which greatly improves the extraction The efficiency of character images.
- the target detection network includes a first target detection sub-network and a second target detection sub-network.
- the specific method for extracting multiple person images from the video data is: each frame of image in the video data is input to the first A target detection sub-network and a second target detection sub-network to obtain the first rectangular frame and the second rectangular frame of each frame of image; according to the first rectangular frame and the second rectangular frame of each frame of image, determine the target rectangle of each frame of image Frame; According to the target rectangular frame of each frame of image, at least one person image is extracted from each frame of image.
- the first rectangular frame and the second rectangular frame are used to frame the character in each frame of the video data
- the target rectangular frame is used to frame the character image in each frame of the video data. If there is an image in each frame People, there is at least one target rectangular frame in each frame of image.
- first target detection sub-network and the second target detection sub-network are two improved target detection sub-networks, and both are obtained through migration learning of the traditional target detection network. Refer to the specific migration learning process The above description will not be repeated here. Through the first target detection sub-network and the second target detection sub-network, the person image extracted from the video data is more accurate.
- the specific method for determining the target rectangular frame of each frame of image is: respectively obtaining the position information of the first rectangular frame and the position information of the second rectangular frame of each frame of image; combining the position information of the first rectangular frame of each frame of image Compare with the position information of the corresponding second rectangular frame to obtain the rectangular frame comparison result of each frame of image; determine according to the rectangular frame comparison result of each frame image, the position information of the first rectangular frame and the position information of the second rectangular frame The target rectangle of each frame of image.
- the position information of the first rectangular frame includes four first position coordinates of the four corner points of the first rectangular frame
- the position information of the second rectangular frame includes four second coordinates of the four corner points of the second rectangular frame.
- Position coordinates, the four corner points of the first rectangular frame in each frame of image correspond to the four corner points of the second rectangular frame one-to-one
- the four first position coordinates correspond to the four second position coordinates one-to-one.
- the position coordinates are compared with the corresponding second position coordinates to obtain four comparison results of each frame of image; the four corresponding target position coordinates are determined according to the four comparison results of each frame image, and the four target position coordinates are determined
- the target rectangle in each frame of image includes the comparison result of the abscissa and the ordinate, and the comparison result uniquely corresponds to the target position coordinate.
- the method of determining a target position coordinate can be: respectively determine the larger abscissa in a comparison result.
- Coordinates and the smaller ordinate use the larger abscissa and smaller ordinate as the abscissa and ordinate of the corresponding target position coordinates, so that the area of the target rectangle determined by the selected target position coordinates is larger , So that the target rectangular frame includes the first rectangular frame and the second rectangular frame.
- a rectangular coordinate system is established with the lower left corner of each frame of image in the video data as the origin, and the four corner points of the first rectangular frame K1 are A1, A2, A3, and A4, respectively.
- the four corners of the second rectangular frame K2 are B1, B2, B3, and B4; compare the position coordinates of A1 with the position coordinates of B1, compare the position coordinates of A2 with the position coordinates of B2, and compare the position coordinates of A3 Compare the position coordinates with the position coordinates of B3 and compare the position coordinates of A4 with the position coordinates of B4 to obtain four comparison results including the abscissa and the ordinate; select the abscissa of A1 and the ordinate of B1 as the first
- the target location coordinates namely C1 in Figure 2
- select the ordinate of A4 and the abscissa of B4 as the second target location coordinates, that is, C2 in Figure 2
- Step S102 Detect each of the person images through a preset bone key point detection network to obtain multiple bone key points in each of the person images.
- the server After obtaining multiple person images through the improved target detection network, the server detects each person image through the preset bone key point detection network to obtain multiple bone key points in each person image.
- the bone key points are human bone key points, which are used to represent the human bone structure, and the bone key point detection network is implemented based on a neural network.
- the implementation of the bone key point detection network is: collect bone key point data from a database labeled with a large amount of human data set as a model sample data set, where the database can be selected as the MSCOCO database; Design a bone key point detection network based on a neural network, where the neural network can be selected as a high-resolution network (HRNet); through the model sample data set to iteratively train the bone key point detection network to convergence, you can get the bone key point detection The internet.
- HRNet high-resolution network
- the multiple bone key points are improved multiple bone key points, that is, the basis of the seventeen bone key points such as nose, eyes, ears, shoulders, elbows, hands, hips, knees, and ankles in the MSCOCO database
- a key point of the bones of the neck is added to make the detected human bone structure more accurate.
- each bone key point when obtaining multiple bone key points in each person image, each bone key point needs to be tracked to ensure that the bone key points of the target person in each subsequent person image will not be lost.
- the tracking method is specifically as follows: number each bone key point in different positions of the human body, where the number includes the number of each character image and each character image The number of each bone key point in the image; select a bone key point in a character image as the target bone key point, and obtain each bone key point in the next frame of the character image; based on a preset
- the sparse optical flow algorithm calculates the key point similarity between the target bone key point and each bone key point in the next frame of person image according to the target bone key point and each bone key point in the next frame of person image; Determine the maximum key point similarity between the target bone key point and each bone key point in the next frame of person image, and use the bone key point in the next frame of person image corresponding to the maximum key point similarity as the to-be-
- selecting a bone key point in a person image can be selected as any bone key point in the first person image, which is convenient for tracking the entire walking process of the target person.
- the preset sparse optical flow algorithm can be set according to the actual situation. This application does not make specific restrictions here.
- the bone key points in the next frame of the person image corresponding to the maximum key point similarity are selected as the target bone key points to be tracked , It can ensure that the bone key points in the next frame of character image selected are correct. Using it as the target bone key point to be tracked can continue to perform the process of each bone key point in the next frame of the target bone key point. Tracking, to ensure the continuity of the key points of the bones in the tracking position.
- OKS p represents the key point similarity between the bone key point numbered i in the character image and a bone key point in the next frame of the character image (hereinafter referred to as the two bone key points),
- d pi represents the Euclidean distance of the key point of the skeleton numbered i in the character image, Represents the area occupied by the target person in the image of the person,
- ⁇ i is the standard deviation between the two bone key points
- v pi is the attribute of the bone key points
- the attributes of bone key points include invisible and visible.
- Step S103 Determine walking information of the target person according to the multiple bone key points in each person image, where the walking information includes at least one of walking state, walking time, walking distance, and walking speed .
- the server After obtaining the multiple bone key points in each person image, the server analyzes the multiple bone key points of the target person in each person image, and can determine the walking information of the target person.
- the walking information includes at least one of walking status, walking duration, walking distance, and walking speed.
- step S103 includes: sub-step S1031 to sub-step S1032.
- Sub-step S1031 according to the multiple bone key points in each person image, determine multiple bone key points of the target person.
- the key point is to determine the walking information of the target person.
- the walking information of the target person is determined according to multiple key bone points of the target person.
- the walking information of the target person can be determined. Specifically, any bone key point of the target person is selected as the target bone key point, and the position information of the target bone key point in each person image is obtained; according to the position information of the target bone key point in each person image, Determine the walking distance of the target person; calculate the time required for the walking distance to reach the preset distance threshold, and use the time required for the walking distance to reach the preset distance threshold as the walking time of the target person.
- the position information includes the coordinate information of the key points of the target bone
- the key points of the target bone may be selected as the key points of the bones with small changes in the position of the person during walking, such as the key points of the bones at the neck or shoulder position.
- the preset distance threshold can be set according to the actual situation. This application does not specifically limit it here. It can be selected as 45 meters. In some scenarios, the space size of the video data processing site cannot reach the preset size. The distance threshold. At this time, the walking distance of the target person can be calculated by the way back and forth. By taking the time required for the walking distance to reach the preset distance threshold as the walking time of the target person, and using the walking time of the target person to evaluate the walking ability of the target person, the evaluation result is more accurate, and the target person is excluded from walking. The impact of unstable speed.
- the specific method for determining the walking information of the target person can also be: selecting multiple bone key points of the target person as the bone key point set, and obtaining each bone key point in the bone key point set in each character image According to the position of each bone key point in each character image, determine the walking distance of the target person corresponding to each bone key point; calculate the target walking time corresponding to each walking distance reaching the preset distance threshold, And calculate the average of the target walking time as the walking time of the target person.
- the determined walking time of the target person can be made more accurate, which indirectly improves the accuracy of evaluating the walking ability index of the target person.
- the multiple bone key points of the target person and the multiple bone key points of the person other than the target person are determined, according to The multiple bone key points of the target person and the multiple bone key points of the person other than the target person determine the walking state of the target person, where the walking state includes the assisted walking state and the independent walking state.
- the walking state includes the assisted walking state and the independent walking state.
- the method for determining the walking state of the target person in the walking process is specifically as follows: in each person image, determine whether the bone key points of the target person overlap with the bone key points of characters other than the target person; if the target person If the bone key points of the target person overlap with the bone key points of characters other than the target person, the walking state of the target person during walking is determined to be the auxiliary walking state. If the bone key points of the target person are the same as those of the person other than the target person If the key points of the bones do not overlap, it is determined that the walking state of the target person in the walking process is the independent walking state.
- Step S104 Determine the walking ability index of the target person according to the walking information.
- the walking ability index of the target person can be determined according to the walking information of the target person. It should be noted that the above walking ability index can be expressed by numbers or grades. Take the walking ability index as a number for example. Optionally, the value range of the walking ability index can be set to 0-100, which is understandable Yes, the larger the value of the walking ability index, the better the walking ability of the target person, and the smaller the value of the walking ability index, the worse the walking ability of the target person.
- the walking ability index of the target person can be determined through a table containing the mapping relationship between the walking information and the walking ability index.
- the mapping relationship table between the walking information and the walking ability index can be set according to the actual situation, and this application does not specifically limit it here.
- the assisted walking state duration and independent walking state duration of the target person determine the assisted walking state duration and independent walking state duration of the target person; obtain the total walking duration of the target person, and calculate the assisted walking state duration and independent walking state separately The ratio of the duration to the total walking time; according to the ratio of the total walking time and the assisted walking state time to the total walking time, the walking ability index of the target person is determined.
- the total walking time is the length of time required for the walking distance of the target person to reach the preset distance threshold.
- the preset distance threshold can be selected as 45.
- the duration of the assisted walking state is the total walking time occupied by the assisted walking state.
- the duration of the independent walking state is the time occupied by the independent walking state in the total walking time.
- the longer the total walking time, the lower the walking ability index of the target person, the greater the proportion of the total walking time in the assisted walking state, the lower the walking ability index of the target person, and the walking state of the target person can be directly and quickly Get the walking ability index of the target person.
- the initial walking ability index corresponding to the total walking time and the weighting coefficient corresponding to the ratio of the auxiliary walking state duration to the total walking time are obtained, and the product of the initial walking ability index and the weighting coefficient is calculated, and the initial walking ability index The product of the weight coefficient is used as the walking ability index of the target person. It should be noted that the initial walking ability index corresponding to the total walking time and the weight coefficient corresponding to the ratio of the auxiliary walking state time to the total walking time can be set based on the actual situation, which is not specifically limited in this application.
- the target person’s assisted walking time and independent walking time are 20s and 40s, respectively, and the target person’s total walking time is 60s, the ratios of assisted walking time and independent walking time to the total walking time are 33.3% and 66.7%.
- the target person’s initial walking ability index is 80.
- the above-mentioned weight coefficient is 0.82, which is the target
- the walking ability index of the character is the product of the initial walking ability index 80 and the weight coefficient 0.82, so the walking ability index of the target character is 65.6.
- the walking ability index of the target person is determined according to the walking time corresponding to the preset walking distance threshold of the target person in the walking information, that is, the mapping relationship table between the pre-stored walking time length and the walking ability index is queried, The walking ability index corresponding to the walking time is taken as the walking ability index of the target person. It is understandable that the walking ability index of the target person can also be determined according to the walking speed corresponding to the preset walking distance threshold of the target person in the walking information, that is, to query the mapping relationship table between the pre-stored walking speed and the walking ability index , The walking ability index corresponding to the walking speed is taken as the walking ability index of the target person.
- FIG. 4 is a schematic diagram of a scene in which the video data processing method provided by this embodiment is implemented.
- the user can record the walking process of the target person through the electronic device to obtain video data including the walking situation information of the target person.
- the electronic device can also directly obtain the video data, and the user can then record the video data through the electronic device. It is sent to the server, and the server evaluates the walking ability of the target person in the video data.
- the video data processing method provided by the above embodiments can accurately extract multiple person images from video data through the target detection network, and accurately obtain multiple bone key points in each person image through the bone key point detection network. Then the walking information of the target person is determined according to the multiple bones in each person image, and the walking ability index of the target person can be accurately and quickly determined according to the walking information.
- the entire video data processing process does not require human intervention, which is greatly improved. This improves the accuracy and speed of human walking ability assessment.
- FIG. 5 is a schematic flowchart of another video data processing method provided by an embodiment of the application.
- the video data processing method includes steps S201 to S206.
- Step S201 Obtain video data, and extract multiple person images from the video data through a preset target detection network, where the video data includes walking condition information of the target person.
- the server After the server obtains the video data, it extracts multiple person images from the video data through a preset target detection network.
- the video data is composed of several frames of images.
- the several frames of images include the walking situation information of the target person.
- the walking situation information records the walking process of the target person.
- the person image includes the image of the person extracted from the several frame images.
- the preset target detection network can automatically detect the person in each frame of the video data, and at least one person image can be extracted from each frame of the image of the person through the preset target detection network.
- the target detection network includes a first target detection sub-network, a second target detection sub-network, and a frame check layer.
- the specific method for extracting multiple person images from the video data is: The images are respectively input to the first target detection sub-network and the second target detection sub-network to obtain the first rectangular frame and the second rectangular frame of each frame of the image; based on the frame check layer, the first rectangular frame is performed through the second rectangular frame Check and determine whether the first rectangular frame passes the check; if the first rectangular frame passes the check, the target frame is determined according to the positional relationship between the first rectangular frame and the second rectangular frame; according to the target frame, from Images of people are extracted from each frame of image.
- the above-mentioned check result includes pass and fail, and the above-mentioned frame check layer is used to check the first rectangular frame.
- the method of verifying the first rectangular frame through the second rectangular frame is specifically: obtaining the position coordinates of the four corner points of the first rectangular frame, which are recorded as the first position coordinates of the four first corner points, and obtaining the first position coordinates of the four first corner points.
- the second position coordinates of the four corner points of the two rectangular boxes are recorded as the second position coordinates of the four second corner points, the first corner point corresponds to the second corner point one-to-one, and the first position coordinate and the second position coordinate One-to-one correspondence; calculate the coordinate difference between each first position coordinate and the corresponding second position coordinate to obtain four coordinate differences, and determine whether the four coordinate differences are less than or equal to the preset threshold, if these four If the coordinate difference is less than or equal to the preset threshold, it is determined that the first rectangular frame passes the verification. If at least one coordinate difference among the four coordinate differences is greater than the preset threshold, it is determined that the first rectangular frame fails the verification. It should be noted that the foregoing preset threshold may be set based on actual conditions, and this solution does not specifically limit this.
- the method for determining the target rectangular frame is specifically as follows: if the first rectangular frame passes the verification, then determining the distance between the first rectangular frame and the second rectangular frame.
- the positional relationship between the first rectangular box and the second rectangular box if the positional relationship between the first rectangular box and the second rectangular box is an intersection relationship, where the positional relationship includes an intersection relationship, an overlap relationship, an inclusion relationship, and an irrelevant relationship.
- the frame formed by the combination of two rectangular frames is used as the target frame; if the positional relationship between the first rectangular frame and the second rectangular frame is an inclusive relationship, the first rectangular frame or the second rectangular frame that contains one side is used as the target frame; if The positional relationship between the first rectangular frame and the second rectangular frame is an overlapping relationship, the first rectangular frame or the second rectangular frame is taken as the target frame; if there is no intersection between the first rectangular frame and the second rectangular frame Stack, that is, the positional relationship between the first rectangular frame and the second rectangular frame is irrelevant, then the area of the first rectangular frame and the second rectangular frame are compared, and the larger area of the first rectangular frame or the second rectangular frame The frame serves as the target frame.
- the first rectangular frame or the second rectangular frame with a larger output area includes the image of the target person to be evaluated, and the target frame is determined by determining the positional relationship between the first rectangular frame and the second rectangular frame to avoid other The interference of the background makes the extracted image of the person more accurate.
- the corresponding person image is re-passed through the first target detection sub-network and the second target detection sub-network, and the verification is performed again based on the frame verification layer until The first rectangular frame is verified to ensure that the image of the person can be extracted from each frame of image in the video data.
- Step S202 Detect each of the person images through a preset bone key point detection network to obtain multiple bone key points in each of the person images.
- the server After obtaining multiple person images through the improved target detection network, the server detects each person image through the preset bone key point detection network to obtain multiple bone key points in each person image.
- the bone key points are human bone key points, which are used to represent the human bone structure, and the bone key point detection network is implemented based on a neural network.
- Step S203 Determine the walking information of the target person according to multiple key bone points in each person image.
- the key point is to determine the walking information of the target person.
- the walking information of the target person includes at least one of the walking state, walking duration, walking speed, and walking distance of the target person.
- the walking information of the target person includes the standing state of the target person.
- the standing state includes the upright state and the non-upright state of the target person.
- the torso aspect ratio of the target person is calculated through the key points of the target person's skeleton. When the torso aspect ratio is less than the preset torso aspect ratio, the target person is considered to be in an upright state, and when the torso aspect ratio is greater than or equal to the preset torso aspect ratio, the target person is considered to be in a non-upright state.
- the preset torso aspect ratio can be set according to actual conditions, and this application is not specifically limited here, and can be 0.4.
- the evaluation process will also record the bone key point data of each frame of image.
- the bone key point movement changes during the entire evaluation process can form a playback video.
- the target person's standing state or non-upright state can be used in the playback video after the evaluation. In the display. By determining the standing state of the target person, it is possible to further understand the walking situation of the target person.
- the walking information of the target person includes the body posture data of the target person, and according to the body posture data, the body bending degree C, stride W, and walking speed V of the assessee are determined; according to the body of the assessee
- the degree of curvature C, stride length W, and walking speed V determine the fall index Fe of the subject.
- the degree of body curvature C, the angle between the straight line formed by the neck and thigh joints and the straight line of the leg joints is a dynamic value during the evaluation process, and the maximum detected angle is taken as the degree of body curvature C, stride length W represents the maximum distance between the left and right feet of the subject during walking.
- the correction coefficient is generally about 100.
- the fall index Fe is smaller, and when the body bending degree C is larger, the fall index Fe is larger, and this formula can affect the fall
- the various data are mapped to the interval from 0 to 1 after calculation.
- the fall index indicates the degree to which the target person is prone to fall.
- the fall index can be used as an influencing factor of the walking ability index of the target person. It is understandable that the higher the fall index, the easier the target person will fall, the lower the walking ability index, the lower the fall index, the less likely the target person will fall, and the higher the walking ability index.
- Step S204 Determine the first walking ability index of the target person according to the walking state in the walking information.
- the first walking ability index of the target person corresponding to the walking state is determined. That is, it is determined that the walking state of the target person in each person image is the assisted walking state or the independent walking state, and the first walking state of the corresponding target person is determined according to the assisted walking state and/or the independent walking state in the walking information of the target person Ability index, where the first walking ability index is a preliminarily determined walking ability index.
- the target person according to multiple bone key points of the target person and multiple bone key points of characters other than the target person, it is determined whether the target person has an assisted walking state in the walking information; if the target person is in the walking information If there is an assisted walking state, determine the assisted walking duration of the target person in the walking information according to the bone key points of the target person in each person image and the bone key points of the person other than the target person in each person image; According to the auxiliary walking time of the target person in the walking information, the first walking ability index of the target person is determined.
- the preset walking index can be set according to actual conditions, which is not specifically limited in this application, and can be selected as 100.
- the assisted walking time of the target person in the walking information is 20 seconds
- the total walking time of the video data during the walking process is 30 seconds
- the percentage of the assisted walking time to the total walking time is 60%
- the absolute value of the difference from the percentage is 40%
- the preset walking index is 100
- the first walking ability index is the absolute value of the difference multiplied by the preset walking index equal to 40.
- the first walking ability index of the target person is a preset Walking index.
- the walking index can be set according to actual conditions, and this application is not specifically limited here, and can be selected as 100.
- the target person's walking state is completely assisted in walking, it is understandable that the first walking ability index of the target person is zero.
- Step S205 Determine a second walking ability index of the target person according to the walking speed in the walking information.
- the target person After determining the first walking ability index of the target person, determine the average walking speed of the target person according to the key points of the bones of the target person in each character image, and determine the second walking ability of the target person according to the average walking speed of the target person index. That is, the pre-stored mapping relationship table between walking speed and walking ability index is obtained, and the mapping relationship table is queried, and the walking ability index corresponding to the average walking speed is used as the second walking ability index of the target person. It should be noted that the above-mentioned mapping relationship table between walking speed and walking ability index can be set based on actual conditions, which is not specifically limited in this application. Calculating the average walking speed of the target person can reduce the instability caused by the walking speeds of different sizes, so that the determined second walking ability index of the target person is more accurate.
- the average walking speed of the target person is 0.5m/s.
- the average walking speed corresponds to the second walking ability index of 75, then the second walking ability index of the target person in this walking ability evaluation
- the walking ability index is 75.
- the method for determining the average walking speed of the target person is specifically as follows: select any bone key point as the target bone key point, and arbitrarily select two different frames of person images, and from the selected two different frames of person images Obtain the positions of the key points of the target bone respectively, and obtain two position coordinates through weighted average processing; calculate the time difference between the two selected character images, and determine the time difference of the target character according to the obtained two position coordinates Calculate the average walking speed of the target person according to the walking distance and the time difference.
- the result of the walking ability evaluation is more accurate.
- Step S206 Determine the walking ability index of the target person according to the first walking ability index and the second walking ability index.
- the walking ability index of the target person can be comprehensively determined. Specifically, based on the preset first weighting coefficient and the second weighting coefficient, a weighted average of the first walking ability index and the second walking ability index is calculated to obtain the walking ability index of the target person.
- the specific method for obtaining the above-mentioned weighted average is: adding the product of the first weighting coefficient and the first walking ability index to the second weighting coefficient and the second walking ability index to obtain the walking ability index of the target person.
- the preset first weighting coefficient and second weighting coefficient can be set according to specific conditions, and this application is not specifically limited here.
- the first weighting coefficient and the second weighting coefficient are 0.4 respectively. And 0.6.
- the first weight coefficient and the second weight coefficient are 0.4 and 0.6, respectively, the first walking ability index is 40, and the second walking ability index is 75.
- the first walking ability index and the second walking ability index are weighted. On average, that is, the product of the first weight coefficient 0.4 and the first walking ability index 40 plus the second weight coefficient 0.6 and the second walking ability index 75, the walking ability index of the target person can be obtained as 61.
- the video data processing method provided by the above embodiments accurately obtains multiple bone key points in each person image through the target detection network and the bone key point detection network, and determines the target based on the multiple bone key points in each person image
- the walking information of the character, and the first walking ability index is determined according to the walking state in the walking information, which can improve the accuracy of the subsequent walking ability index.
- the corresponding The second walking ability index Based on the preset mapping table and the walking speed in the walking information, the corresponding The second walking ability index, based on the first walking ability index and the second walking ability index, can accurately determine the walking ability index of the target person, which greatly improves the accuracy of the person's walking ability evaluation.
- FIG. 6 is a schematic block diagram of a video data processing apparatus according to an embodiment of the application.
- the video data processing device 300 includes: an extraction module 301, a detection module 302, a first determination module 303, and a second determination module 304.
- the extraction module 301 is configured to obtain video data, and extract multiple person images from the video data through a preset target detection network, where the video data includes information about the walking situation of the target person;
- the detection module 302 is configured to detect each of the person images through a preset bone key point detection network to obtain multiple bone key points in each of the person images;
- the first determining module 303 is configured to determine the walking information of the target person according to multiple bone key points in each of the person images, where the walking information includes walking state, walking time, walking distance, and walking speed At least one of
- the second determining module 304 is configured to determine the walking ability index of the target person according to the walking information.
- the extraction module 301 is further used for:
- At least one person image is extracted from each frame of image.
- the extraction module 301 is further used for:
- the target rectangular frame of each frame of image is determined.
- the first determining module 303 is further configured to:
- the walking state of the target person is determined according to the multiple bone key points of the target person and the multiple bone key points of the person other than the target person.
- the first determining module 303 includes:
- the first determining sub-module 3031 is configured to determine multiple bone key points of the target person according to the multiple bone key points in each person image;
- the second determining sub-module 3032 is configured to determine the walking information of the target person according to multiple key bone points of the target person.
- the second determining submodule 3032 is further configured to:
- the time required for the walking distance to reach a preset distance threshold is calculated, and the time required for the walking distance to reach the preset distance threshold is taken as the walking time of the target person.
- FIG. 8 is a schematic block diagram of another video data processing apparatus provided by an embodiment of the application.
- the video data processing device 400 includes:
- the extraction module 401 is configured to obtain video data, and extract a plurality of person images from the video data through a preset target detection network, wherein the video data includes the walking situation information of the target person.
- the detection module 402 is configured to detect each person image through a preset bone key point detection network to obtain multiple bone key points in each person image.
- the first determining module 403 is configured to determine the walking information of the target person according to multiple key bone points in each person image.
- the second determining module 404 is configured to determine the first walking ability index of the target person according to the walking state in the walking information.
- the third determining module 405 is configured to determine the second walking ability index of the target person according to the walking speed in the walking information.
- the fourth determining module 406 is configured to determine the walking ability index of the target person according to the first walking ability index and the second walking ability index.
- the apparatus provided in the foregoing embodiment may be implemented in the form of a computer program, and the computer program may run on the computer device as shown in FIG. 9.
- FIG. 9 is a schematic block diagram of the structure of a computer device provided by an embodiment of the application.
- the computer device can be a server or an electronic device.
- the computer device includes a processor, a memory, and a network interface connected through a system bus, where the memory may include a storage medium and an internal memory.
- the non-volatile storage medium can store an operating system and a computer program.
- the computer program includes program instructions, and when the program instructions are executed, the processor can execute any video data processing method.
- the processor is used to provide computing and control capabilities and support the operation of the entire computer equipment.
- the internal memory provides an environment for the operation of the computer program in the storage medium.
- the processor can execute any video data processing method.
- the network interface is used for network communication, such as sending assigned tasks.
- the network interface is used for network communication, such as sending assigned tasks.
- FIG. 9 is only a block diagram of part of the structure related to the solution of the present application, and does not constitute a limitation on the computer device to which the solution of the present application is applied.
- the specific computer device may Including more or fewer parts than shown in the figure, or combining some parts, or having a different arrangement of parts.
- the processor may be a central processing unit (Central Processing Unit, CPU), and the processor may also be other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), and application specific integrated circuits (Application Specific Integrated Circuits). Circuit, ASIC), Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc.
- the general-purpose processor may be a microprocessor or the processor may also be any conventional processor.
- the processor is used to run a computer program stored in a memory to implement the following steps:
- the walking information includes at least one of a walking state, a walking duration, a walking distance, and a walking speed;
- the walking ability index of the target person is determined.
- the processor when the processor is used to determine the walking information of the target person according to multiple key bone points in each of the person images, it is used to achieve:
- the walking information of the target person is determined according to the multiple key bone points of the target person.
- the processor when the processor is used to determine the walking information of the target person according to multiple key bone points in each of the person images, it is used to achieve:
- the walking state of the target person is determined according to the multiple bone key points of the target person and the multiple bone key points of the person other than the target person.
- the processor when the processor is used to determine the walking information of the target person according to multiple key bone points of the target person, it is used to realize:
- the time required for the walking distance to reach a preset distance threshold is calculated, and the time required for the walking distance to reach the preset distance threshold is taken as the walking time of the target person.
- the processor when used to determine the walking ability index of the target person according to the walking information and a preset mapping relationship table, it is used to achieve:
- the walking ability index of the target person is determined.
- the processor realizes that the target detection network includes a first target detection sub-network and a second target detection sub-network; the preset target detection network is used to extract more information from the video data.
- the preset target detection network is used to extract more information from the video data.
- the target rectangular frame of each frame of image is determined; according to the target rectangular frame of each frame of image, at least one person image is extracted from each frame of image.
- the processor when the processor implements the determination of the target rectangular frame of each frame of image according to the first rectangular frame and the second rectangular frame of each frame of image, it is used to implement:
- the target rectangular frame of each frame of image is determined.
- the embodiment of the present application also provides a computer-readable storage medium, the computer-readable storage medium stores a computer program, the computer program includes program instructions, and the method implemented when the program instructions are executed can refer to this Apply for various embodiments of video data processing methods.
- the computer-readable storage medium may be the internal storage unit of the computer device described in the foregoing embodiment, for example, the hard disk or memory of the computer device.
- the computer-readable storage medium may also be an external storage device of the computer device, and the computer-readable storage medium may be non-volatile or volatile, such as a plug-in type provided on the computer device.
- Hard disk Smart Media Card (SMC), Secure Digital (SD) card, Flash Card (Flash Card), etc.
- the video data may also be stored in a node of a blockchain.
- the blockchain referred to in this application is a new application mode of computer technology such as distributed data storage, point-to-point transmission, consensus mechanism, and encryption algorithm.
- Blockchain essentially a decentralized database, is a series of data blocks associated with cryptographic methods. Each data block contains a batch of network transaction information for verification. The validity of the information (anti-counterfeiting) and the generation of the next block.
- the blockchain can include the underlying platform of the blockchain, the platform product service layer, and the application service layer.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Public Health (AREA)
- Medical Informatics (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Psychiatry (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Social Psychology (AREA)
- Human Computer Interaction (AREA)
- Biomedical Technology (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Software Systems (AREA)
- Pathology (AREA)
- Computational Linguistics (AREA)
- Epidemiology (AREA)
- Primary Health Care (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (20)
- 一种视频数据处理方法,其中,包括:获取视频数据,并通过预设的目标检测网络,从所述视频数据中提取多个人物图像,其中,所述视频数据包括目标人物的行走情况信息;通过预设的骨骼关键点检测网络,对每个所述人物图像进行检测,得到每个所述人物图像中的多个骨骼关键点;根据每个所述人物图像中的多个骨骼关键点,确定所述目标人物的行走信息,其中,所述行走信息包括行走状态、行走时长、行走距离和行走速度中的至少一种;根据所述行走信息,确定所述目标人物的行走能力指数。
- 如权利要求1所述的视频数据处理方法,其中,所述根据每个所述人物图像中的多个骨骼关键点,确定所述目标人物的行走信息,包括:根据每个所述人物图像中的多个骨骼关键点,确定所述目标人物的多个骨骼关键点;根据所述目标人物的多个骨骼关键点,确定所述目标人物的行走信息。
- 如权利要求1所述的视频数据处理方法,其中,所述根据每个所述人物图像中的多个骨骼关键点,确定所述目标人物的行走信息,包括:根据每个所述人物图像中的多个骨骼关键点,确定所述目标人物的多个骨骼关键点和除所述目标人物之外的人物的多个骨骼关键点;根据所述目标人物的多个骨骼关键点和除所述目标人物之外的人物的多个骨骼关键点,确定所述目标人物的行走状态。
- 如权利要求2所述的视频数据处理方法,其中,所述根据所述目标人物的多个骨骼关键点,确定所述目标人物的行走信息,包括:选定所述目标人物的任意一个骨骼关键点作为目标骨骼关键点,并获取所述目标骨骼关键点在每个所述人物图像中的位置信息;根据所述目标骨骼关键点在每个所述人物图像中的位置信息,确定所述目标人物的行走距离;计算所述行走距离达到预设的距离阈值所需的时长,将所述行走距离达到预设的距离阈值所需的时长作为所述目标人物的行走时长。
- 如权利要求1所述的视频数据处理方法,其中,所述根据所述行走信息,确定所述目标人物的行走能力指数,包括:根据所述行走信息中的行走状态,确定所述目标人物的第一行走能力指数;根据所述行走信息中的行走速度,确定所述目标人物的第二行走能力指数;根据所述第一行走能力指数和所述第二行走能力指数,确定所述目标人物的行走能力指数。
- 如权利要求1-5中任一项所述的视频数据处理方法,其中,所述目标检测网络包括第一目标检测子网络和第二目标检测子网络;所述通过预设的目标检测网络,从所述视频数据中提取多个人物图像,包括:将所述视频数据中的每帧图像分别输入至所述第一目标检测子网络和所述第二目标检测子网络,得到每帧图像的第一矩形框和第二矩形框;根据每帧图像的第一矩形框和第二矩形框,确定每帧图像的目标矩形框;根据每帧图像的目标矩形框,从每帧图像中提取出至少一个人物图像。
- 如权利要求6所述的视频数据处理方法,其中,所述根据每帧图像的第一矩形框和第二矩形框,确定每帧图像的目标矩形框,包括:分别获取每帧图像的所述第一矩形框的位置信息和所述第二矩形框的位置信息;将每帧图像的所述第一矩形框的位置信息与对应的所述第二矩形框的位置信息进行 比较,得到每帧图像的矩形框比较结果;根据每帧图像的矩形框比较结果、第一矩形框的位置信息和第二矩形框的位置信息,确定每帧图像的目标矩形框。
- 一种视频数据处理装置,其中,所述视频数据处理装置包括:提取模块,用于获取视频数据,并通过预设的目标检测网络,从所述视频数据中提取多个人物图像,其中,所述视频数据包括目标人物的行走情况信息;检测模块,用于通过预设的骨骼关键点检测网络,对每个所述人物图像进行检测,得到每个所述人物图像中的多个骨骼关键点;第一确定模块,用于根据每个所述人物图像中的多个骨骼关键点,确定所述目标人物的行走信息,其中,所述行走信息包括行走状态、行走时长、行走距离和行走速度中的至少一种;第二确定模块,用于根据所述行走信息,确定所述目标人物的行走能力指数。
- 一种计算机设备,其中,所述计算机设备包括处理器、存储器、以及存储在所述存储器上并可被所述处理器执行的计算机程序,其中所述计算机程序被所述处理器执行时,实现如下的步骤:获取视频数据,并通过预设的目标检测网络,从所述视频数据中提取多个人物图像,其中,所述视频数据包括目标人物的行走情况信息;通过预设的骨骼关键点检测网络,对每个所述人物图像进行检测,得到每个所述人物图像中的多个骨骼关键点;根据每个所述人物图像中的多个骨骼关键点,确定所述目标人物的行走信息,其中,所述行走信息包括行走状态、行走时长、行走距离和行走速度中的至少一种;根据所述行走信息,确定所述目标人物的行走能力指数。
- 如权利要求9所述的计算机设备,其中,所述根据每个所述人物图像中的多个骨骼关键点,确定所述目标人物的行走信息,包括:根据每个所述人物图像中的多个骨骼关键点,确定所述目标人物的多个骨骼关键点;根据所述目标人物的多个骨骼关键点,确定所述目标人物的行走信息。
- 如权利要求9所述的计算机设备,其中,所述根据每个所述人物图像中的多个骨骼关键点,确定所述目标人物的行走信息,包括:根据每个所述人物图像中的多个骨骼关键点,确定所述目标人物的多个骨骼关键点和除所述目标人物之外的人物的多个骨骼关键点;根据所述目标人物的多个骨骼关键点和除所述目标人物之外的人物的多个骨骼关键点,确定所述目标人物的行走状态。
- 如权利要求10所述的计算机设备,其中,所述根据所述目标人物的多个骨骼关键点,确定所述目标人物的行走信息,包括:选定所述目标人物的任意一个骨骼关键点作为目标骨骼关键点,并获取所述目标骨骼关键点在每个所述人物图像中的位置信息;根据所述目标骨骼关键点在每个所述人物图像中的位置信息,确定所述目标人物的行走距离;计算所述行走距离达到预设的距离阈值所需的时长,将所述行走距离达到预设的距离阈值所需的时长作为所述目标人物的行走时长。
- 如权利要求9所述的计算机设备,其中,所述根据所述行走信息,确定所述目标人物的行走能力指数,包括:根据所述行走信息中的行走状态,确定所述目标人物的第一行走能力指数;根据所述行走信息中的行走速度,确定所述目标人物的第二行走能力指数;根据所述第一行走能力指数和所述第二行走能力指数,确定所述目标人物的行走能力 指数。
- 如权利要求9-13中任一项所述的计算机设备,其中,所述目标检测网络包括第一目标检测子网络和第二目标检测子网络;所述通过预设的目标检测网络,从所述视频数据中提取多个人物图像,包括:将所述视频数据中的每帧图像分别输入至所述第一目标检测子网络和所述第二目标检测子网络,得到每帧图像的第一矩形框和第二矩形框;根据每帧图像的第一矩形框和第二矩形框,确定每帧图像的目标矩形框;根据每帧图像的目标矩形框,从每帧图像中提取出至少一个人物图像。
- 如权利要求14所述的计算机设备,其中,所述根据每帧图像的第一矩形框和第二矩形框,确定每帧图像的目标矩形框,包括:分别获取每帧图像的所述第一矩形框的位置信息和所述第二矩形框的位置信息;将每帧图像的所述第一矩形框的位置信息与对应的所述第二矩形框的位置信息进行比较,得到每帧图像的矩形框比较结果;根据每帧图像的矩形框比较结果、第一矩形框的位置信息和第二矩形框的位置信息,确定每帧图像的目标矩形框。
- 一种计算机可读存储介质,其中,所述计算机可读存储介质上存储有计算机程序,其中所述计算机程序被处理器执行时,实现如下步骤:获取视频数据,并通过预设的目标检测网络,从所述视频数据中提取多个人物图像,其中,所述视频数据包括目标人物的行走情况信息;通过预设的骨骼关键点检测网络,对每个所述人物图像进行检测,得到每个所述人物图像中的多个骨骼关键点;根据每个所述人物图像中的多个骨骼关键点,确定所述目标人物的行走信息,其中,所述行走信息包括行走状态、行走时长、行走距离和行走速度中的至少一种;根据所述行走信息,确定所述目标人物的行走能力指数。
- 如权利要求16所述的计算机可读存储介质,其中,所述根据每个所述人物图像中的多个骨骼关键点,确定所述目标人物的行走信息,包括:根据每个所述人物图像中的多个骨骼关键点,确定所述目标人物的多个骨骼关键点;根据所述目标人物的多个骨骼关键点,确定所述目标人物的行走信息。
- 如权利要求16所述的计算机可读存储介质,其中,所述根据每个所述人物图像中的多个骨骼关键点,确定所述目标人物的行走信息,包括:根据每个所述人物图像中的多个骨骼关键点,确定所述目标人物的多个骨骼关键点和除所述目标人物之外的人物的多个骨骼关键点;根据所述目标人物的多个骨骼关键点和除所述目标人物之外的人物的多个骨骼关键点,确定所述目标人物的行走状态。
- 如权利要求17所述的计算机可读存储介质,其中,所述根据所述目标人物的多个骨骼关键点,确定所述目标人物的行走信息,包括:选定所述目标人物的任意一个骨骼关键点作为目标骨骼关键点,并获取所述目标骨骼关键点在每个所述人物图像中的位置信息;根据所述目标骨骼关键点在每个所述人物图像中的位置信息,确定所述目标人物的行走距离;计算所述行走距离达到预设的距离阈值所需的时长,将所述行走距离达到预设的距离阈值所需的时长作为所述目标人物的行走时长。
- 如权利要求16所述的计算机可读存储介质,其中,所述根据所述行走信息,确定所述目标人物的行走能力指数,包括:根据所述行走信息中的行走状态,确定所述目标人物的第一行走能力指数;根据所述行走信息中的行走速度,确定所述目标人物的第二行走能力指数;根据所述第一行走能力指数和所述第二行走能力指数,确定所述目标人物的行走能力指数。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010032164.8A CN111242030A (zh) | 2020-01-13 | 2020-01-13 | 视频数据处理方法、装置、设备及计算机可读存储介质 |
CN202010032164.8 | 2020-01-13 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021143103A1 true WO2021143103A1 (zh) | 2021-07-22 |
Family
ID=70875988
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/105904 WO2021143103A1 (zh) | 2020-01-13 | 2020-07-30 | 视频数据处理方法、装置、设备及计算机可读存储介质 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN111242030A (zh) |
WO (1) | WO2021143103A1 (zh) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113537121A (zh) * | 2021-07-28 | 2021-10-22 | 浙江大华技术股份有限公司 | 身份识别方法和装置、存储介质及电子设备 |
CN113727187A (zh) * | 2021-08-31 | 2021-11-30 | 平安科技(深圳)有限公司 | 基于骨架迁移的动画视频处理方法、装置及相关设备 |
WO2023185034A1 (zh) * | 2022-03-31 | 2023-10-05 | 上海商汤智能科技有限公司 | 动作检测方法、装置、电子设备及存储介质 |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111242030A (zh) * | 2020-01-13 | 2020-06-05 | 平安国际智慧城市科技股份有限公司 | 视频数据处理方法、装置、设备及计算机可读存储介质 |
CN113902084A (zh) * | 2020-07-06 | 2022-01-07 | 阿里体育有限公司 | 运动计数方法、装置、电子设备及计算机存储介质 |
CN111767888A (zh) * | 2020-07-08 | 2020-10-13 | 北京澎思科技有限公司 | 对象状态检测方法、计算机设备、存储介质和电子设备 |
CN112597898B (zh) * | 2020-12-24 | 2021-11-23 | 珠高电气检测有限公司 | 基于实时计步的电力人员安全状态智能识别方法及介质 |
CN112597903B (zh) * | 2020-12-24 | 2021-08-13 | 珠高电气检测有限公司 | 基于步幅测量的电力人员安全状态智能识别方法及介质 |
CN113705482B (zh) * | 2021-08-31 | 2024-03-22 | 江苏唯宝体育科技发展有限公司 | 一种基于人工智能的身体健康监测管理系统及方法 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005287806A (ja) * | 2004-03-31 | 2005-10-20 | Biophilia Kenkyusho Kk | 立位歩行機能評価判定標準化技術 |
CN104346606A (zh) * | 2014-10-30 | 2015-02-11 | 东北大学 | 异常步态分析方法及系统 |
CN109919137A (zh) * | 2019-03-28 | 2019-06-21 | 广东省智能制造研究所 | 一种行人结构化特征表达方法 |
CN111242030A (zh) * | 2020-01-13 | 2020-06-05 | 平安国际智慧城市科技股份有限公司 | 视频数据处理方法、装置、设备及计算机可读存储介质 |
-
2020
- 2020-01-13 CN CN202010032164.8A patent/CN111242030A/zh active Pending
- 2020-07-30 WO PCT/CN2020/105904 patent/WO2021143103A1/zh active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005287806A (ja) * | 2004-03-31 | 2005-10-20 | Biophilia Kenkyusho Kk | 立位歩行機能評価判定標準化技術 |
CN104346606A (zh) * | 2014-10-30 | 2015-02-11 | 东北大学 | 异常步态分析方法及系统 |
CN109919137A (zh) * | 2019-03-28 | 2019-06-21 | 广东省智能制造研究所 | 一种行人结构化特征表达方法 |
CN111242030A (zh) * | 2020-01-13 | 2020-06-05 | 平安国际智慧城市科技股份有限公司 | 视频数据处理方法、装置、设备及计算机可读存储介质 |
Non-Patent Citations (1)
Title |
---|
YAN, TIEBIN: "Timed "Up-Go" Test for Evaluating the Reliability and Concurrent Validity of Walking Function in Stroke Patients", CHINESE JOURNAL OF REHABILITATION THEORY AND PRACTICE, vol. 6, no. 3, 25 September 2000 (2000-09-25), CN, pages 115 - 117, XP009529277, ISSN: 1006-9771 * |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113537121A (zh) * | 2021-07-28 | 2021-10-22 | 浙江大华技术股份有限公司 | 身份识别方法和装置、存储介质及电子设备 |
CN113727187A (zh) * | 2021-08-31 | 2021-11-30 | 平安科技(深圳)有限公司 | 基于骨架迁移的动画视频处理方法、装置及相关设备 |
CN113727187B (zh) * | 2021-08-31 | 2022-10-11 | 平安科技(深圳)有限公司 | 基于骨架迁移的动画视频处理方法、装置及相关设备 |
WO2023185034A1 (zh) * | 2022-03-31 | 2023-10-05 | 上海商汤智能科技有限公司 | 动作检测方法、装置、电子设备及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
CN111242030A (zh) | 2020-06-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021143103A1 (zh) | 视频数据处理方法、装置、设备及计算机可读存储介质 | |
CN110276366A (zh) | 使用弱监督模型来检测对象 | |
WO2021114892A1 (zh) | 基于环境语义理解的人体行为识别方法、装置、设备及存储介质 | |
Ahmed et al. | Social distance monitoring framework using deep learning architecture to control infection transmission of COVID-19 pandemic | |
Mehmood et al. | Efficient image recognition and retrieval on IoT-assisted energy-constrained platforms from big data repositories | |
CN111832468B (zh) | 基于生物识别的手势识别方法、装置、计算机设备及介质 | |
WO2018028546A1 (zh) | 一种关键点的定位方法及终端、计算机存储介质 | |
WO2020259213A1 (zh) | 行为识别的方法、装置、终端设备及存储介质 | |
WO2018228218A1 (zh) | 身份识别方法、计算设备及存储介质 | |
WO2020107847A1 (zh) | 基于骨骼点的跌倒检测方法及其跌倒检测装置 | |
WO2019071660A1 (zh) | 票据信息识别方法、电子装置及可读存储介质 | |
WO2021151338A1 (zh) | 医学影像图片分析方法、装置、电子设备及可读存储介质 | |
WO2022105118A1 (zh) | 基于图像的健康状态识别方法、装置、设备及存储介质 | |
WO2021051547A1 (zh) | 暴力行为检测方法及系统 | |
WO2021057316A1 (zh) | 一种人体完整度数据标注方法、装置及终端设备 | |
CN111986194B (zh) | 医学标注图像检测方法、装置、电子设备及存储介质 | |
WO2021082692A1 (zh) | 一种行人图片标注方法、装置、存储介质和智能设备 | |
WO2021184847A1 (zh) | 一种遮挡车牌字符识别方法、装置、存储介质和智能设备 | |
WO2023109631A1 (zh) | 数据处理方法、装置、设备、存储介质及程序产品 | |
WO2021223738A1 (zh) | 模型参数的更新方法、装置、设备及存储介质 | |
WO2021217937A1 (zh) | 姿态识别模型的训练方法及设备、姿态识别方法及其设备 | |
JP2022542199A (ja) | キーポイントの検出方法、装置、電子機器および記憶媒体 | |
CN113177133B (zh) | 一种图像检索方法、装置、设备及存储介质 | |
WO2021068613A1 (zh) | 面部识别方法、装置、设备及计算机可读存储介质 | |
CN111353325A (zh) | 关键点检测模型训练方法及装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20913430 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20913430 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205 DATED 15.03.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20913430 Country of ref document: EP Kind code of ref document: A1 |