CN115410121A - Video-based automatic determination method for joint seal person, electronic device and storage medium - Google Patents

Video-based automatic determination method for joint seal person, electronic device and storage medium Download PDF

Info

Publication number
CN115410121A
CN115410121A CN202210987107.4A CN202210987107A CN115410121A CN 115410121 A CN115410121 A CN 115410121A CN 202210987107 A CN202210987107 A CN 202210987107A CN 115410121 A CN115410121 A CN 115410121A
Authority
CN
China
Prior art keywords
image
human body
target object
bid
coordinate system
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210987107.4A
Other languages
Chinese (zh)
Inventor
杨玉春
胡东平
叶新江
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Merit Interactive Co Ltd
Original Assignee
Merit Interactive Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Merit Interactive Co Ltd filed Critical Merit Interactive Co Ltd
Priority to CN202210987107.4A priority Critical patent/CN115410121A/en
Publication of CN115410121A publication Critical patent/CN115410121A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Software Systems (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Human Computer Interaction (AREA)
  • Computational Linguistics (AREA)
  • Closed-Circuit Television Systems (AREA)

Abstract

The invention provides a video-based automatic judgment method for tight-lock personnel, which comprises the steps of firstly, acquiring a corresponding monitoring video based on related information of a target object, then acquiring all images containing the target object from the monitoring video, then, generating action tracks of the target object based on all the acquired images for protecting the target object, and then, acquiring personnel with space intersection with the target object for each frame of image in each action track. The invention also provides electronic equipment and a storage medium. The invention can quickly and accurately find out the personnel with the intersection with the target object in the time space.

Description

Video-based automatic judgment method for contact person, electronic equipment and storage medium
Technical Field
The present invention relates to the field of data processing, and in particular, to a method for automatically determining a person in contact with a skin, an electronic device, and a storage medium.
Background
In some application scenarios, it is necessary to acquire related people having intersection with the target people on the temporal-spatial information. The current methods for acquiring related people mainly include two methods, one is to manually collect the action track of the target person and then to visit the information of the related people on the action track as much as possible; the other method is to determine whether the person is the related person through a set identifier on the mobile terminal, and then acquire information of the related person based on the acquired action information and the action track of the target person. The first method can acquire relatively accurate personnel information, but is time-consuming and labor-consuming. Compared with the first mode, the second mode uses an information means to acquire the information of the related personnel, so that the time and the labor can be saved, but the acquired information of the related personnel is not accurate due to low information density and large position error.
Disclosure of Invention
Aiming at the technical problem, the technical scheme adopted by the invention is as follows:
the embodiment of the invention provides a video-based automatic judgment method for joint sealing personnel, which comprises the following steps:
s100, acquiring a first monitoring video about a target object;
s110, decoding the first monitoring video to obtain n1 frame images; identifying the human body information in any frame image i in the n1 frame images to obtain an identification information table C of the image i i ,C i The j-th row in (b) includes identification information (BID) of the j-th human body identified in the image i ij ,P ij ,G ij ,t i ,CID i ),BID ij For the ID, P of the j-th body identified in image i ij For the image feature vector of the j-th body identified in image i, G ij Is the position, t, of the j-th body identified in the image i i CID as the photographing time of image i i ID of the camera for image i; j takes values from 1 to m (i), and m (i) is the number of human bodies identified in the image i; the value of i is 1 to n1;
s120, based on any identification information table C i Obtaining a similarity set S i =(S i1 ,S i2 ,…,S ij ,…,S im(i) ) Wherein S is ij Is C i P in (1) ij Similarity with an image feature vector P in the image information of the target object;
s130, if max (S) i ) If the number of the images is more than or equal to K, storing the image i into the first target image set; k is a set similarity threshold;
s140, acquiring a motion track sequence of the target object based on the first target image set;
s150, for any frame image r in t (p) frame images corresponding to any action track sequence p, identifying information table C based on image r r Acquiring the distance between any human body except the target object in the image r and the target object in a world coordinate system;
s160, acquiring identification information corresponding to the distance within the set distance threshold value, and generating a first set;
s170, carrying out duplicate removal processing on the information in the first set to obtain a first target set.
The present invention also provides a non-transitory computer readable storage medium having stored therein at least one instruction or at least one program, the at least one instruction or the at least one program being loaded and executed by a processor to implement the aforementioned method.
The invention also provides an electronic device comprising a processor and the aforementioned non-transitory computer-readable storage medium.
The invention has at least the following beneficial effects:
the automatic judgment method for the tight-lock person based on the video comprises the steps of firstly obtaining a corresponding monitoring video based on relevant information of a target object, then obtaining all images containing the target object from the monitoring video, then generating action tracks of the target object based on all the obtained images including the target object, and then obtaining a person with space intersection with the target object for each frame of image in each action track.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present invention, the drawings needed to be used in the description of the embodiments will be briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
Fig. 1 is a flowchart of a method for automatically determining a person in close contact based on a video according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Fig. 1 is a flowchart of a method for automatically determining a person in close contact based on a video according to an embodiment of the present invention.
The embodiment of the invention provides a video-based automatic judgment method for a joint sealing person, which comprises the following steps of:
s100, a first monitoring video about the target object is obtained.
In an exemplary embodiment of the invention, the target object may be a person who needs to be restricted from freedom for a particular period of time, for example, in one particular application scenario, the target object may be a person diagnosed as having an infectious disease.
In an embodiment of the present invention, the first surveillance video may be obtained based on the identity information of the target object and the geographic area passed by within the set time period. The identity information of the target object includes image information of the target object. Those skilled in the art know that it is prior art to acquire corresponding surveillance video based on the identity information of the target object and the geographic area that has passed within a set period of time. The set time period may be determined based on actual conditions, and may be, for example, 7 days or 14 days, etc.
Further, the corresponding image feature vector P may also be acquired based on the image information of the target object. High-dimensional feature vectors can be extracted from image information through an image feature extractor, and corresponding image feature vectors are obtained. Those skilled in the art know that extracting feature vectors of an image, forming a set of image feature vectors, may be a prior art, for example, extracting by means of a deep neural network, etc.
S110, decoding the first monitoring video to obtain n1 frame images; identifying the human body information in any frame image i in the n1 frame images to obtain an identification information table C of the image i i ,C i The j-th row in (b) includes identification information (BID) of the j-th human body identified in the image i ij ,P ij ,G ij ,t i ,CID i ),BID ij ID, P of the j-th body identified in image i ij For the image feature vector of the j-th body identified in image i, G ij Is the position, t, of the j-th body identified in the image i i CID as the photographing time of image i i ID of the camera for image i; j takes values from 1 to m (i), and m (i) is the number of human bodies identified in the image i; i takes on values from 1 to n1.
In an embodiment of the present invention, the ID of each human body may be customized or generated in a time sequence appearing in the video data, for example, assigned in sequence in a natural number sequence, and the like.
In the embodiment of the invention, each human body in each frame of image is automatically detected by the neural network detector, and the detection frame of the human body is identified. Wherein G is ij The midpoint position of the bottom side of the detection frame for the jth pedestrian identified in the image i. G ij =(x ij ,y ij ),x ij For the abscissa, y, of the jth individual in the image i in the image coordinate system ij Is the ordinate of the jth individual in image i in the image coordinate system, in pixels.
S120, based on any identification information table C i Obtaining a similarity set S i =(S i1 ,S i2 ,…,S ij ,…,S im(i) ) Wherein S is ij Is C i P in (1) ij And the similarity with the image feature vector P in the image information of the target object.
In the embodiment of the present invention, the similarity between two image feature vectors, for example, a cosine distance, a euclidean distance, a hamming distance, and the like, may be calculated by a set similarity calculation method.
In the embodiment of the invention, the camera device can be a camera.
S130, if max (S) i ) If the number of the images is more than or equal to K, storing the image i into the first target image set; and K is a set similarity threshold.
If max (S) i ) And K or more, the human body corresponding to the target object is found in the image i. In the embodiment of the invention, the first target image set is stored as ordered data, and is initialized to be a list containing a head element, and the length of the list can be expanded.
In the embodiment of the present invention, K may be determined based on an existing manner, for example, K may be 0.85 to 0.95.
S140, acquiring a motion track sequence of the target object based on the first target image set.
In the embodiment of the present invention, image slices of a target object with temporal and spatial continuity can be integrated into a continuous track sequence based on a target tracking algorithm. It is known to those skilled in the art that, based on a target tracking algorithm, the integration of image slices of a target object with temporal and spatial continuity into one continuous trajectory sequence may be an existing method. Each action track sequence comprises the shooting time of i in any image in the first target image set and max (S) in any image i i ) The corresponding ID, image feature vector and position of the human body.
Through S140, the data of the target object in the continuous frames can be deduplicated, and the subsequent processing efficiency is improved.
S150, for any moving railAny frame image r in t (p) frame images corresponding to the trace sequence p, and identification information table C based on the image r r And acquiring the distance between any human body except the target object in the image r and the target object in the world coordinate system.
Further, S150 may specifically include:
s151, for any frame image r in t (p) frame images corresponding to any action track sequence p, identifying information table C based on image r r Obtaining a set of position information RG p r =(RG p r1 ,RG p r2 ,…,RG p s ,…,RG p rm(r) ),RG p rs For the position G of the s-th body in the image r rs Mapping to a location in a world coordinate system; the value of r is 1 to t (p), the value of s is 1 to m (r), the value of p is 1 to H, and H is the number of the action track sequences.
The person skilled in the art knows that deriving the position in the world coordinate system based on the position in the image coordinate system may be prior art, e.g. by means of camera calibration. Camera calibration may be prior art and may for example comprise the steps of: (1) making a calibration chessboard; (2) Shooting an image of the calibration chessboard by a camera to serve as a first test image; (3) calculating camera internal parameters based on the first test image; (4) Arranging a plurality of coordinate marks on the ground of an area actually shot by a camera; (5) Shooting an image of the ground of the actually shot area as a second test image; (6) Acquiring camera external parameters according to the camera internal parameters, the positions of the coordinate marks in the world coordinate system and the positions in the second test image; (7) And calculating a coordinate transformation formula based on the camera internal parameters and the camera external parameters.
S152, based on any position information set RG p r Obtaining a distance set D p r =(D p r1 ,D p r2 ,…,D p rt ,…,D p r(m(r)-1) ) Wherein D is p rt For the distance between the position of the tth individual in the image r, except the target object, in the world coordinate system and the position of the target object in the world coordinate systemSeparating; t is 1 to m (r) -1.
In an embodiment of the present invention, the distance between the two positions may be a euclidean distance.
S160, acquiring the identification information corresponding to the distance within the set distance threshold, and generating a first set.
Further, S160 may specifically include:
traverse D p r If D is p rt If D is less than or equal to D, then D is added p rt Storing the corresponding identification information of the human body into a first set; generating a first set; d is a set distance threshold.
In the embodiment of the invention, the first set stores ordered data, the ordered data is initialized to a set containing a header element, and the length of combination can be expanded. The set distance threshold may be determined based on actual conditions, for example, may be 5 meters.
S170, carrying out duplicate removal processing on the information in the first set to obtain a first target set.
The technical effect of S100 to S170 is that, based on the image of the target object, a person who is spatially intersected with the target object in the presence of the target object is obtained, and a person who is temporally and spatially associated with the target object (hereinafter referred to as a first associated person for convenience of description) can be quickly and accurately found.
Further, the method provided by the embodiment of the invention further comprises the following steps:
s200, aiming at any BID in the first target set q The following operations are performed:
s210, acquiring BID q And a second monitoring video of the geographic area where the corresponding human body passes in a second specified time period.
The second specified time period may be determined based on actual circumstances, and may be, for example, 7 days or 14 days, etc.
S220, decoding the second monitoring video to obtain an n2 frame image; identifying the human body information in any frame image a in the n2 frame images to obtain an identification information table C of the image a a ,C a The b-th line in (b) includes identification information (BID) of the b-th human body identified in the image a ab ,P ab ,G ab ,t a ,CID a ),BID ab ID, P, of the b-th body identified in image a ab For the image feature vector of the b-th body identified in image a, G ab Is the position of the b-th body identified in image a in image i, t a CID as the photographing time of image a a ID of the camera for image a; the value of b is 1 to m (a), and m (a) is the number of human bodies identified in the image a; and a takes a value from 1 to n2.
S230, based on any identification information table C a Obtaining a similarity set S a =(S a1 ,S a2 ,…,S ab ,…,S a(m(a)-1) ) Wherein S is ab Is C a P in (1) ab And BID q Corresponding image feature vector P q The similarity between them.
S240, if max (S) a ) If the number of the images a is more than or equal to K, storing the images a into a second target image set; k is a set similarity threshold, and may be, for example, 0.85 to 0.95.
S250, acquiring BID based on the second target image set q And (4) corresponding action track sequence of the human body.
The specific implementation of S220 to S250 can refer to the foregoing S110 to S140, and detailed descriptions thereof are omitted for avoiding redundancy.
S260, identifying information table C of any frame image e in t (p 1) frame images corresponding to any action track sequence p1 acquired in S250 e Obtaining the target object and BID in the image e q Any human body other than the corresponding human body and BID q And the corresponding human body is in distance in the world coordinate system, and identification information corresponding to the distance within the set distance threshold is acquired to generate a second set.
Further, S260 may specifically include:
s261, if C e Including the ID of the target object, a first information table C is obtained 1 a Executing S262; otherwise, executing S263; c 1 a Is C a And an information table obtained after the identification information of the target object is deleted.
S263, obtaining a first position information set RG1 p1 e =(RG1 p1 e ,RG1 p1 e ,…,RG1 p1 s1 ,…,RG1 p1 e(m(e)-1) ),RG1 p1 es1 Is C 1 a S1 th human body position G in (1) es1 Mapping to a location in a world coordinate system; the value of e is 1 to t (p 1), the value of S1 is 1 to m (e) -1, the value of p1 is 1 to H1, and H1 is the number of the action track sequences obtained in S250; s264 is executed.
S263, obtain the second position information set RG2 p1 e =(RG2 p1 e ,RG2 p1 e ,…,RG2 p1 s2 ,…,RG2 p1 em(e) ),RG1 p1 es2 Is C a S2 th human body position G es2 Mapping to a location in a world coordinate system; s2 takes on a value of 1 to m (e); s265 is executed.
S264, based on any first position information set RG1 p1 e Obtaining a first distance set D1 p1 e =(D1 p1 e1 ,D1 p1 e2 ,…,D1 p1 et1 ,…,D1 p1 e(m(e)-2) ) Wherein D1 p1 et1 For dividing BID in image e q Corresponding position and BID of t1 th person except human body in world coordinate system q The distance between the positions of the corresponding human body in the world coordinate system; t1 takes the value from 1 to m (r) -2; s266 is executed.
S265, based on any second position information set RG2 p1 e Obtaining a second distance set D2 p1 e =(D2 p1 e1 ,D2 p1 e2 ,…,D2 p1 et2 ,…,D2 p1 e(m(e)-1) ) Wherein, D2 p1 et2 For dividing BID in image e q Corresponding position and BID of t2 th person except human body in world coordinate system q The distance between the positions of the corresponding human body in the world coordinate system; t2 takes the value from 1 to m (r) -1; s267 is performed.
S266, traverse D1 p1 e If D1 is p1 et1 If D is less than or equal to D, then D1 is added p1 et1 The corresponding human body identification information is stored into the second set.
S267, go through D2 p1 e If D2 is p1 et2 D is less than or equal to D, then D2 is added p1 et2 The corresponding human body identification information is stored into the second set.
In the embodiment of the invention, the second set stores ordered data, the ordered data is initialized to a set containing a header element, and the length of combination can be expanded.
The technical effect of S210 to S260 is that, compared to the foregoing embodiment, the person associated with each first associated person, i.e., the second associated person associated with the target object, can be further confirmed, so that all persons associated with the target object can be found as completely as possible to provide accurate information for the relevant department.
Embodiments of the present invention also provide a non-transitory computer-readable storage medium, which may be configured in an electronic device to store at least one instruction or at least one program for implementing a method of the method embodiments, where the at least one instruction or the at least one program is loaded into and executed by a processor to implement the method provided by the above embodiments.
Embodiments of the present invention also provide an electronic device comprising a processor and the aforementioned non-transitory computer-readable storage medium.
Embodiments of the present invention also provide a computer program product comprising program code means for causing an electronic device to carry out the steps of the method according to various exemplary embodiments of the invention described above in the present description, when said program product is run on the electronic device.
Although some specific embodiments of the present invention have been described in detail by way of illustration, it should be understood by those skilled in the art that the above illustration is only for the purpose of illustration and is not intended to limit the scope of the invention. It will also be appreciated by those skilled in the art that various modifications may be made to the embodiments without departing from the scope and spirit of the invention. The scope of the invention is defined by the appended claims.

Claims (9)

1. A video-based automatic judgment method for a joint sealing person is characterized by comprising the following steps:
s100, acquiring a first monitoring video about a target object;
s110, decoding the first monitoring video to obtain n1 frame images; identifying the human body information in any frame image i in the n1 frame images to obtain an identification information table C of the image i i ,C i The jth line in (b) includes identification information (BID) of the jth human body identified in the image i ij ,P ij ,G ij ,t i ,CID i ),BID ij For the ID, P of the j-th body identified in image i ij For the image feature vector of the j-th body identified in image i, G ij Is the position, t, of the j-th body identified in the image i i CID as the photographing time of image i i ID of the camera for image i; j takes values from 1 to m (i), and m (i) is the number of human bodies identified in the image i; the value of i is 1 to n1;
s120, based on any identification information table C i Obtaining a similarity set S i =(S i1 ,S i2 ,…,S ij ,…,S im(i) ) Wherein S is ij Is C i P in (1) ij Similarity with an image feature vector P in the image information of the target object;
s130, if max (S) i ) If the number of the images is more than or equal to K, storing the image i into a first target image set; k is a set similarity threshold;
s140, acquiring a motion track sequence of the target object based on the first target image set;
s150, for any frame image r in t (p) frame images corresponding to any action track sequence p, identifying information table C based on image r r Acquiring any human body except the target object in the image r and the target object in the worldDistance in a coordinate system;
s160, acquiring identification information corresponding to the distance within the set distance threshold value, and generating a first set;
s170, carrying out duplicate removal processing on the information in the first set to obtain a first target set.
2. The method according to claim 1, wherein in S100, the first surveillance video is obtained based on the identity information of the target object and a geographical area that has passed within a set period of time.
3. The method according to claim 1, wherein S150 specifically comprises:
s151, for any frame image r in t (p) frame images corresponding to any action track sequence p, identifying information table C based on image r r Obtaining a set of position information RG p r =(RG p r1 ,RG p r2 ,…,RG p s ,…,RG p rm(r) ),RG p rs As the position G of the s-th body in the image r rs Mapping to a location in a world coordinate system; the value of r is 1 to t (p), the value of s is 1 to m (r), the value of p is 1 to H, and H is the number of the action track sequences;
s152, based on any position information set RG p r Obtaining a distance set D p r =(D p r1 ,D p r2 ,…,D p rt ,…,D p r(m(r)-1) ) Wherein D is p rt The distance between the position of the tth individual in the image r except the target object in the world coordinate system and the position of the target object in the world coordinate system; t is 1 to m (r) -1.
4. The method according to claim 1, wherein S160 specifically comprises:
traverse D p r If D is p rt If D is less than or equal to D, then D is added p rt Storing the corresponding identification information of the human body into a first set; generating a first set; d is a set distance threshold.
5. The method of claim 1, further comprising the steps of:
s200, aiming at any BID in the first target set q The following operations are performed:
s210, acquiring BID q A second monitoring video of a geographic area where the corresponding human body passes in a second designated time period;
s220, decoding the second monitoring video to obtain n2 frame images; identifying the human body information in any frame image a in the n2 frame images to obtain an identification information table C of the image a a ,C a The b-th line in (b) includes identification information (BID) of the b-th human body identified in the image a ab ,P ab ,G ab ,t a ,CID a ),BID ab For the b-th body ID, P identified in image a ab For the image feature vector of the b-th body identified in image a, G ab Is the position of the b-th body identified in image a in image i, t a CID as the photographing time of image a a ID of the camera for image a; the value of b is 1 to m (a), and m (a) is the number of human bodies identified in the image a; the value of a is 1 to n2;
s230, based on any identification information table C a Obtaining a similarity set S a =(S a1 ,S a2 ,…,S ab ,…,S a(m(a)-1) ) Wherein S is ab Is C a P in (1) ab And BID q Corresponding image feature vector P q The similarity between them;
s240, go to if max (S) a ) If the number of the images a is more than or equal to K, storing the images a into a second target image set;
s250, acquiring BID based on the second target image set q Corresponding action track sequence of human body;
s260, identifying information table of any frame image e in t (p 1) frame images corresponding to any action track sequence p1 acquired in S250C e Obtaining the target object and BID in the image e q Any human body other than the corresponding human body and BID q The distance of the corresponding human body in the world coordinate system is obtained, identification information corresponding to the distance within a set distance threshold is obtained, and a second set is generated;
and S270, carrying out duplicate removal processing on the information in the second set to obtain a second target set.
6. The method according to claim 5, wherein S260 specifically comprises:
s261, if C e Including the ID of the target object, a first information table C is obtained 1 a Executing S262; otherwise, executing S263; c 1 a Is C a An information table obtained after the identification information of the target object is deleted;
s262, obtaining a first position information set RG1 p1 e =(RG1 p1 e ,RG1 p1 e ,…,RG1 p1 s1 ,…,RG1 p1 e(m(e)-1) ),RG1 p1 es1 Is C 1 a S1 th human body position G es1 Mapping to a location in a world coordinate system; the value of e is 1 to t (p 1), the value of S1 is 1 to m (e) -1, the value of p1 is 1 to H1, and H1 is the number of the action track sequences obtained in S250; executing S264;
s263, obtain the second position information set RG2 p1 e =(RG2 p1 e ,RG2 p1 e ,…,RG2 p1 s2 ,…,RG2 p1 em(e) ),RG1 p1 es2 Is C a S2 th human body position G in (1) es2 Mapping to a location in a world coordinate system; s2 takes on a value of 1 to m (e); executing S265;
s264, based on any first position information set RG1 p1 e Obtaining a first distance set D1 p1 e =(D1 p1 e1 ,D1 p1 e2 ,…,D1 p1 et1 ,…,D1 p1 e(m(e)-2) ) Which isIn (D1) p1 et1 For dividing BID in image e q Corresponding position and BID of t1 th human body except the human body in the world coordinate system q The distance between the positions of the corresponding human body in the world coordinate system; t1 is 1 to m (r) -2; executing S266;
s265, based on any second position information set RG2 p1 e Obtaining a second distance set D2 p1 e =(D2 p1 e1 ,D2 p1 e2 ,…,D2 p1 et2 ,…,D2 p1 e(m(e)-1) ) Wherein, D2 p1 et2 Removing BID from image e q Corresponding position and BID of t2 th person except human body in world coordinate system q The distance between the positions of the corresponding human body in the world coordinate system; t2 takes the value from 1 to m (r) -1; executing S267;
s266, go through D1 p1 e If D1 is p1 et1 If D is less than or equal to D, then D1 is added p1 et1 Storing the corresponding identification information of the human body into a second set;
s267, traverse D2 p1 e If D2 is p1 et2 If D is less than or equal to D, adding D2 p1 et2 The corresponding identification information of the human body is stored into the second set.
7. The method according to claim 1, characterized in that the image of each human body recognized in image i is identified by a detection box; wherein, G ij The midpoint position of the bottom side of the detection frame for the jth pedestrian identified in the image i.
8. A non-transitory computer readable storage medium having stored therein at least one instruction or at least one program, the at least one instruction or the at least one program being loaded and executed by a processor to implement the method of any one of claims 1-7.
9. An electronic device comprising a processor and the non-transitory computer readable storage medium of claim 8.
CN202210987107.4A 2022-08-17 2022-08-17 Video-based automatic determination method for joint seal person, electronic device and storage medium Pending CN115410121A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210987107.4A CN115410121A (en) 2022-08-17 2022-08-17 Video-based automatic determination method for joint seal person, electronic device and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210987107.4A CN115410121A (en) 2022-08-17 2022-08-17 Video-based automatic determination method for joint seal person, electronic device and storage medium

Publications (1)

Publication Number Publication Date
CN115410121A true CN115410121A (en) 2022-11-29

Family

ID=84159628

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210987107.4A Pending CN115410121A (en) 2022-08-17 2022-08-17 Video-based automatic determination method for joint seal person, electronic device and storage medium

Country Status (1)

Country Link
CN (1) CN115410121A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115966313A (en) * 2023-03-09 2023-04-14 创意信息技术股份有限公司 Integrated management platform based on face recognition

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115966313A (en) * 2023-03-09 2023-04-14 创意信息技术股份有限公司 Integrated management platform based on face recognition
CN115966313B (en) * 2023-03-09 2023-06-09 创意信息技术股份有限公司 Integrated management platform based on face recognition

Similar Documents

Publication Publication Date Title
US11113526B2 (en) Training methods for deep networks
JP6614611B2 (en) Apparatus, program, and method for tracking object in consideration of similarity between images
US11205276B2 (en) Object tracking method, object tracking device, electronic device and storage medium
US20180101732A1 (en) Image processing apparatus, image processing system, method for image processing, and computer program
CN103679742B (en) Method for tracing object and device
CN111512317A (en) Multi-target real-time tracking method and device and electronic equipment
US9489582B2 (en) Video anomaly detection based upon a sparsity model
CN111104925B (en) Image processing method, image processing apparatus, storage medium, and electronic device
JP6280020B2 (en) Moving object tracking device
Bashar et al. Multiple object tracking in recent times: A literature review
CN110675426B (en) Human body tracking method, device, equipment and storage medium
CN104376575A (en) Pedestrian counting method and device based on monitoring of multiple cameras
US11354923B2 (en) Human body recognition method and apparatus, and storage medium
Han et al. Deep learning-based multi-cattle tracking in crowded livestock farming using video
CN115410121A (en) Video-based automatic determination method for joint seal person, electronic device and storage medium
JP7488674B2 (en) OBJECT RECOGNITION DEVICE, OBJECT RECOGNITION METHOD, AND OBJECT RECOGNITION PROGRAM
CN112381774A (en) Cow body condition scoring method and system based on multi-angle depth information fusion
CN113743293B (en) Fall behavior detection method and device, electronic equipment and storage medium
CN115760922A (en) Target tracking method, device, system, terminal equipment and storage medium
Chung et al. Low-complexity and reliable moving objects detection and tracking for aerial video surveillance with small uavs
CN111862153B (en) Long-time multi-target tracking method for pedestrians
Huang et al. Motion characteristics estimation of animals in video surveillance
CN111898471A (en) Pedestrian tracking method and device
CN111191524A (en) Sports people counting method
CN114092516B (en) Multi-target tracking detection method, device, equipment and medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination