WO2021114985A1 - 一种同行对象识别方法、装置、服务器及系统 - Google Patents
一种同行对象识别方法、装置、服务器及系统 Download PDFInfo
- Publication number
- WO2021114985A1 WO2021114985A1 PCT/CN2020/127500 CN2020127500W WO2021114985A1 WO 2021114985 A1 WO2021114985 A1 WO 2021114985A1 CN 2020127500 W CN2020127500 W CN 2020127500W WO 2021114985 A1 WO2021114985 A1 WO 2021114985A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- objects
- peer
- capture
- time interval
- time
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 80
- 238000004458 analytical method Methods 0.000 claims abstract description 32
- 238000012806 monitoring device Methods 0.000 claims description 104
- 238000012544 monitoring process Methods 0.000 claims description 46
- 238000009434 installation Methods 0.000 claims description 20
- 238000004590 computer program Methods 0.000 claims description 15
- 230000000717 retained effect Effects 0.000 claims description 8
- 238000010586 diagram Methods 0.000 description 12
- 238000004364 calculation method Methods 0.000 description 3
- 238000013500 data storage Methods 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- 230000006399 behavior Effects 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000001914 filtration Methods 0.000 description 2
- 230000037237 body shape Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 230000001815 facial effect Effects 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000000284 resting effect Effects 0.000 description 1
- 230000003442 weekly effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/52—Surveillance or monitoring of activities, e.g. for recognising suspicious objects
- G06V20/53—Recognition of crowd images, e.g. recognition of crowd congestion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
Definitions
- This application relates to the field of security monitoring technology, in particular to a peer object recognition method, device, server and system.
- Peer object recognition is a key technology in the security field. It can identify whether two objects have a peer relationship. On the one hand, it can clarify the social relationship between the objects, discover potential dangers and prevent them in time. On the other hand, in danger When it happens, the object information and the time and location of the danger can be obtained in time, providing important clues for the traceability of the incident.
- the specific object in each image is first identified, and then it is judged whether other objects appear continuously in the same image with the specific object. If an object appears continuously in the same image as the specific object, it is determined
- the relationship between the object and the specific object is a peer object relationship.
- the purpose of the embodiments of the present application is to provide a peer object recognition method, device, server, and system to meet the increasingly complex object analysis requirements.
- the specific technical solutions are as follows:
- an embodiment of the present application provides a peer object recognition method, which includes:
- a peer object between the at least two objects is generated once recording.
- an embodiment of the present application provides a peer object recognition method, which includes:
- the database stores the captured location in the object image collected by the monitoring equipment Records of peer objects between at least two objects whose interval is within a preset range and the time interval of the capture time is within the first preset time interval;
- an embodiment of the present application provides a peer object recognition device, which includes:
- the acquisition module is used to acquire the object image collected by the monitoring equipment
- the analysis module is used to analyze the object image and determine the time and location of each object captured in the object image
- the recognition and recording module is configured to generate the at least two objects once if the distance between the capture locations of at least two objects is within a preset range and the time interval of the capture time of the at least two objects is within the first preset time interval. Peer object records between objects.
- an embodiment of the present application provides a peer object recognition device, which includes:
- the obtaining module is used to obtain the peer object query request, where the peer object query request includes the target object to be queried;
- the search module is used to find all the peer object records where the target object appears in the database, and count the number of occurrences of each object that is the same object as the target object in all the peer object records of the target object.
- the database stores the data collected by the monitoring equipment A record of a peer object between at least two objects in the object image where the distance between the captured locations is within a preset range and the time interval of the capture time is within the first preset time interval;
- the output module is used to output the peer records of the objects whose occurrence times are greater than the second preset threshold among all the peer object records of the target object.
- an embodiment of the present application provides a server, including a processor and a memory, where:
- Memory used to store computer programs
- the processor is configured to implement the method provided in the first aspect or the method provided in the second aspect of the embodiments of the present application when executing the computer program stored in the memory.
- the embodiments of the present application provide a non-temporary storage medium.
- the non-temporary storage medium stores a computer program.
- the computer program is executed by a processor, the method provided in the first aspect of the embodiments of the present application or The method provided in the second aspect.
- an embodiment of the present application provides an application program for execution at runtime: the method provided in the first aspect of the embodiment of the present application or the method provided in the second aspect.
- an embodiment of the present application provides a peer object recognition system, which includes multiple monitoring devices and servers;
- the server is used to obtain the object image collected by the monitoring equipment; analyze the object image to determine the time and location of each object captured in the object image; if the distance between the capture locations of at least two objects is within the preset range and When the time interval of the capture time of the at least two objects is within the first preset time interval, then a peer object record between the at least two objects is generated once.
- the method, device, server and system for peer object recognition acquire object images collected by monitoring equipment, analyze the object images, and determine the time and location of each object captured in the object image, if at least The distance between the capture locations of the two objects is within a preset range, and the time interval of the capture time of the at least two objects is within the first preset time interval, then a peer object record between the at least two objects is generated once.
- the capture time and location of each object captured in the object image can be obtained, and it is finally determined that the distance between the capture locations is within the preset range and the time interval of the capture time is within the first preset time interval.
- At least two objects are peer objects, a peer object record is generated once, and the peer object records between the objects are recorded locally to provide query basis for query personnel.
- the peer object records of each object in the local record can be output instead of only output Peer object records of specific objects can meet the increasingly complex object analysis needs.
- FIG. 1 is a schematic flowchart of a peer object recognition method according to an embodiment of this application
- FIG. 2 is a schematic flowchart of a peer object recognition method according to another embodiment of this application.
- FIG. 3 is a schematic flowchart of a peer object recognition method according to another embodiment of this application.
- FIG. 4 is a schematic flowchart of a peer object recognition method according to still another embodiment of this application.
- FIG. 5 is a schematic flowchart of a peer object recognition method according to another embodiment of this application.
- FIG. 6 is a schematic diagram of a single-screen snapshot of a peer-to-peer scene in an embodiment of the application
- FIG. 7 is a schematic diagram of a scene where a monitoring device collects multiple pictures to capture a peer-to-peer snapshot in an embodiment of the application;
- FIG. 8 is a schematic diagram of a multi-screen capture peer scene collected by multiple monitoring devices according to an embodiment of the application.
- FIG. 9 is a schematic diagram of a peer-captured peer-to-peer scene for filtering repeated peers according to an embodiment of the application.
- FIG. 10 is a schematic diagram of a scene of the principle of peer object matching in four scenarios in an embodiment of the application.
- FIG. 11 is a schematic flowchart of a peer object recognition method according to still another embodiment of this application.
- FIG. 12 is a schematic structural diagram of a peer object recognition device according to an embodiment of the application.
- FIG. 13 is a schematic structural diagram of a peer object recognition device according to another embodiment of the application.
- FIG. 14 is a schematic structural diagram of a server according to an embodiment of the application.
- FIG. 15 is a schematic structural diagram of a peer object recognition system according to an embodiment of the application.
- the peer object information of the specific object In some scenarios, it is necessary to monitor the behavior of some specific objects. Therefore, it is generally necessary to know the peer object information of the specific object.
- the corresponding peer object recognition method after acquiring multiple object images collected by the monitoring equipment, first identify the specific object in each object image, and then judge whether other objects appear in the same object image consecutively with the specific object , If an object continuously appears in the same object image with a specific object, it can be determined that the object and the specific object are in the same object relationship. In this way, when there is a query request, the user enters the object information of a specific object, and can learn about other objects that have a peer-object relationship with the specific object, providing an analysis basis for monitoring the behavior of the specific object.
- embodiments of the present application provide a peer object recognition method, device, server, and system.
- the method for identifying peer objects provided by the embodiments of the present application will be introduced first.
- the execution subject of the peer object identification method provided by the embodiment of the application may be a server with core processing capabilities, and the method for implementing the peer object identification method provided by the embodiment of the application may be software, hardware circuits, and circuits provided in the execution subject. At least one of logic circuits.
- the method for identifying peer objects may include the following steps:
- S101 Acquire an object image collected by a monitoring device.
- one or more monitoring devices are usually deployed for real-time monitoring.
- object images can be collected.
- the object image refers to an image including an object; the object here is a target that needs to be monitored, for example, a vehicle, an animal, or a pedestrian, etc.
- the monitoring device may be a camera, a camera, and other capture devices.
- the monitoring device once an object passes through the monitoring device, the monitoring device automatically recognizes the target, captures the object, and uploads the captured image of the object to a server for peer object recognition.
- Monitoring equipment can be set up in areas where objects flow relatively large. When setting up monitoring equipment, conditions such as angle, height, and brightness need to be guaranteed to meet the requirements for capturing objects.
- S102 Analyze the object image, and determine the time and location of each object captured in the object image.
- the server used for peer object recognition provides intelligent analysis services. After obtaining the object image, the object image can be analyzed. The analysis process is to extract the capture time and location of each object in the object image. Object features are distinguished. Object features refer to the attribute characteristics of the object, for example, one or more of facial features, clothes color, height, or body shape.
- the capture time refers to the time to capture the target, which can usually be read from the image attributes.
- the capture address refers to the location of the target when the target is captured. It can usually be represented by the latitude and longitude information of the monitoring device or the erection position information. Capture the address.
- the latitude and longitude information collected by the positioning module can be used to represent the capture address; for the monitoring equipment without the positioning module, you can The location information of the monitoring equipment is used to represent the capture address.
- the capture address can also be the actual latitude and longitude of the object.
- the coordinates of the object in the image can be obtained, and then the coordinates of the object in the world coordinate system can be obtained according to the pre-established correspondence between the image coordinates and the world coordinates. In order to get the actual longitude and latitude of the target.
- the method for establishing the corresponding relationship between the image coordinates and the world coordinates can refer to the calibration method of camera external parameters in the related art, which will not be repeated here.
- the capture time and location of an object can also form a piece of capture data, and multiple pieces of capture data can be obtained by analyzing an object image or multiple object images.
- the peer object record between the at least two objects indicates that the at least two objects are peer objects.
- the object record of the at least two objects in the same line may include the identity of the two objects, such as ID, etc., and may also include the captured images or image features of the two objects, where the image feature of the object may be depth Learn the characteristics of objects extracted by the network.
- the embodiment of this application provides a rule for judging peer objects, that is, different objects pass through a certain area one after another (a few seconds interval, different pictures are captured) or at the same time (the same picture is captured) (the distance between the captured locations is within a preset range), then Treat it as a peer object, and generate a peer object record between the objects once.
- the distance between the capture locations of at least two objects is within a preset range (for example, the same building, four directions of an intersection, the same square, etc.), and the capture time of the at least two objects If the interval is within the first preset time interval (ie, several seconds or at the same time), the at least two objects are identified as peer objects, and a peer object record between the at least two objects is generated once.
- a preset range for example, the same building, four directions of an intersection, the same square, etc.
- the peer object records of each object in the local record can be output instead of only output
- the peer object record of a specific object the obtained peer object record is more comprehensive, which can meet the increasingly complex object analysis needs.
- every time at least two objects are identified as peer objects a peer object record between the at least two objects will be generated once. There will be many peer object records for different objects. In this way, the peer object record is output The output is based on the number of generated peer object records, which provides support for the accuracy of output peer object records.
- S103 can be specifically implemented through the following steps:
- filter conditions are added: the same at least two objects are recognized as peer objects multiple times in a short period of time, and only one record of peer objects in the second preset time interval is retained. Delete other peer object records of the at least two objects in the second preset time interval. That is, in this case, no matter how many times it is recognized as a peer, it will only be recorded as one peer record.
- the peer object records of each target target can be obtained.
- the peer object rules two or more objects are continuously The number of times the objects are identified as peers is greater than or equal to the first preset threshold (for example, 20 peers are identified in a day), it can be determined that these objects are peers with each other, and only one record of the peers between these objects is recorded.
- the first preset threshold for example, 20 peers are identified in a day
- the last record of the peers within the second preset time interval can be retained.
- the records of the peers recorded at other times can also be retained, and there is no specific limitation here.
- the amount of data recorded by peer objects that need to be stored is reduced, effectively saving storage resources, and due to the limitation of the judgment rules, there may be a single misjudgment (for example, a misjudgment only due to accidental rubbings) For peers), and after multiple identifications, the objects with multiple peer relationships are recorded as a peer record, which can reduce the misjudgment of peers.
- the step of obtaining the object image collected by the monitoring device may specifically be: obtaining at least one object image collected by one monitoring device; or obtaining multiple object images collected by multiple monitoring devices .
- the steps of analyzing the object image to determine the time and location of each object captured in the object image can be specifically as follows: analyzing an object image collected by a monitoring device to determine the object image captured The capture time and location of each object, where the capture time is the timestamp of the object image collected by the monitoring device, and the capture location is the installation location of the monitoring device; or, multiple object images captured by a monitoring device Analyze and determine the capture time and location of each object captured in each object image, where the capture time is the time stamp of each object image captured by the monitoring device, and the capture location is the installation location of the monitoring device; or, for many Multiple object images collected by a monitoring device are analyzed to determine the capture time and location of each object captured in each object image.
- the capture time is the time stamp of each object image captured by each monitoring device, and the capture location is the capture location.
- the recording step can specifically be: for at least two objects in an object image collected by a monitoring device, generating a peer object record between the at least two objects; or, if a monitoring device collects multiple objects If the time interval between the capture times of at least two objects in the image is within the first preset time interval, then a peer object record between the at least two objects is generated once; or, if multiple object images captured by multiple monitoring devices If the distance between the capture locations of at least two objects is within a preset range, and the time interval of the capture time of the at least two objects is within the first preset time interval, then a peer object record between the at least two objects is generated once .
- One monitoring device can monitor a specified area, and multiple monitoring devices can also monitor the specified area at the same time.
- the installation location of the monitoring device (such as the installed latitude and longitude) can be used to determine whether multiple monitoring devices are monitoring. Specify the area range. Because of the needs of the scene, multiple monitoring devices may be arranged at the same point. These monitoring devices have complementary capture capabilities, or the installation locations of multiple monitoring devices have different geographical latitudes and longitudes, but collaborative work can make up for the blind spots of monitoring.
- the first scene to capture the peer object is a single-screen capture of the peer object scene.
- a monitoring device monitors a monitoring area, analyzes the object image collected by the monitoring device at a certain moment, and uses the time stamp of the object image collected by the monitoring device as Capture time and the installation location of the monitoring device as the capture location.
- the capture time i.e. time stamp
- capture location i.e. installation location
- the second scene of capturing peer objects is a multi-screen capturing of peer objects of the same monitoring device.
- a monitoring device monitors a monitoring area, and multiple object images collected by the monitoring device in a certain period of time are analyzed, and the monitoring device Collect the time stamp of each object image as the capture time of each object in each object image, and use the installation location of the monitoring device as the capture location of each object in each object image, then when at least two objects are captured in each object image , Because there are the same capture location (i.e. installation location), it is only necessary to determine whether the time interval of the capture time (i.e. time stamp) of at least two objects is within the first preset time interval. If so, the at least two objects are generated once. Peer object records between objects.
- the third scene of capturing peer objects is the scene of capturing peer objects in multiple images of different monitoring devices.
- the embodiment of the present application also provides a peer object recognition method. As shown in FIG. 2, the method may include the following steps.
- S201 Obtain an object image collected by a monitoring device.
- S202 Analyze the object image, and determine the time and location of each object captured in the object image.
- S201 and S202 in the embodiment shown in FIG. 2 are the same as S101 and S102 in the embodiment shown in FIG. 1, and will not be repeated here.
- S203 Identify the object characteristics of each object, and cluster each object based on the object characteristics of each object.
- the server used for peer object recognition also provides an object comparison service for object feature comparison.
- the comparison process can be a comparison between the object features of each object, or the acquired object and the known object.
- the object feature model is compared.
- the purpose of the comparison is to cluster the objects, and the objects of the same type are grouped together. After clustering, different types of objects can be distinguished by marking the identity tags of each object.
- the specific method for clustering objects may be: comparing the object features of each object to obtain the object feature similarity between the objects; the object feature similarity is greater than or Objects that are equal to the preset similarity threshold are labeled with the same identity label; objects whose feature similarity is less than the preset similarity threshold are labeled with different identity labels.
- the method of comparing the object features may be to compare the object features of each object to obtain the similarity of the object features between the objects, for example, to compare the clothes color and hair of the objects respectively.
- Object features such as length, height, etc., compare the object feature similarity between the two objects. The higher the similarity, the greater the probability that the two objects are the same object. Therefore, in the embodiments of the present application, A preset similarity threshold is set. If the object feature similarity is greater than or equal to the preset similarity threshold, the two objects can basically be regarded as the same object and marked with the same identity tag. If the object feature similarity is less than the preset similarity If the threshold is higher, the two objects can basically be identified as objects of different rows and marked with different identity tags.
- object targets can be clustered.
- Objects with similarity of object features greater than or equal to the preset similarity threshold are labeled with the same identity tags, and objects of the same category are classified into the same category.
- the comparison process of object features can also be to compare the object features of each object with a known object feature model.
- the object feature model is a model of a known object established based on experience. Through comparison, it can be more Know the identity of the subject clearly.
- objects belonging to the same class are regarded as the same object. If the distance between the capture locations of at least two objects of different types is within the preset range, and the time interval of the capture time of the at least two objects of different types is within the first preset time interval, then the at least two objects are generated once. Peer object records between objects of different classes. After clustering the objects, the objects are divided more accurately. Objects belonging to the same category will not be recorded as a peer object record, but objects belonging to different categories will be recorded as a peer object record.
- the embodiment of the present application also provides a peer object recognition method. As shown in FIG. 3, the method may include the following steps.
- S301 Obtain an object image collected by a monitoring device.
- S302 Analyze the object image, and determine the time and location of each object captured in the object image.
- S301 and S302 in the embodiment shown in FIG. 3 are the same as S101 and S102 in the embodiment shown in FIG. 1, and will not be repeated here.
- S303 Store the capture time and capture location of each object in the first database.
- the capture time and location of each object can be stored in the first database to achieve the purpose of big data storage and provide a big data basis for subsequent peer object judgment.
- S304 Extract the capture time and capture location of at least two objects from the first database.
- S305 If the distance between the capture locations of the at least two objects is within a preset range, and the time interval of the capture time of the at least two objects is within the first preset time interval, generate a time between the at least two objects Of peer object records.
- the first database and the second database may be the same database or different databases, and both are within the protection scope of this application. In one embodiment, the first database and the second database are different databases.
- the first database is used to store the captured images of the object, the characteristics of the object, the time and location of the capture of the object; the second database is used to record peer objects Record, so as to facilitate the classification and management of data.
- the embodiment of the present application also provides a peer object recognition method. As shown in FIG. 4, the method may include the following steps.
- S401 Obtain an object image collected by a monitoring device.
- S402 Analyze the object image, and determine the time and location of each object captured in the object image.
- S401 to S403 in the embodiment shown in FIG. 4 are the same as S101 to S103 in the embodiment shown in FIG. 1, and will not be repeated here.
- S404 Acquire a peer object query request, where the peer object query request includes the object information of the target object to be queried.
- a peer object query request is sent to the server through the query client (usually a computer).
- the peer object query request includes the object information of the target object to be queried, for example, the target object
- the object information of can include one or more of the target object's name, ID, characteristics and other information.
- S405 According to the object information of the target object, search for all the peer object records where the target object appears, and count the appearance times of each object that is the same as the target object in all the peer object records of the target object.
- the server After the server receives the peer object query request, based on the target object's object information, it searches the recorded peer object records for all the peer object records where the target object appears. These peer object records include other objects that have a peer object relationship with the target object. Count the number of appearances of other objects, and get the number of appearances of each object that is the same object as the target object.
- S406 Output the peer object records of the object whose appearance times are greater than the second preset threshold among all the peer object records of the target object.
- Outputting the peer object records with the number of occurrences greater than the second preset threshold that is, selecting and outputting the peer object records of the object with the number of occurrences greater than the second preset threshold among all the peer object records of the target object.
- the number of occurrences of the object is greater than the second preset threshold, which means that the number of times the object and the target object are in the same line is greater than the second preset threshold. Since the inquirer generally pays attention to the objects that often have a peer relationship with the target object, and there is a close relationship between these objects and the target object, therefore, when giving feedback to the inquirer, the number of peers with the target object can be reported to be greater than the number of peers. 2.
- the peer record of the subject with a preset threshold where the second preset threshold can be set based on experience, and each peer record of the output target object contains objects whose number of peers with the target object is greater than the second preset threshold and target.
- the target object's peer object record is searched out from the recorded peer object records, and based on the number of times each object has walked with the target object, the output of the peer object record with the target object is output.
- the peer record of the subject whose peer count is greater than the second preset threshold. According to the real-time needs of the query personnel, the accurate query results of the peer object records can be output to the query personnel in a targeted manner, which meets the query requirements of the query personnel.
- FIG. 5 is a schematic flowchart of a peer object recognition method provided by an embodiment of the application.
- the monitoring equipment collects the object image, it reports the object image to the server that provides the central service.
- the intelligent analysis service in the server analyzes the object image and determines the time and location of each object captured in the object image. Call the captured data (the captured time and location of each captured object) through the object application, and call out the object characteristics.
- the comparison service will compare the object characteristics to determine the same object to facilitate the assignment of identities Identify and return the comparison result to the target application, send the snapshot data and comparison result to the data warehouse (the first database) for storage, and extract historical snapshot data from the data warehouse when performing peer object recognition.
- the specific calculation process is as described in the above method embodiment, which will not be repeated here, and the calculation result is finally output to the database (second database) for storage.
- FIG. 6 it is a schematic diagram of a single-screen capture of a peer-to-peer object scene.
- the monitored device 1 captures the image.
- the obtained object image is uploaded to the server, and the intelligent analysis service analyzes the object image to obtain object feature 1 and object feature 2, and compare the two object features through the comparison service to determine the same object.
- the object binding identity ID 1 represented by the object feature 1 is the binding identity ID 2 of the object represented by the object feature 2.
- the position of object 1 (the object bound by ID 1) in each image and the position of object 2 (the object bound by ID 2) in each image can be obtained, so that a snapshot of object 1 at time t1 can be obtained Record (including the capture time and location) and the capture record of object 2 at t1, and save the capture record to the data warehouse.
- peer objects different objects sequentially (several seconds interval, different screen captures) or at the same time (same screen (Snapshot) After passing through a certain area, it is regarded as a peer object, and it can be determined that the object 1 and the object 2 are peer objects, and a peer object record is generated.
- FIG 7 it is a schematic diagram of a scene where a monitoring device collects multiple pictures to capture peer objects.
- the monitored device 1 captures the image, and there are objects 1 and Object 2, the two captured object images are uploaded to the server, and the intelligent analysis service analyzes the object image to obtain object feature 1 and object feature 2, and compare the two object features through the comparison service to determine the same object.
- the object binding identity ID 1 represented by the object feature 1 is the binding identity ID 2 of the object represented by the object feature 2.
- the position of object 1 (the object bound by ID 1) in the image and the position of object 2 (the object bound by ID 2) in the image can be obtained, so that the snapshot record of object 1 at time t1 and The snapshot record of object 2 at time t2, save the snapshot record to the data warehouse, according to the peer object rule: different objects pass through a certain area in succession (a few seconds interval, different screen captures) or at the same time (same screen capture).
- the peer object since the difference between t1 and t2 is less than the first preset time interval, it can be determined that the object 1 and the object 2 are peer objects, and a peer object record is generated.
- FIG 8 it is a schematic diagram of a scene where multiple monitoring devices collect multiple images to capture peer objects.
- object 1 and object 2 are captured by monitoring device 1 and monitoring device 2 in the same monitoring area 1 at time t1 and t2, respectively.
- the two captured object images are uploaded to the server, and the intelligent analysis service analyzes the object image to obtain object feature 1 and object feature 2, and compare the two object features through the comparison service to determine the same object.
- the object binding identity ID 1 represented by the object feature 1 is the binding identity ID 2 of the object represented by the object feature 2.
- the position of object 1 (the object bound by ID 1) in the image and the position of object 2 (the object bound by ID 2) in the image can be obtained, so that the snapshot record of object 1 at time t1 and The snapshot record of object 2 at time t2, save the snapshot record to the data warehouse, according to the peer object rule: different objects pass through a certain area in succession (a few seconds interval, different screen captures) or at the same time (same screen capture).
- the peer object since the difference between t1 and t2 is less than the first preset time interval, it can be determined that the object 1 and the object 2 are peer objects, and a peer object record is generated.
- the following takes the same camera capture as an example to describe in detail the implementation of filtering repeated peer object records to increase the effective peer rate.
- the monitored device 1 captures the image, and there are object 1 and object 2 in the screen.
- the captured object image is uploaded to the server, and the object image is analyzed by the intelligent analysis service to obtain object feature 1 and object feature 2, and the two object features are performed through the comparison service
- the snapshot record of object 1 at time t1 and the snapshot record of object 2 at time t1 are obtained, and the snapshot records are saved in the data warehouse.
- both the time interval between t2 and t1 and the time interval between t3 and t1 are less than the second preset time interval, it is considered that the peer recognition of the same group of objects has occurred in a short period of time, and it is only retained once and recorded as a peer record .
- the peer object records of each object can be obtained, and the rules are calculated according to the peer objects (During a period of time (can be daily, weekly or monthly) two or more peers can be regarded as peers with each other multiple times.) Filter the records of peers to reach the threshold of the number of peers (such as five records a day). The object, as a peer object, gives feedback to the inquirer.
- the execution subject is a system composed of a server and a database. As shown in FIG. 11, the method may include the following steps.
- S1101 Acquire a peer object query request, where the peer object query request includes the target object to be queried.
- the target object to be queried refers to the object that the user wants to query.
- the identity is the object of Zhang San, or the identity is 111111, and so on.
- the peer object query request includes the target object to be queried.
- the target peer object query request includes the identity of the target object to be queried.
- S1102 Search for all peer object records where the target object appears in the database, and count the number of occurrences of each object that is the same object as the target object in all peer object records of the target object, where the database stores the object images collected by the monitoring device The distance between the captured locations is within a preset range and the time interval of the capture time is within the first preset time interval between at least two objects in the same line.
- S1103 Output the peer records of the object whose appearance times are greater than the preset second threshold among all the peer object records of the target object.
- the peer object record where the target object appears is specifically searched from the database, and based on the number of occurrences of each object that is the peer object with the target object, all peers of the target object are output
- the database provides big data storage. In the peer object query scenario, it provides a big data query basis to ensure the integrity of the data and provide a guarantee for the accuracy of the query results.
- an embodiment of the present application provides a peer object recognition device.
- the device may include:
- the obtaining module 1210 is used to obtain the object image collected by the monitoring device
- the analysis module 1220 is used to analyze the object image to obtain the time and location of each object captured in the object image;
- the recognition recording module 1230 is configured to generate the at least two objects once if the distance between the capture locations of at least two objects is within a preset range and the time interval of the capture time of the at least two objects is within the first preset time interval. Peer object records between objects.
- the device may also include:
- the clustering module is used to identify the object characteristics of each object; based on the object characteristics of each object, cluster each object;
- the identification record module 1230 can be specifically used for:
- the distance between the capture locations of the at least two objects is within a preset range and the time interval of the capture time of the at least two objects is within the first preset time interval, then generate A record of the peer objects between at least two objects at a time.
- the obtaining module 1210 can be specifically used for:
- the analysis module 1220 can be specifically used for:
- the capture time is the time stamp of each object image collected by the monitoring device, and the capture location is The installation location of the monitoring equipment;
- the capture time is the time stamp and location of each object image captured by each monitoring device The installation position of each monitoring equipment for collecting each object image;
- the identification record module 1230 can be specifically used for:
- the time interval of the capturing time of at least two objects in the multiple object images collected by one monitoring device is within the first preset time interval, then generate a peer object record between the at least two objects once;
- the distance between the capture locations of at least two objects in the multiple object images collected by multiple monitoring devices is within a preset range, and the time interval of the capture time of the at least two objects is within the first preset time interval, then generate A record of the peer objects between at least two objects at a time.
- the device may also include:
- the storage module is used to store the capture time and location of each object in the first database
- the identification record module 1230 can be specifically used for:
- the identification record module 1230 can be specifically used for:
- the obtaining module 1210 can also be used to:
- the peer object query request includes the object information of the target object to be queried
- the device may also include:
- the search module is used to find all the records of the peers where the target object appears according to the target information of the target, and count the number of occurrences of each object that is the same as the target object in all the records of the target object's peers;
- the output module is used to output the peer object records of the object whose appearance times are greater than the second preset threshold among all the peer object records of the target object.
- the peer object records of each object in the local record can be output instead of only output
- the peer object record of a specific object the obtained peer object record is more comprehensive, which can meet the increasingly complex object analysis needs.
- every time at least two objects are identified as peer objects a peer object record between the at least two objects will be generated once. There will be many peer object records for different objects. In this way, the peer object record is output The output is based on the number of generated peer object records, which provides support for the accuracy of output peer object records.
- the embodiment of the present application also provides a peer object recognition device. As shown in FIG. 13, the device may include:
- the obtaining module 1310 is configured to obtain a peer object query request, where the peer object query request includes the target object to be queried;
- the search module 1320 is used to search for all peer object records where the target object appears in the database, and count the number of occurrences of each object that is the same object as the target object in all the peer object records of the target object.
- the database stores the collection of monitoring equipment Records of peer objects between at least two objects in the object image where the distance between the capture locations is within a preset range and the time interval of the capture time is within the first preset time interval;
- the output module 1330 is configured to output the peer records of the objects whose appearance times are greater than the second preset threshold among all the peer object records of the target object.
- the peer object record in which the target object appears is searched in a targeted manner from the database, and based on the number of times each object travels with the target object, the output occurrence number is greater than the second preset threshold
- the peer object record of the object is greater than the second preset threshold
- the database provides big data storage. In the peer object query scenario, it provides the basis for big data query to ensure the integrity of the data and provide a guarantee for the accuracy of the query results.
- An embodiment of the present application also provides a server, as shown in FIG. 14, including a processor 1401 and a memory 1402, where:
- the memory 1402 is used to store computer programs
- the processor 1401 is configured to execute any of the peer object recognition methods provided in the embodiment of the present application when executing the computer program stored in the memory 1402.
- the foregoing memory may include RAM (Random Access Memory, random access memory), and may also include NVM (Non-Volatile Memory, non-volatile memory), such as at least one disk storage.
- NVM Non-Volatile Memory, non-volatile memory
- the memory may also be at least one storage device located far away from the foregoing processor.
- the above-mentioned processor may be a general-purpose processor, including CPU (Central Processing Unit), NP (Network Processor, network processor), etc.; it may also be DSP (Digital Signal Processing, digital signal processor), ASIC (Application Specific Integrated Circuit, FPGA (Field-Programmable Gate Array, Field Programmable Gate Array) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components.
- CPU Central Processing Unit
- NP Network Processor, network processor
- DSP Digital Signal Processing, digital signal processor
- ASIC Application Specific Integrated Circuit
- FPGA Field-Programmable Gate Array, Field Programmable Gate Array
- other programmable logic devices discrete gates or transistor logic devices, discrete hardware components.
- the processor reads the computer program stored in the memory and runs the computer program to achieve: obtain the object image collected by the monitoring equipment, analyze the object image, and determine the captured object in the object image Capture time and capture location. If the distance between the capture locations of at least two objects is within a preset range, and the time interval of the capture time of the at least two objects is within the first preset time interval, then the at least two objects are generated once Peer object records between objects. Through the analysis of the object image, the capture time and location of each object captured in the object image can be obtained, and it is finally determined that the distance between the capture locations is within the preset range and the time interval of the capture time is within the first preset time interval.
- At least two objects are peer objects, a peer object record is generated once, and the peer object records between the objects are recorded locally to provide query basis for query personnel.
- the peer object records of each object in the local record can be output instead of only output
- the peer object record of a specific object the obtained peer object record is more comprehensive, which can meet the increasingly complex object analysis needs.
- every time at least two objects are identified as peer objects a peer object record between the at least two objects will be generated once. There will be many peer object records for different objects. In this way, the peer object record is output The output is based on the number of generated peer object records, which provides support for the accuracy of output peer object records.
- the embodiment of the present application provides a non-temporary storage medium that stores a computer program in the non-temporary storage medium.
- the computer program is executed by a processor, any of the peer object identification methods provided in the embodiments of the present application is implemented.
- the non-temporary storage medium stores a computer program that executes the peer object recognition method provided by the embodiment of this application at runtime, so it can achieve: obtain the object image collected by the monitoring device, analyze the object image, and determine The capture time and location of each object captured in the object image, if the distance between the capture locations of at least two objects is within the preset range and the time interval of the capture time of the at least two objects is within the first preset time interval , Then a peer object record between the at least two objects is generated once. Through the analysis of the object image, the capture time and location of each object captured in the object image can be obtained, and it is finally determined that the distance between the capture locations is within the preset range and the time interval of the capture time is within the first preset time interval.
- At least two objects are peer objects, a peer object record is generated once, and the peer object records between the objects are recorded locally to provide query basis for query personnel.
- the peer object records of each object in the local record can be output instead of only output
- the peer object record of a specific object the obtained peer object record is more comprehensive, which can meet the increasingly complex object analysis needs.
- every time at least two objects are identified as peer objects a peer object record between the at least two objects will be generated once. There will be many peer object records for different objects. In this way, the peer object record is output The output is based on the number of generated peer object records, which provides support for the accuracy of output peer object records.
- the embodiment of the present application also provides an application program for executing at runtime: any of the peer object recognition methods provided in the embodiment of the present application.
- the embodiment of the present application provides a peer object recognition system.
- the system includes one or more monitoring devices 1510 and a server 1520;
- Monitoring equipment 1510 used to collect object images
- the server 1520 is used to obtain the object image collected by the monitoring equipment; analyze the object image to determine the time and location of each object captured in the object image; if the distance between the capture locations of at least two objects is within a preset range, And the time interval of the capture time of the at least two objects is within the first preset time interval, then a peer object record between the at least two objects is generated once.
- the system may also include a first database and a second database;
- the first database is used to store the time and location of each object captured by the server after analyzing the object image
- the second database is used to store peer object records generated by the server.
- the system also includes a client;
- the server is also used to obtain a peer object query request, where the peer object query request includes the object information of the target object to be queried; according to the object information of the target object, find all the peer object records where the target object appears, and count all the target objects The number of occurrences of each object that is the same as the target object in the peer object record;
- the client terminal is used to display the peer object records of the object whose appearance times are greater than the second preset threshold among all the peer object records of the target object.
- the server obtains the object image collected by the monitoring device, analyzes the object image, and determines the time and location of each object captured in the object image. If the distance between the capture locations of at least two objects is within a preset range And the time interval of the capture time of the at least two objects is within the first preset time interval, then a peer object record between the at least two objects is generated once.
- the capture time and location of each object captured in the object image can be obtained, and it is finally determined that the distance between the capture locations is within the preset range and the time interval of the capture time is within the first preset time interval.
- At least two objects are peer objects, a peer object record is generated once, and the peer object records between the objects are recorded locally to provide query basis for query personnel.
- the peer object records of each object in the local record can be output instead of only output
- the peer object record of a specific object the obtained peer object record is more comprehensive, which can meet the increasingly complex object analysis needs.
- every time at least two objects are identified as peer objects a peer object record between the at least two objects will be generated once. There will be many peer object records for different objects. In this way, the peer object record is output The output is based on the number of generated peer object records, which provides support for the accuracy of output peer object records.
Abstract
Description
Claims (20)
- 一种同行对象识别方法,其特征在于,所述方法包括:获取监控设备采集的对象图像;对所述对象图像进行分析,确定所述对象图像中抓拍到各对象的抓拍时间及抓拍地点;若至少两个对象的抓拍地点的间距在预设范围内、且所述至少两个对象的抓拍时间的时间间隔在第一预设时间间隔内,则生成一次所述至少两个对象之间的同行对象记录。
- 根据权利要求1所述的方法,其特征在于,在所述若至少两个对象的抓拍地点的间距在预设范围内、且所述至少两个对象的抓拍时间的时间间隔在第一预设时间间隔内,则生成一次所述至少两个对象之间的同行对象记录之前,所述方法还包括:识别所述各对象的对象特征;基于所述各对象的对象特征,对所述各对象进行聚类;所述若至少两个对象的抓拍地点的间距在预设范围内、且所述至少两个对象的抓拍时间的时间间隔在第一预设时间间隔内,则生成一次所述至少两个对象之间的同行对象记录,包括:针对属于不同类的至少两个对象,若所述至少两个对象的抓拍地点的间距在预设范围内、且所述至少两个对象的抓拍时间的时间间隔在第一预设时间间隔内,则生成一次所述至少两个对象之间的同行对象记录。
- 根据权利要求1所述的方法,其特征在于,所述获取监控设备采集的对象图像,包括:获取一个监控设备采集的至少一张对象图像;或者,获取多个监控设备采集的多张对象图像;所述对所述对象图像进行分析,确定所述对象图像中抓拍到各对象的抓拍时间及抓拍地点,包括:对一个监控设备采集的一张对象图像进行分析,确定该张对象图像中抓拍到各对象的抓拍时间及抓拍地点,其中,所述抓拍时间为该监控设备采集该张对象图像的时间戳,所述抓拍地点为该监控设备的安装位置;或者,对一个监控设备采集的多张对象图像进行分析,确定各张对象图像中抓拍到各对象的抓拍时间及抓拍地点,其中,所述抓拍时间为该监控设备采集 各张对象图像的时间戳,所述抓拍地点为该监控设备的安装位置;或者,对多个监控设备采集的多张对象图像进行分析,确定各张对象图像中抓拍到各对象的抓拍时间及抓拍地点,其中,所述抓拍时间为各监控设备采集各张对象图像的时间戳,所述抓拍地点为采集各张对象图像的各监控设备的安装位置;所述若至少两个对象的抓拍地点的间距在预设范围内、且所述至少两个对象的抓拍时间的时间间隔在第一预设时间间隔内,则生成一次所述至少两个对象之间的同行对象记录,包括:针对一个监控设备采集的一张对象图像中的至少两个对象,生成一次所述至少两个对象之间的同行对象记录;或者,若一个监控设备采集的多张对象图像中至少两个对象的抓拍时间的时间间隔在第一预设时间间隔内,则生成一次所述至少两个对象之间的同行对象记录;或者,若多个监控设备采集的多张对象图像中至少两个对象的抓拍地点的间距在预设范围内、且所述至少两个对象的抓拍时间的时间间隔在第一预设时间间隔内,则生成一次所述至少两个对象之间的同行对象记录。
- 根据权利要求1所述的方法,其特征在于,在所述对所述对象图像进行分析,确定所述对象图像中抓拍到各对象的抓拍时间及抓拍地点之后,所述方法还包括:将所述各对象的抓拍时间及抓拍地点存储至第一数据库;所述若至少两个对象的抓拍地点的间距在预设范围内、且所述至少两个对象的抓拍时间的时间间隔在第一预设时间间隔内,则生成一次所述至少两个对象之间的同行对象记录,包括:从所述第一数据库中,提取出至少两个对象的抓拍时间及抓拍地点;若所述至少两个对象的抓拍地点的间距在预设范围内、且所述至少两个对象的抓拍时间的时间间隔在第一预设时间间隔内,则生成一次所述至少两个对象之间的同行对象记录;存储所述同行对象记录至第二数据库。
- 根据权利要求1所述的方法,其特征在于,所述若至少两个对象的抓 拍地点的间距在预设范围内、且所述至少两个对象的抓拍时间的时间间隔在第一预设时间间隔内,则生成一次所述至少两个对象之间的同行对象记录,包括:识别抓拍地点的间距在预设范围内、且抓拍时间的时间间隔在第一预设时间间隔内的至少两个对象;统计在第二预设时间间隔内连续生成所述至少两个对象之间的同行对象记录的次数;若所述次数大于或等于第一预设阈值,则保留所述第二预设时间间隔内生成的一次同行对象记录。
- 根据权利要求1-5任一项所述的方法,其特征在于,在所述若至少两个对象的抓拍地点的间距在预设范围内、且所述至少两个对象的抓拍时间的时间间隔在第一预设时间间隔内,则生成一次所述至少两个对象之间的同行对象记录之后,所述方法还包括:获取同行对象查询请求,所述同行对象查询请求包括待查询的目标对象的对象信息;根据所述目标对象的对象信息,查找出现所述目标对象的所有同行对象记录,并统计目标对象的所有同行对象记录中与所述目标对象为同行对象的各对象出现次数;输出所述目标对象的所有同行对象记录中出现次数大于第二预设阈值的对象的同行对象记录。
- 一种同行对象识别方法,其特征在于,所述方法包括:获取同行对象查询请求,所述同行对象查询请求包括待查询的目标对象;从数据库中查找出现所述目标对象的所有同行对象记录,并统计目标对象的所有同行对象记录中与所述目标对象为同行对象的各对象出现次数,所述数据库存储有将监控设备采集的对象图像中抓拍地点的间距在预设范围内、且抓拍时间的时间间隔在第一预设时间间隔内的至少两个对象之间的同行对象记录;输出所述目标对象的所有同行对象记录中出现次数大于第二预设阈值的对象的同行记录。
- 一种同行对象识别装置,其特征在于,所述装置包括:获取模块,用于获取监控设备采集的对象图像;分析模块,用于对所述对象图像进行分析,确定所述对象图像中抓拍到 各对象的抓拍时间及抓拍地点;识别记录模块,用于若至少两个对象的抓拍地点的间距在预设范围内、且所述至少两个对象的抓拍时间的时间间隔在第一预设时间间隔内,则生成一次所述至少两个对象之间的同行对象记录。
- 根据权利要求8所述的装置,其特征在于,所述装置还包括:聚类模块,用于识别所述各对象的对象特征;基于所述各对象的对象特征,对所述各对象进行聚类;所述识别记录模块,具体用于:针对属于不同类的至少两个对象,若所述至少两个对象的抓拍地点的间距在预设范围内、且所述至少两个对象的抓拍时间的时间间隔在第一预设时间间隔内,则生成一次所述至少两个对象之间的同行对象记录。
- 根据权利要求8所述的装置,其特征在于,所述获取模块,具体用于:获取一个监控设备采集的至少一张对象图像;或者,获取多个监控设备采集的多张对象图像;所述分析模块,具体用于:对一个监控设备采集的一张对象图像进行分析,确定该张对象图像中抓拍到各对象的抓拍时间及抓拍地点,其中,所述抓拍时间为该监控设备采集该张对象图像的时间戳,所述抓拍地点为该监控设备的安装位置;或者,对一个监控设备采集的多张对象图像进行分析,确定各张对象图像中抓拍到各对象的抓拍时间及抓拍地点,其中,所述抓拍时间为该监控设备采集各张对象图像的时间戳,所述抓拍地点为该监控设备的安装位置;或者,对多个监控设备采集的多张对象图像进行分析,确定各张对象图像中抓拍到各对象的抓拍时间及抓拍地点,其中,所述抓拍时间为各监控设备采集各张对象图像的时间戳,所述抓拍地点为采集各张对象图像的各监控设备的安装位置;所述识别记录模块,具体用于:针对一个监控设备采集的一张对象图像中的至少两个对象,生成一次所述至少两个对象之间的同行对象记录;或者,若一个监控设备采集的多张对象图像中至少两个对象的抓拍时间的时间间隔在第一预设时间间隔内,则生成一次所述至少两个对象之间的同行对象记录;或者,若多个监控设备采集的多张对象图像中至少两个对象的抓拍地点的间距在预设范围内、且所述至少两个对象的抓拍时间的时间间隔在第一预设时间间隔内,则生成一次所述至少两个对象之间的同行对象记录。
- 根据权利要求8所述的装置,其特征在于,所述装置还包括:存储模块,用于将所述各对象的抓拍时间及抓拍地点存储至第一数据库;所述识别记录模块,具体用于:从所述第一数据库中,提取出至少两个对象的抓拍时间及抓拍地点;若所述至少两个对象的抓拍地点的间距在预设范围内、且所述至少两个对象的抓拍时间的时间间隔在第一预设时间间隔内,则生成一次所述至少两个对象之间的同行对象记录;存储所述同行对象记录至第二数据库。
- 根据权利要求8所述的装置,其特征在于,所述识别记录模块,具体用于:识别抓拍地点的间距在预设范围内、且抓拍时间的时间间隔在第一预设时间间隔内的至少两个对象;统计在第二预设时间间隔内连续生成所述至少两个对象之间的同行对象记录的次数;若所述次数大于或等于第一预设阈值,则保留所述第二预设时间间隔内生成的一次同行对象记录。
- 根据权利要求8-12任一项所述的装置,其特征在于,所述获取模块,还用于:获取同行对象查询请求,所述同行对象查询请求包括待查询的目标对象的对象信息;所述装置还包括:查找模块,用于根据所述目标对象的对象信息,查找出现所述目标对象的所有同行对象记录,并统计目标对象的所有同行对象记录中与所述目标对象为同行对象的各对象出现次数;输出模块,用于输出所述目标对象的所有同行对象记录中出现次数大于 第二预设阈值的对象的同行对象记录。
- 一种同行对象识别装置,其特征在于,所述装置包括:获取模块,用于获取同行对象查询请求,所述同行对象查询请求包括待查询的目标对象;查找模块,用于从数据库中查找出现所述目标对象的所有同行对象记录,并统计目标对象的所有同行对象记录中与所述目标对象为同行对象的各对象出现次数,所述数据库存储有将监控设备采集的对象图像中抓拍地点的间距在预设范围内、且抓拍时间的时间间隔在第一预设时间间隔内的至少两个对象之间的同行对象记录;输出模块,用于输出所述目标对象的所有同行对象记录中出现次数大于第二预设阈值的对象的同行记录。
- 一种服务器,其特征在于,包括处理器和存储器,其中,所述存储器,用于存放计算机程序;所述处理器,用于执行所述存储器上所存放的计算机程序时,实现权利要求1-6或7任一项所述的方法。
- 一种非临时性存储介质,其特征在于,所述非临时性存储介质内存储有计算机程序,所述计算机程序被处理器执行时,实现权利要求1-6或7任一项所述的方法。
- 一种应用程序,其特征在于,用于在运行时执行:权利要求1-6或7任一项所述的方法。
- 一种同行对象识别系统,其特征在于,所述系统包括一个或多个监控设备及服务器;所述监控设备,用于采集对象图像;所述服务器,用于获取所述监控设备采集的所述对象图像;对所述对象图像进行分析,确定所述对象图像中抓拍到各对象的抓拍时间及抓拍地点;若至少两个对象的抓拍地点的间距在预设范围内、且所述至少两个对象的抓拍时间的时间间隔在第一预设时间间隔内,则生成一次所述至少两个对象之间的同行对象记录。
- 根据权利要求18所述的系统,其特征在于,所述系统还包括第一数据库及第二数据库;所述第一数据库,用于存储所述服务器对所述对象图像进行分析得到的抓拍到各对象的抓拍时间及抓拍地点;所述第二数据库,用于存储所述服务器生成的所述同行对象记录。
- 根据权利要求18所述的系统,其特征在于,所述系统还包括客户端;所述服务器,还用于获取同行对象查询请求,所述同行对象查询请求包括待查询的目标对象的对象信息;根据所述目标对象的对象信息,查找出现所述目标对象的所有同行对象记录,并统计目标对象的所有同行对象记录中与所述目标对象为同行对象的各对象出现次数;所述客户端,用于显示所述目标对象的所有同行对象记录中出现次数大于第二预设阈值的对象的同行对象记录。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911259493.X | 2019-12-10 | ||
CN201911259493.XA CN111435435A (zh) | 2019-12-10 | 2019-12-10 | 一种同行人识别方法、装置、服务器及系统 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021114985A1 true WO2021114985A1 (zh) | 2021-06-17 |
Family
ID=71580953
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/127500 WO2021114985A1 (zh) | 2019-12-10 | 2020-11-09 | 一种同行对象识别方法、装置、服务器及系统 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN111435435A (zh) |
WO (1) | WO2021114985A1 (zh) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114449144A (zh) * | 2022-01-04 | 2022-05-06 | 航天科工智慧产业发展有限公司 | 多路相机的抓拍联动装置及方法 |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111435435A (zh) * | 2019-12-10 | 2020-07-21 | 杭州海康威视数字技术股份有限公司 | 一种同行人识别方法、装置、服务器及系统 |
CN112651335A (zh) * | 2020-12-25 | 2021-04-13 | 深圳集智数字科技有限公司 | 一种同行人识别方法、系统、设备及存储介质 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108229335A (zh) * | 2017-12-12 | 2018-06-29 | 深圳市商汤科技有限公司 | 关联人脸识别方法和装置、电子设备、存储介质、程序 |
CN108733819A (zh) * | 2018-05-22 | 2018-11-02 | 深圳云天励飞技术有限公司 | 一种人员档案建立方法和装置 |
CN110334231A (zh) * | 2019-06-28 | 2019-10-15 | 深圳市商汤科技有限公司 | 一种信息处理方法及装置、存储介质 |
CN110348347A (zh) * | 2019-06-28 | 2019-10-18 | 深圳市商汤科技有限公司 | 一种信息处理方法及装置、存储介质 |
CN111435435A (zh) * | 2019-12-10 | 2020-07-21 | 杭州海康威视数字技术股份有限公司 | 一种同行人识别方法、装置、服务器及系统 |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108959667A (zh) * | 2017-05-18 | 2018-12-07 | 株式会社日立制作所 | 行人跟随行为仿真方法及行人跟随行为仿真装置 |
CN109889773A (zh) * | 2017-12-06 | 2019-06-14 | 中国移动通信集团四川有限公司 | 评标室人员的监控的方法、装置、设备和介质 |
CN108288025A (zh) * | 2017-12-22 | 2018-07-17 | 深圳云天励飞技术有限公司 | 一种车载视频监控方法、装置及设备 |
CN109117714B (zh) * | 2018-06-27 | 2021-02-26 | 北京旷视科技有限公司 | 一种同行人员识别方法、装置、系统及计算机存储介质 |
CN109299683B (zh) * | 2018-09-13 | 2019-12-10 | 嘉应学院 | 一种基于人脸识别和行为大数据的安防评估系统 |
CN109934176B (zh) * | 2019-03-15 | 2021-09-10 | 艾特城信息科技有限公司 | 行人识别系统、识别方法及计算机可读存储介质 |
CN110084103A (zh) * | 2019-03-15 | 2019-08-02 | 深圳英飞拓科技股份有限公司 | 一种基于人脸识别技术的同行人分析方法及系统 |
CN110276272A (zh) * | 2019-05-30 | 2019-09-24 | 罗普特科技集团股份有限公司 | 确认标签人员的同行人员关系的方法、装置、存储介质 |
CN110532929A (zh) * | 2019-08-23 | 2019-12-03 | 深圳市驱动新媒体有限公司 | 一种同行人分析方法和装置以及设备 |
-
2019
- 2019-12-10 CN CN201911259493.XA patent/CN111435435A/zh active Pending
-
2020
- 2020-11-09 WO PCT/CN2020/127500 patent/WO2021114985A1/zh active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108229335A (zh) * | 2017-12-12 | 2018-06-29 | 深圳市商汤科技有限公司 | 关联人脸识别方法和装置、电子设备、存储介质、程序 |
CN108733819A (zh) * | 2018-05-22 | 2018-11-02 | 深圳云天励飞技术有限公司 | 一种人员档案建立方法和装置 |
CN110334231A (zh) * | 2019-06-28 | 2019-10-15 | 深圳市商汤科技有限公司 | 一种信息处理方法及装置、存储介质 |
CN110348347A (zh) * | 2019-06-28 | 2019-10-18 | 深圳市商汤科技有限公司 | 一种信息处理方法及装置、存储介质 |
CN111435435A (zh) * | 2019-12-10 | 2020-07-21 | 杭州海康威视数字技术股份有限公司 | 一种同行人识别方法、装置、服务器及系统 |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114449144A (zh) * | 2022-01-04 | 2022-05-06 | 航天科工智慧产业发展有限公司 | 多路相机的抓拍联动装置及方法 |
CN114449144B (zh) * | 2022-01-04 | 2024-03-05 | 航天科工智慧产业发展有限公司 | 多路相机的抓拍联动装置及方法 |
Also Published As
Publication number | Publication date |
---|---|
CN111435435A (zh) | 2020-07-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021114985A1 (zh) | 一种同行对象识别方法、装置、服务器及系统 | |
US10970333B2 (en) | Distributed video storage and search with edge computing | |
EP3253042B1 (en) | Intelligent processing method and system for video data | |
Gauen et al. | Comparison of visual datasets for machine learning | |
CN105336077B (zh) | 数据处理设备和操作其的方法 | |
RU2632473C1 (ru) | Способ обмена данными между ip видеокамерой и сервером (варианты) | |
US20210357624A1 (en) | Information processing method and device, and storage medium | |
CN103942811A (zh) | 分布式并行确定特征目标运动轨迹的方法与系统 | |
CN102422286A (zh) | 利用图像获取参数和元数据自动和半自动的图像分类、注释和标签 | |
CN111222373B (zh) | 一种人员行为分析方法、装置和电子设备 | |
JP2022518459A (ja) | 情報処理方法および装置、記憶媒体 | |
CN109522421B (zh) | 一种网络设备的产品属性识别方法 | |
CN111209776A (zh) | 同行人识别方法、装置、处理服务器、存储介质及系统 | |
Xu et al. | Video analytics with zero-streaming cameras | |
CN106534784A (zh) | 一种用于视频分析数据结果集的采集分析存储统计系统 | |
CN112770265B (zh) | 一种行人身份信息获取方法、系统、服务器和存储介质 | |
CN109299307B (zh) | 一种基于结构分析的商标检索预警方法及装置 | |
CN112256682B (zh) | 一种多维异构数据的数据质量检测方法及装置 | |
CN110717358B (zh) | 访客人数统计方法、装置、电子设备及存储介质 | |
WO2023236514A1 (zh) | 一种跨摄像头的多目标追踪方法、装置、设备及介质 | |
WO2021184628A1 (zh) | 一种图像处理方法及装置 | |
CN111767432A (zh) | 共现对象的查找方法和装置 | |
CN114003672B (zh) | 一种道路动态事件的处理方法、装置、设备和介质 | |
WO2021104513A1 (zh) | 对象展示方法、装置、电子设备及存储介质 | |
CN111382281B (zh) | 基于媒体对象的内容的推荐方法、装置、设备和存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20897883 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20897883 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 23.05.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20897883 Country of ref document: EP Kind code of ref document: A1 |