WO2019020103A1 - 目标识别方法、装置、存储介质和电子设备 - Google Patents
目标识别方法、装置、存储介质和电子设备 Download PDFInfo
- Publication number
- WO2019020103A1 WO2019020103A1 PCT/CN2018/097374 CN2018097374W WO2019020103A1 WO 2019020103 A1 WO2019020103 A1 WO 2019020103A1 CN 2018097374 W CN2018097374 W CN 2018097374W WO 2019020103 A1 WO2019020103 A1 WO 2019020103A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- determined
- target
- path
- information
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/46—Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
- G06V20/47—Detecting features for summarising video content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/217—Validation; Performance evaluation; Active pattern learning techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/23—Clustering techniques
- G06F18/232—Non-hierarchical techniques
- G06F18/2321—Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/29—Graphical models, e.g. Bayesian networks
- G06F18/295—Markov models or related models, e.g. semi-Markov models; Markov random fields; Networks embedding Markov models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/06—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
- G06N3/063—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N7/00—Computing arrangements based on specific mathematical models
- G06N7/01—Probabilistic graphical models, e.g. probabilistic networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/143—Segmentation; Edge detection involving probabilistic approaches, e.g. Markov random field [MRF] modelling
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/48—Matching video sequences
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/52—Surveillance or monitoring of activities, e.g. for recognising suspicious objects
- G06V20/54—Surveillance or monitoring of activities, e.g. for recognising suspicious objects of traffic, e.g. cars on the road, trains or boats
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/047—Probabilistic or stochastic networks
Definitions
- Embodiments of the present invention relate to the field of artificial intelligence technologies, and in particular, to a target identification method, apparatus, storage medium, and electronic device.
- Re-identification of vehicles is an important content in the field of computer vision and public safety. It has important application value in many aspects such as vehicle detection and tracking, route estimation and abnormal behavior detection.
- vehicle re-identification techniques are based on the appearance information of the vehicle. Different from pedestrian re-identification, the difficulty in re-identifying the vehicle by simply using the vehicle appearance information lies in the appearance of many vehicles such as vehicles (such as color, Model, shape, etc.) are very similar. Especially in different vehicles of the same brand, the difference will be even smaller. Detecting and recognizing the unique details of vehicle identification information such as vehicle license plate information and vehicle interior decorations such as interior decorations may be caused by factors such as poor monitoring lens angle, poor lighting conditions, and lens blur. The robustness of detection and recognition is degraded, resulting in inaccurate detection and recognition results.
- the embodiment of the invention provides a technical solution for target recognition.
- a target recognition method includes: acquiring a first image and a second image, wherein the first image and the second image respectively include a target to be determined; and generating a predicted path based on the first image and the second image, The two ends of the prediction path respectively correspond to the first image and the second image; perform validity determination on the prediction path, and determine, in the first image and the second image, to be determined based on the determination result Whether the target is the same target to be determined.
- the target to be determined is a vehicle.
- the method before the generating the prediction path based on the first image and the second image, the method further includes: according to time information, spatial information, image feature information of the first image, and Determining, by the time information of the second image, the spatial information, the image feature information, a preliminary identical probability value of the target to be determined respectively included in the first image and the second image; And generating, by the second image, the prediction path, including: when the preliminary same probability value is greater than a preset value, generating the prediction path based on the first image and the second image.
- the difference between the difference and the spatial information is input to a Siamese Convolutional Neural Network (Siamese-CNN) to obtain preliminary identical probability values of the target to be determined in the first image and the second image.
- Siamese Convolutional Neural Network Siamese Convolutional Neural Network
- the generating a prediction path based on the first image and the second image including: according to feature information of the first image, time information of the first image, the first The spatial information of the image, the feature information of the second image, the time information of the second image, and the spatial information of the second image are used to generate a predicted path of the target to be determined by a probability model.
- the generating, by using the probability model, the predicted path of the target to be determined comprising: determining, by using a chained Markov Random Field (MRF), the acquired image set to include the Generating the information of the target, and all the images having the spatio-temporal sequence relationship with the first image and the second image; generating the prediction of the target to be determined according to the determined time information and spatial information corresponding to all the images path.
- MRF chained Markov Random Field
- the generating the predicted path of the target to be determined according to the determined time information and the spatial information of all the images including: generating time based on the determined time information and spatial information of all the images.
- the first image is a first node and the second image is a predicted path of the tail node, wherein the predicted path corresponds to at least one intermediate node in addition to the first node and the tail node.
- the MRF determines, from the acquired image set, all the images including the information of the object to be determined and the spatio-temporal sequence relationship with the first image and the second image, including : taking a position corresponding to the spatial information of the first image as a starting position, and using a position corresponding to the spatial information of the second image as a termination position, acquiring all the positions from the starting position to the ending position Position information of the image capturing device; generating, according to the relationship between the positions indicated by the position information of all the image capturing devices, the image capturing device corresponding to the starting position as the starting point, and the image capturing device corresponding to the ending position as the end point, generating at least one a device path, wherein each device path includes information of at least one other camera device in addition to the camera device of the start point and the camera device of the end point; for each device path, the time of the first image
- the time corresponding to the information is the start time
- the time corresponding to the time information of the second image is the end time, from each of
- the generating, according to the determined time information and spatial information corresponding to all the images, a prediction path that is the first image as the first node and the second image as the tail node including: a device path, generating, according to the determined time series relationship of the image, a plurality of connected intermediate nodes having a spatio-temporal sequence relationship; generating and the current device path according to the first node, the tail node, and the intermediate node Corresponding image path having a spatiotemporal sequence relationship; determining, from the image path corresponding to each device path, a maximum probability image path with the first image as the head node and the second image as the tail node as the to-be Determine the predicted path of the target.
- the maximum probability image path with the first image as the first node and the second image as the tail node is determined as the target to be determined from the image path corresponding to each device path.
- a prediction path comprising: obtaining, for each image path corresponding to each device path, a probability of having information of the same target to be determined between images of two adjacent nodes in the image path; according to each of the image paths a probability of having information of the same target to be determined between images of two adjacent nodes, calculating a probability that the image path is a predicted path of the target to be determined; and each image path is used as a predicted path of the target to be determined The probability of determining the maximum probability image path as the predicted path of the target to be determined.
- the validity of the prediction path is determined, and based on the determination result, determining whether the target to be determined in the first image and the second image are the same, including: predicting the prediction by using a neural network
- the path performs validity determination, and based on the determination result, determines whether the target to be determined in the first image and the second image is the same target to be determined.
- determining, by the neural network, the validity of the prediction path, and determining, according to the determination result, whether the target to be determined in the first image and the second image is the same target to be determined including: Obtaining time information of adjacent images in the prediction path, acquiring time differences of adjacent images; acquiring spatial differences of adjacent images according to spatial information of adjacent images; determining to be determined according to adjacent images Feature information of the target, obtaining feature differences of the target to be determined in the adjacent image; inputting time difference, spatial difference, and feature difference of adjacent images in the predicted path into the long and short time memory network (Long Short-Term Memory, LSTM), obtaining an identification probability of the target to be determined of the predicted path; determining, according to the recognition probability of the target to be determined of the predicted path, whether the target to be determined in the first image and the second image are the same The target is to be determined.
- LSTM Long Short-Term Memory
- acquiring feature differences of the target to be determined in the adjacent image according to the feature information of the target to be determined in the adjacent image including: acquiring, by the Siamese-CNN, the to-be-determined in the adjacent image respectively Feature information of the target; acquiring feature differences of the target to be determined in the adjacent image according to the separately acquired feature information.
- an object recognition apparatus configured to acquire a first image and a second image, where the first image and the second image respectively include a target to be determined; and a generating module configured to be based on the first image and the a second image, a prediction path is generated, the two ends of the prediction path respectively correspond to the first image and the second image; and the first determining module is configured to perform validity determination on the prediction path, based on the determination result Determining whether the target to be determined in the first image and the second image is the same target to be determined.
- the target to be determined is a vehicle.
- the device further includes: a second determining module, configured to: according to time information, spatial information, image feature information of the first image, and time information, spatial information of the second image, And determining, by the image feature information, preliminary preliminary probability values of the to-be-determined target respectively included in the first image and the second image; the generating module, comprising: a first generating sub-module configured to perform the preliminary same probability When the value is greater than the preset value, the predicted path is generated based on the first image and the second image.
- a second determining module configured to: according to time information, spatial information, image feature information of the first image, and time information, spatial information of the second image, And determining, by the image feature information, preliminary preliminary probability values of the to-be-determined target respectively included in the first image and the second image
- the generating module comprising: a first generating sub-module configured to perform the preliminary same probability When the value is greater than the preset value, the predicted path is generated based on the first image and the second image.
- the second determining module includes: a first determining submodule configured to: the first image and the second image, and the first image and the second image
- the difference between the time information and the difference in the spatial information is input to the Siamese-CNN, and preliminary preliminary probability values of the target to be determined in the first image and the second image are obtained.
- the generating module includes: a second generating submodule configured to: according to feature information of the first image, time information of the first image, spatial information of the first image, The feature information of the second image, the time information of the second image, and the spatial information of the second image are used to generate a predicted path of the target to be determined by using a probability model.
- the second generation sub-module includes: a first determining unit configured to determine, by the MRF, the information that includes the target to be determined from the acquired image set, and the first image and The second image has all the images of the spatio-temporal sequence relationship; the first generating unit is configured to generate the predicted path of the target to be determined according to the determined time information and spatial information of all the images.
- the first generating unit includes: a second generating unit, configured to generate, according to the determined time information and spatial information corresponding to all the images, the first image as a first node and the first The two images are a predicted path of the tail node, wherein the predicted path corresponds to at least one intermediate node in addition to the first node and the tail node.
- the first determining unit is configured to: take a position corresponding to the spatial information of the first image as a starting position, and use a position corresponding to the spatial information of the second image as a ending position to obtain Position information of all the image pickup apparatuses from the start position to the end position; according to the relationship between the positions indicated by the position information of all the image pickup apparatuses, starting from the image pickup apparatus corresponding to the start position, Generating at least one device path with the imaging device corresponding to the termination position as an end point, wherein each device path includes information of at least one other imaging device in addition to the imaging device of the starting point and the imaging device of the end point For each device path, the time corresponding to the time information of the first image is taken as the start time, and the time corresponding to the time information of the second image is the end time, from each other camera device on the current path In the captured image, a map of information including the target to be determined captured by a previous imaging device adjacent to the current imaging device is determined An image having information that sets a time series
- the second generating unit is configured to generate, according to the determined time series relationship of the image, a plurality of intermediate nodes having a spatio-temporal sequence relationship for each device path;
- the node, the tail node, and the intermediate node generate an image path having a spatio-temporal sequence relationship corresponding to the current device path; and determining, by the image path corresponding to each device path, that the first image is the first node
- the maximum probability image path with the second image as the tail node is used as the predicted path of the target to be determined.
- the second generating unit is further configured to: acquire, for each image path corresponding to each device path, an image of each two adjacent nodes in the image path having the same target to be determined Probability of information; calculating a probability of the image path as a predicted path of the target to be determined according to a probability of having information of the same target to be determined between images of two adjacent nodes in the image path; The image path determines the maximum probability image path as the predicted path of the target to be determined as the probability of the predicted path of the target to be determined.
- the first determining module includes: a second determining submodule configured to determine validity of the predicted path by using a neural network, and determine the first image and the second according to the determining result. Whether the target to be determined in the image is the same target to be determined.
- the second determining sub-module includes: a first acquiring unit, configured to acquire a time difference of adjacent images according to time information of adjacent images in the predicted path; The spatial information of the image is obtained, and the spatial difference of the adjacent image is obtained; according to the feature information of the target to be determined in the adjacent image, the feature difference of the target to be determined in the adjacent image is acquired; the second acquiring unit is configured as Inputting the time difference, the spatial difference, and the feature difference of the adjacent images in the obtained prediction path into the LSTM to obtain the recognition probability of the target to be determined of the predicted path; the second determining unit is configured to be according to the predicted path Determining the recognition probability of the target, determining whether the target to be determined in the first image and the second image is the same target to be determined.
- the first acquiring unit is configured to acquire feature information of the target to be determined in the adjacent image by using Siamese-CNN, and acquire the adjacent image according to the separately obtained feature information. The difference in characteristics of the target to be determined.
- a computer readable storage medium having stored thereon computer program instructions, wherein the program instructions, when executed by a processor, implement the first aspect of the embodiments of the present invention The steps of the target recognition method.
- an electronic device comprising: a processor, a memory, a communication component, and a communication bus, wherein the processor, the memory, and the communication component complete each other through the communication bus
- the memory is for storing at least one executable instruction that causes the processor to perform the steps of the object recognition method according to the first aspect of the embodiments of the present invention.
- a prediction path that the target to be determined may pass is generated; and the validity of the prediction path is determined to determine the first image and Whether the targets to be determined in the second image are the same.
- the validity judgment is a possibility of judging whether the current prediction path is a travel route of the same target to be determined, and the higher the probability, the target to be determined in the first image and the second image is the same target to be determined. The possibility is also higher. Thereby, it is possible to perform more accurate detection and recognition on whether the target to be determined in different images is the same target to be determined.
- FIG. 1 is a schematic flow chart of a target recognition method according to a first embodiment of the present invention
- FIG. 2 is a schematic flow chart of a target recognition method according to Embodiment 2 of the present invention.
- FIG. 3 is a schematic flow chart of a target recognition method according to Embodiment 3 of the present invention.
- FIG. 4 is a block diagram showing the structure of a target recognition apparatus according to Embodiment 4 of the present invention.
- Figure 5 is a block diagram showing the structure of a target recognition apparatus according to Embodiment 5 of the present invention.
- FIG. 6 is a structural block diagram of an object recognition apparatus according to Embodiment 6 of the present invention.
- FIG. 7 is a schematic structural diagram of an electronic device according to Embodiment 7 of the present invention.
- FIG. 1 is a schematic flow chart of a target recognition method according to a first embodiment of the present invention. As shown in FIG. 1, the target recognition method of this embodiment includes the following steps:
- step S102 the first image and the second image are acquired.
- the content to be determined is included in both the first image and the second image from the content included in the image.
- the first image and the second image may both be captured still images, or video images in a sequence of video frames, and the like.
- the target to be determined may include a pedestrian, a drone, a vehicle, and the like. It is to be understood that the embodiment is not limited thereto, and any movable object is included in the range of the target to be determined.
- step S104 a prediction path is generated based on the first image and the second image.
- the target may be determined based on the feature information of the target to be determined included in the first image and the second image and the spatiotemporal information included in the first image and the second image.
- the route of travel is predicted, and the reliability of the target identification to be determined is enhanced by the route prediction result.
- it is necessary to further find a possible travel route of the target to be determined in the image, wherein the target to be determined captured on the travel route The images should all be associated with the first image and the second image in time and space.
- step S106 the validity of the prediction path is determined, and based on the determination result, it is determined whether the target to be determined in the first image and the second image is the same target to be determined.
- the validity judgment is a possibility of determining whether a predicted path is a travel route of the same target to be determined, and the higher the probability, the target to be determined in the first image and the second image is the same target to be determined.
- the result of the validity judgment may specifically be an effective probability, or may be directly “effective”.
- the object recognition method based on the information included in the first image and the second image, a prediction path through which the target to be determined may pass is generated; and the validity of the prediction path is determined to determine the first image and Whether the targets to be determined in the second image are the same.
- the validity judgment is a possibility of judging whether the current prediction path is a travel route of the same target to be determined, and the higher the probability, the target to be determined in the first image and the second image is the same target to be determined. The possibility is also higher. Thereby, it is possible to perform more accurate detection and recognition on whether the target to be determined in different images is the same target to be determined.
- the target recognition method of this embodiment may be performed by any suitable device having image or data processing capability, including but not limited to: camera, terminal, mobile terminal, PC, server, in-vehicle device, entertainment device, advertising device, personal digital Assistants (PDAs), tablets, laptops, handheld game consoles, smart glasses, smart watches, wearables, virtual display devices or display enhancement devices (such as Google Glass, Oculus Rift, Hololens, Gear VR).
- PDAs personal digital Assistants
- tablets laptops, handheld game consoles, smart glasses, smart watches, wearables, virtual display devices or display enhancement devices (such as Google Glass, Oculus Rift, Hololens, Gear VR).
- FIG. 2 a flow chart of a target recognition method according to a second embodiment of the present invention is shown.
- the object recognition method of the embodiment of the present invention is described by taking the object to be determined as a vehicle as an example, but those skilled in the art should understand that in practical applications, other objects to be determined may refer to the embodiment. Achieve the corresponding target recognition operation.
- step S202 the first image and the second image are acquired.
- the first image and the second image each include a target to be determined, and the target to be determined is a vehicle.
- step S204 according to the feature information of the first image, the time information of the first image, the spatial information of the first image, the feature information of the second image, and the time information of the second image. And the spatial information of the second image, and the predicted path of the target to be determined is generated by a probability model.
- the travel route of the vehicle is more stable and regular, and the accuracy of judgment and recognition is higher. Therefore, the characteristic information of the vehicle (which can characterize the appearance of the vehicle) and the image can be jointly utilized. Temporal and spatial information to predict the travel route of the vehicle, and to enhance the credibility of the vehicle identification with the route prediction result.
- the time information of the image is used to indicate the time when the image is captured, and may be regarded as the time when the target (such as a vehicle) to be determined passes the shooting device; the spatial information of the image is used to indicate the position of the captured image, which may be considered as the location of the shooting device.
- the location may also be considered as the location of the target to be determined when the vehicle is photographed;
- the feature information of the image is used to indicate a feature of the target to be determined in the image, such as a feature of the vehicle, according to which the vehicle may be determined.
- Information such as appearance. It can be understood that the information included in the image related to the embodiment may include, but is not limited to, time information of the image, spatial information of the image, and feature information of the image.
- the probability model can be an MRF.
- a random field can be thought of as a collection of random variables corresponding to the same sample space. In general, when there is a dependency between these random variables, the random field can be considered to have practical significance.
- the random field contains two elements, the site and the phase space. When a value of phase space is randomly assigned to each location according to a certain distribution, the whole is called a random field.
- the MRF is a random field with a Markov nature restriction.
- the Markov property refers to the distribution of a random variable sequence in chronological order. The distribution characteristics at the N+1th time are independent of the value of the random variable before N.
- An MRF corresponds to an undirected graph. Each node on the undirected graph corresponds to a random variable, and the edge between the nodes indicates a probability dependency between the random variables corresponding to the node. Therefore, the structure of MRF essentially reflects a priori knowledge, that is, which variables have dependencies to consider and which can be ignored.
- At least one predicted path of the target to be determined in the first image and the second image may be generated by the MRF, and then the optimal path is determined therefrom as the predicted path of the target to be determined.
- the MRF may be adopted according to the feature information of the first image, the time information of the first image, the spatial information of the first image, the feature information of the second image, the time information of the second image, and the spatial information of the second image. Generating a predicted path of the target to be determined.
- all the images including the information of the object to be determined and having a spatio-temporal sequence relationship with the first image and the second image may be determined from the acquired image set by the chain MRF;
- the time information and the spatial information corresponding to all the images generate a predicted path of the target to be determined.
- spatio-temporal data refers to data having both time and space dimensions, including information of two dimensions of time and space.
- spatio-temporal data can be regarded as a time-series set with spatial correlation and a real-time null sequence.
- the data in the set can be thought of as data with a spatiotemporal sequence relationship.
- all the images having the spatio-temporal sequence relationship with the first image and the second image have meaning that the spatio-temporal data included in the all images and the spatio-temporal data included in the first image and the spatio-temporal data included in the second image are respectively Time and space are related.
- the first image is used as the path head node image
- the second image is the path tail node image
- the time information and spatial information corresponding to all the images determined by the chain MRF are generated, and the first image is used as the first node.
- the second image is a predicted path of the tail node, wherein the predicted path corresponds to the at least one intermediate node in addition to the first node and the tail node.
- the position corresponding to the spatial information of the image is the starting position
- the position corresponding to the spatial information of the second image is the ending position
- the position information of all the imaging devices from the starting position to the ending position is acquired; a relationship between the locations indicated by the location information of all the imaging devices, starting from the imaging device corresponding to the initial location, and generating an at least one device path with the imaging device corresponding to the termination location as an end point, wherein each The device path includes information of at least one other imaging device in addition to the imaging device of the starting point and the imaging device of the end point; for each device path, the time corresponding to the time information of the first image is taken as the starting time, The time taken by the time information of the second image is the end time, and the picture taken from each of the other camera devices on the current path , It is determined that the current captured image information imaging apparatus adjacent
- the determined image may be determined for each device path.
- the time series relationship generates a plurality of connected intermediate nodes having a spatio-temporal sequence relationship; and according to the first node, the tail node, and the intermediate node, generating an image path having a spatio-temporal sequence relationship corresponding to the current device path;
- a maximum probability image path with the first image as the first node and the second image as the tail node is determined as the predicted path of the target to be determined.
- each node on the chain is a camera, and the variable space of the node is taken by the camera.
- Siamese-CNN can be regarded as the potential energy function of adjacent nodes in MRF.
- the maximum subsequence and (Max-Sum) algorithm can be used to minimize (optimize) the product value of the potential energy function to obtain the most probable prediction path.
- the predicted path includes the geographic location of the camera through which the vehicle passes, the time taken, and related information of the captured image.
- setting p indicates information of the first image (including feature information, time information, and spatial information)
- q indicates information of the second image (including feature information, time information, and spatial information), which may be from a plurality of pieces by chain MRF
- One way to determine the optimal path in the predicted path can be achieved by maximizing the following formula (1):
- P represents the predicted path (ie, the predicted path through which the vehicle is likely to pass);
- X represents the camera;
- N represents the number of cameras on a predicted path, from X1 to XN, where x 1 represents the vehicle photographed by X1 Information of the image, and so on, x N represents information of the image of the vehicle captured by the XN,
- Indicates the potential energy function ie, the output of Siamese-CNN, a probability value between 0 and 1
- Representing the potential energy function pair between x i and x i+1 , x i and x i+1 are considered to contain information for the same vehicle. If x i and x i+1 do contain information about the same vehicle, then There will be a larger value, otherwise there will be a smaller value.
- the time constraint described in the formula (2) can be used to make the formula (2) satisfy the formula (3), namely:
- t is the time, with The optimal selection of the information of the image corresponding to x i and the optimal selection of the information of the image corresponding to x i+1 ;
- X represents the camera;
- N represents the number of cameras on a predicted path, from X1 to XN,
- x N represents information of an image of the vehicle photographed by XN.
- the information of the image includes time information, spatial information, and feature information of the image.
- the above formula (1) can be optimized to the following formula (4) to obtain an optimal path, that is, a maximum probability path through which the vehicle can pass.
- the first image is the prediction path head node A
- the second image is the prediction path tail node D.
- the possible driving route of the vehicle includes: route 1: A->B->C -> D; Route 2: A->E->D; Route 3: A->F->G->H->D.
- step S206 the validity of the prediction path is determined by the neural network, and based on the determination result, whether the target to be determined in the first image and the second image is the same target to be determined is determined.
- the neural network may be any suitable neural network that can implement feature extraction or target object recognition, including but not limited to a convolutional neural network, an enhanced learning neural network, a generation network in an anti-neural network, and the like.
- the configuration of the specific structure in the neural network may be appropriately set by a person skilled in the art according to actual needs, such as the number of layers of the convolution layer, the size of the convolution kernel, the number of channels, and the like, which are not limited in the embodiment of the present invention.
- the neural network can be an LSTM.
- LSTM is a time recurrent neural network, a variant of a cyclic neural network (RNN) that is better at processing sequence information.
- RNN cyclic neural network
- the predicted path of the vehicle may also be considered as a sequence information, which is processed by LSTM to determine the validity of the predicted path.
- the validity judgment is a possibility of judging whether a predicted path will be the same traveling route of the target to be determined, and the higher the probability, the possibility that the target to be determined in the first image and the second image is the same target to be determined. The higher the sex.
- the time difference of the adjacent image may be acquired according to the time information of the adjacent image in the prediction path; and the spatial difference of the adjacent image is obtained according to the spatial information of the adjacent image; Obtaining feature information of the target to be determined in the adjacent image, acquiring feature differences of the target to be determined in the adjacent image; inputting time difference, spatial difference, and feature difference of adjacent images in the obtained predicted path into the LSTM, Obtaining a recognition probability of the target to be determined of the predicted path; determining, according to the recognition probability of the target to be determined of the predicted path, whether the target to be determined in the first image and the second image is the same target to be determined.
- the specific setting of the criterion for determining whether the target is the same target to be determined may be appropriately set by a person skilled in the art according to actual needs, and the embodiment of the present invention does not limit this.
- the time difference of adjacent images can be obtained by subtracting the time information of the two; the spatial difference of the adjacent images can be obtained by calculating the distance between the two; and the feature difference of the adjacent images can pass the feature vectors of the two Subtraction is obtained.
- the feature information of the target to be determined in the adjacent image may be respectively acquired by Siamese-CNN by using Siamese-CNN; and the feature information acquired according to the separately acquired feature information is obtained.
- the Siamese-CNN in this step may be the same as or different from the Siamese-CNN in step S204.
- the judgment mode adopted is that the LSTM is used for judging, and the input of the LSTM is the time difference (ie, time difference) between the adjacent nodes on the route, the distance difference (ie, the spatial difference), and the difference in appearance thereof. (ie, feature difference), as described above, the difference in appearance can be obtained by directly subtracting the feature vectors output from the two images after input to Siamese-CNN.
- the output of the LSTM is a probability value by which the validity of the predicted path can be determined to determine whether the vehicles in the two images are indeed the same vehicle.
- a prediction path that the target to be determined in the image may pass is generated; and the validity of the prediction path is determined to determine the first Whether the target to be determined in an image and the second image is the same.
- the validity judgment is a possibility of judging whether the current prediction path is a travel route of the same target to be determined, and the higher the probability, the target to be determined in the first image and the second image is the same target to be determined. The possibility is also higher. Thereby, it is possible to perform more accurate detection and recognition on whether the target to be determined in different images is the same target to be determined.
- the target recognition method of this embodiment may be performed by any suitable device having image or data processing capability, including but not limited to: camera, terminal, mobile terminal, PC, server, in-vehicle device, entertainment device, advertising device, personal digital Assistants (PDAs), tablets, laptops, handheld game consoles, smart glasses, smart watches, wearables, virtual display devices or display enhancement devices (such as Google Glass, Oculus Rift, Hololens, Gear VR).
- PDAs personal digital Assistants
- tablets laptops, handheld game consoles, smart glasses, smart watches, wearables, virtual display devices or display enhancement devices (such as Google Glass, Oculus Rift, Hololens, Gear VR).
- FIG. 3 a flow chart of a target recognition method according to Embodiment 3 of the present invention is shown.
- the object recognition method of the embodiment of the present invention is described by taking the object to be determined as a vehicle as an example, but those skilled in the art should understand that in practical applications, other objects to be determined may refer to the embodiment. Achieve the corresponding target recognition operation.
- step S302 the first image and the first image are determined according to time information, spatial information, image feature information of the first image, and time information, spatial information, and image feature information of the second image.
- the first image and the second image both contain information of the target to be determined.
- the first image and the second image have a spatio-temporal sequence relationship, and each includes information of a corresponding target to be determined, and based on comprehensive consideration of time information, spatial information, and image feature information of the image, the field
- the technician can initially determine the preliminary identical probability values of the targets to be determined in the two images by any suitable method.
- a preliminary identical probability value of the target to be determined respectively included in the first image and the second image may be obtained using Siamese-CNN.
- Siamese-CNN is a CNN with at least two branches. It can receive multiple inputs at the same time and output the similarity of the multiple inputs (which can be expressed in the form of probability). Taking the double branch as an example, two images can be input simultaneously to the Siamese-CNN through the double branch, and the Siamese-CNN will output the similarity between the two images, or output a judgment result of whether the two images are similar.
- the Siamese-CNN in this embodiment includes three branches, two of which are used to receive an input image, and one branch is used to receive a difference (time difference) between time information between two images input and a difference in spatial information. (space difference).
- the similarity (such as appearance similarity) of the target object (the vehicle in this embodiment) in the output image is detected, and the difference of the input time information and the difference of the spatial information are detected.
- the similarity of the target object in time and space in the output image According to the similarity of the two aspects, the target object in the image, such as the preliminary identical probability value of the vehicle in the present embodiment, can be further determined.
- the first image and the second image, and the difference of the time information between the first image and the second image and the difference of the spatial information can be input into the Siamese-CNN to obtain the first image.
- preliminary preliminary probability values of the target to be determined in the second image After obtaining the preliminary same probability value, it is initially determined according to the preliminary same probability value that the first image and the second image have the same target to be determined.
- comparing the preliminary same probability value with a preset value when the preliminary same probability value is less than or equal to a preset value, determining that the first image and the second image do not have the same target to be determined And determining, when the preliminary same probability value is greater than the preset value, that the first image and the second image have the same target to be determined.
- the preset value may be appropriately set by a person skilled in the art according to the actual situation, which is not limited by the embodiment of the present invention.
- Siamese-CNN can effectively judge the degree of similarity of target objects such as vehicles in two images with spatio-temporal information, but not limited to Siamese-CNN.
- Other methods or neural networks that have similar functions or can achieve the same purpose are also applicable. The solution of the embodiment of the present invention.
- step S304 when the preliminary same probability value is greater than a preset value, the predicted path is generated based on the first image and the second image.
- the route to be determined such as the route of the vehicle
- the vehicle's feature information which can characterize the appearance of the vehicle
- space-time information can be used in combination to make the route of the vehicle. It is estimated that the credibility of vehicle re-identification will be enhanced by route estimation results.
- the first image and the second image are images having a spatiotemporal sequence relationship, and based on this, it is necessary to further find a possible travel route of the vehicle in the image, wherein the vehicle photographed on the travel route
- the images should all have a spatiotemporal sequence relationship with the first image and the second image.
- the predicted path of the target to be determined is generated using the MRF according to the information of the first image and the information of the second image.
- the specific implementation process is similar to the step S204 in the foregoing Embodiment 2, and details are not described herein again.
- step S306 the validity of the prediction path is determined, and based on the determination result, whether the target to be determined in the first image and the second image is the same target to be determined is re-identified.
- the validity judgment is a possibility of determining whether a predicted path is a travel route of the same target to be determined, and the higher the probability, the target to be determined in the first image and the second image is the same target to be determined. The possibility is also higher.
- the initially determined result is itself erroneous, that is, the vehicle in the first image and the vehicle in the second image may not be the same vehicle, but are mistaken. Recognized as the same vehicle. If the two are not the same vehicle, the probability that the two have the same driving route within a reasonable time range is very low. Therefore, the prediction path is determined according to the information of the first image and the information of the second image. The validity is also low, whereby re-judgement and recognition of whether the vehicle in the first image and the second image is the same vehicle can be achieved.
- the validity of the prediction path is determined by using the LSTM, and whether the target to be determined in the first image and the second image is the same target to be determined is re-identified according to the determination result.
- the specific implementation process is similar to the step S206 in the foregoing Embodiment 2, and details are not described herein again.
- the object recognition method on the basis of initially determining that the to-be-determined targets respectively included in the first image and the second image are the same, determining a prediction path that the target to be determined may pass; Judging the validity of the prediction path, determining whether the initially determined result is accurate, so as to realize whether the target to be determined in the first image and the second image is re-identified by the same target to be determined.
- the validity judgment is a possibility of judging whether the current prediction path is a travel route of the same target to be determined, and the higher the probability, the target to be determined in the first image and the second image is the same target to be determined. The possibility is also higher. Thereby, it is possible to perform more accurate re-detection and recognition on whether the target to be determined in different images is the same target to be determined.
- the target recognition method of this embodiment may be performed by any suitable device having image or data processing capability, including but not limited to: camera, terminal, mobile terminal, PC, server, in-vehicle device, entertainment device, advertising device, personal digital Assistants (PDAs), tablets, laptops, handheld game consoles, smart glasses, smart watches, wearables, virtual display devices or display enhancement devices (such as Google Glass, Oculus Rift, Hololens, Gear VR).
- PDAs personal digital Assistants
- tablets laptops, handheld game consoles, smart glasses, smart watches, wearables, virtual display devices or display enhancement devices (such as Google Glass, Oculus Rift, Hololens, Gear VR).
- FIG. 4 is a block diagram showing the structure of an object recognition apparatus according to Embodiment 4 of the present invention. It can be used to execute the target recognition method flow as described in the first embodiment.
- the target recognition apparatus includes an acquisition module 401, a generation module 402, and a first determination module 403.
- the obtaining module 401 is configured to acquire the first image and the second image, where the first image and the second image respectively include a target to be determined;
- the generating module 402 is configured to generate a prediction path based on the first image and the second image, where two ends of the prediction path respectively correspond to the first image and the second image;
- the first determining module 403 is configured to perform validity determination on the predicted path, and determine, according to the determination result, whether the target to be determined in the first image and the second image is the same target to be determined.
- the target recognition apparatus provided in this embodiment generates a prediction path through which the target to be determined may pass based on the information included in the first image and the second image, and determines the first image by determining the validity of the prediction path. Whether the targets to be determined in the second image are the same.
- the validity judgment is a possibility of judging whether the current prediction path is a travel route of the same target to be determined, and the higher the probability, the target to be determined in the first image and the second image is the same target to be determined. The possibility is also higher. Thereby, it is possible to perform more accurate detection and recognition on whether the target to be determined in different images is the same target to be determined.
- FIG. 5 is a block diagram showing the structure of a target recognition apparatus according to Embodiment 5 of the present invention. It can be used to perform the target recognition method flow as described in the second embodiment.
- the target recognition apparatus includes an acquisition module 501, a generation module 502, and a first determination module 503.
- the acquiring module 501 is configured to acquire the first image and the second image, where the first image and the second image respectively include a target to be determined
- the generating module 502 is configured to be based on the first image and the a second image, a prediction path is generated, and the two ends of the prediction path respectively correspond to the first image and the second image
- the first determining module 503 is configured to perform validity determination on the prediction path, based on the determination result Determining whether the target to be determined in the first image and the second image is the same target to be determined.
- the generating module 502 includes: a second generating submodule 5021, configured to: according to feature information of the first image, time information of the first image, spatial information of the first image And the feature information of the second image, the time information of the second image, and the spatial information of the second image, and the predicted path of the target to be determined is generated by a probability model.
- the second generation sub-module 5021 includes: a first determining unit 5022, configured to determine information including the target to be determined from the acquired image set by using an MRF, and the first The image and the second image both have all images in a spatiotemporal sequence relationship; the first generating unit 5023 is configured to generate a predicted path of the target to be determined according to the determined time information and spatial information corresponding to all the images.
- the first generating unit 5023 includes: a second generating unit 5024, configured to generate, according to the determined time information and spatial information of all the images, the first image as a first node and The second image is a predicted path of the tail node, wherein the predicted path corresponds to at least one intermediate node in addition to the first node and the tail node.
- the first determining unit 5022 is configured to: start a position corresponding to the spatial information of the first image, and use a position corresponding to the spatial information of the second image as a termination position, where Obtaining position information of all the imaging devices from the starting position to the ending position; according to the relationship between the positions indicated by the position information of all the imaging devices, starting from the imaging device corresponding to the starting position Generating at least one device path with the imaging device corresponding to the termination position as an end point, wherein each device path includes at least one other imaging device in addition to the imaging device of the starting point and the imaging device of the end point Information; for each device path, the time corresponding to the time information of the first image is the start time, and the time corresponding to the time information of the second image is the end time, from each other camera on the current path In the image captured by the device, the information of the target to be determined captured by the previous imaging device adjacent to the current imaging device is determined.
- the image has an image that sets a time series relationship and contains information of the object to be determined
- the second generating unit 5024 is configured to: for each device path, generate a plurality of connected intermediate nodes having a spatio-temporal sequence relationship according to the determined time series relationship of the image; The first node, the tail node, and the intermediate node generate an image path having a spatio-temporal sequence relationship corresponding to the current device path; and determining, by the image path corresponding to each device path, the first image as a first node And the maximum probability image path with the second image as the tail node is used as the predicted path of the target to be determined.
- the second generating unit 5024 is further configured to: obtain, for each image path corresponding to each device path, an image of each two adjacent nodes in the image path having the same target to be determined The probability of the information; calculating the probability that the image path is the predicted path of the target to be determined according to the probability of having information of the same target to be determined between the images of every two adjacent nodes in the image path; An image path is used as the probability of the predicted path of the target to be determined to determine a maximum probability image path as the predicted path of the target to be determined.
- the first determining module 503 includes: a second determining submodule 5031, configured to perform validity determination on the predicted path by using a neural network, and determine the first image and based on the determining result. Whether the target to be determined in the second image is the same target to be determined.
- the second determining submodule 5031 includes: a first obtaining unit 5032, configured to acquire a time difference of adjacent images according to time information of adjacent images in the predicted path; The spatial information of the adjacent image is obtained, and the spatial difference of the adjacent image is obtained; and the feature difference of the target to be determined in the adjacent image is acquired according to the feature information of the target to be determined in the adjacent image; the second acquiring unit 5033 And configured to input the time difference, the spatial difference, and the feature difference of the adjacent images in the obtained prediction path into the LSTM, to obtain the recognition probability of the target to be determined of the predicted path; and the second determining unit 5034 is configured to Determining, by the recognition probability of the target to be determined of the predicted path, determining whether the target to be determined in the first image and the second image is the same target to be determined.
- the first acquiring unit 5032 is configured to: respectively acquire feature information of the target to be determined in the adjacent image by using Siamese-CNN; and acquire adjacent images according to the separately acquired feature information. The difference in characteristics of the target to be determined.
- FIG. 6 is a block diagram showing the structure of an object recognition apparatus according to Embodiment 6 of the present invention. It can be used to perform the target recognition method flow as described in the third embodiment.
- the target recognition apparatus includes an acquisition module 601, a generation module 603, and a first determination module 604.
- the obtaining module 601 is configured to acquire the first image and the second image, where the first image and the second image respectively include a target to be determined
- the generating module 603 is configured to be based on the first image and the a second image, a prediction path is generated, the two ends of the prediction path respectively correspond to the first image and the second image
- the first determining module 604 is configured to perform validity determination on the prediction path, based on the determination result Determining whether the target to be determined in the first image and the second image are the same.
- the target to be determined is a vehicle.
- the apparatus further includes: a second determining module 602 configured to: time information according to the first image, spatial information, image feature information, and time information and spatial information of the second image And determining, by the image feature information, preliminary preliminary probability values of the target to be determined respectively included in the first image and the second image; correspondingly, the generating module 603 includes: a first generating submodule 6031 configured to When the preliminary same probability value is greater than a preset value, the predicted path is generated based on the first image and the second image.
- the second determining module 602 includes: a first determining submodule 6021 configured to combine the first image and the second image, and the first image and the second The difference in temporal information between the images and the difference in spatial information are input to the Siamese-CNN, and preliminary preliminary probability values of the objects to be determined in the first image and the second image are obtained.
- Embodiment 7 of the present invention provides an electronic device, which may be, for example, a mobile terminal, a personal computer (PC), a tablet computer, a server, or the like.
- the electronic device 700 includes one or more processors, communication components Etc., the one or more processors, for example: one or more central processing units (CPUs) 701, and/or one or more image processors (GPUs) 713, etc., may be stored in a read-only memory ( Executable instructions in ROM) 702 or executable instructions loaded from random access memory (RAM) 703 from storage portion 708 perform various appropriate actions and processes.
- the communication component includes a communication component 712 and/or a communication interface 709.
- the communication component 712 can include, but is not limited to, a network card, which can include, but is not limited to, an IB (Infiniband) network card
- the communication interface 709 includes a communication interface of a network interface card such as a LAN card, a modem, etc.
- the communication interface 709 is via, for example, the Internet.
- the network performs communication processing.
- the processor can communicate with read only memory 702 and/or random access memory 703 to execute executable instructions, communicate with communication component 712 over communication bus 704, and communicate with other target devices via communication component 712, thereby completing embodiments of the present invention.
- Providing an operation corresponding to any one of the target recognition methods for example, acquiring the first image and the second image, wherein the first image and the second image each include a target to be determined; based on the first image and the a second image, a prediction path is generated, the two ends of the prediction path respectively corresponding to the first image and the second image; determining validity of the prediction path, determining the first image based on the determination result Whether the targets to be determined in the second image are the same.
- RAM 703 various programs and data required for the operation of the device can be stored.
- the CPU 701 or the GPU 713, the ROM 702, and the RAM 703 are connected to each other through a communication bus 704.
- ROM 702 is an optional module.
- the RAM 703 stores executable instructions or writes executable instructions to the ROM 702 at runtime, the executable instructions causing the processor to perform operations corresponding to the above-described communication methods.
- An input/output (I/O) interface 705 is also coupled to communication bus 704.
- the communication component 712 can be integrated or can be configured to have multiple sub-modules (e.g., multiple IB network cards) and be on a communication bus link.
- the following components are connected to the I/O interface 705: an input portion 706 including a keyboard, a mouse, etc.; an output portion 707 including a cathode ray tube (CRT), a liquid crystal display (LCD), and the like, and a speaker; a storage portion 708 including a hard disk or the like And a communication interface 709 including a network interface card such as a LAN card, modem, or the like.
- Driver 710 is also connected to I/O interface 705 as needed.
- a removable medium 711 such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory or the like, is mounted on the drive 710 as needed so that a computer program read therefrom is installed into the storage portion 708 as needed.
- FIG. 7 is only an optional implementation manner.
- the number and type of components in FIG. 7 may be selected, deleted, added, or replaced according to actual needs;
- Different function components can also be implemented in separate settings or integrated settings, such as GPU and CPU detachable settings or GPU can be integrated on the CPU, communication components can be separated, or integrated on the CPU or GPU. and many more.
- an embodiment of the invention includes a computer program product comprising a computer program tangibly embodied on a machine readable medium, the computer program comprising program code for executing the method illustrated in the flowchart, the program code comprising the corresponding execution
- the instructions corresponding to the method steps provided by the embodiment of the present invention for example, acquiring the first image and the second image, wherein the first image and the second image respectively include a target to be determined; based on the first image and the a second image, a prediction path is generated, the two ends of the prediction path respectively corresponding to the first image and the second image; determining validity of the prediction path, determining the first image based on the determination result Whether the targets to be determined in the second image are the same.
- the computer program can be downloaded and installed from the network via a communication component, and/or installed from the removable media 711.
- the above method according to an embodiment of the present invention may be implemented in hardware, firmware, or implemented as software or computer code that may be stored in a recording medium such as a CD ROM, a RAM, a floppy disk, a hard disk, or a magneto-optical disk, or implemented by
- the network downloads computer code originally stored in a remote recording medium or non-transitory machine readable medium and stored in a local recording medium so that the methods described herein can be stored using a general purpose computer, a dedicated processor or programmable
- Such software processing on a recording medium of dedicated hardware such as an ASIC or an FPGA.
- a computer, processor, microprocessor controller or programmable hardware includes storage components (eg, RAM, ROM, flash memory, etc.) that can store or receive software or computer code, when the software or computer code is The processing methods described herein are implemented when the processor or hardware is accessed and executed. Moreover, when a general purpose computer accesses code for implementing the processing shown herein, the execution of the code converts the general purpose computer into a special purpose computer for performing the processing shown herein.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Data Mining & Analysis (AREA)
- Artificial Intelligence (AREA)
- Software Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- General Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- Molecular Biology (AREA)
- Probability & Statistics with Applications (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Neurology (AREA)
- Algebra (AREA)
- Computational Mathematics (AREA)
- Mathematical Analysis (AREA)
- Mathematical Optimization (AREA)
- Pure & Applied Mathematics (AREA)
- Image Analysis (AREA)
- Traffic Control Systems (AREA)
Abstract
Description
Claims (28)
- 一种目标识别方法,包括:获取第一图像及第二图像,所述第一图像及所述第二图像中均包含待确定目标;基于所述第一图像及所述第二图像,生成预测路径,所述预测路径的两端分别对应所述第一图像及所述第二图像;对所述预测路径进行有效性判断,基于判断结果,确定所述第一图像及所述第二图像中的待确定目标是否为同一待确定目标。
- 根据权利要求1所述的方法,其中,所述待确定目标为交通工具。
- 根据权利要求1所述的方法,其中,所述基于所述第一图像及所述第二图像,生成预测路径之前,所述方法还包括:根据所述第一图像的时间信息、空间信息、图像特征信息,以及,所述第二图像的时间信息、空间信息、图像特征信息,确定所述第一图像及所述第二图像中分别包含的待确定目标的初步相同概率值;所述基于所述第一图像及所述第二图像,生成预测路径,包括:当所述初步相同概率值大于预设值时,基于所述第一图像及所述第二图像,生成所述预测路径。
- 根据权利要求3所述的方法,其中,根据所述第一图像的时间信息、空间信息、图像特征信息,以及,所述第二图像的时间信息、空间信息、图像特征信息,确定所述第一图像及所述第二图像中分别包含的待确定目标的初步相同概率值,包括:将所述第一图像和所述第二图像,以及,所述第一图像和所述第二图像之间的时间信息的差异和空间信息的差异输入孪生卷积神经网络Siamese-CNN,获得所述第一图像和第二图像中的待确定目标的初步相同概率值。
- 根据权利要求1-4任一项所述的方法,其中,所述基于所述第一图像及所述第二图像,生成预测路径,包括:根据所述第一图像的特征信息、所述第一图像的时间信息、所述第一图像的空间信息、所述第二图像的特征信息、所述第二图像的时间信息、和所述第二图像的空间信息,通过概率模型生成所述待确定目标的预测路径。
- 根据权利要求5所述的方法,其中,所述通过概率模型生成所述待确定目标的预测路径,包括:通过链状马尔可夫随机场模型MRF从获取的图像集中确定出包含有所述待确定目标的信息、且与所述第一图像和所述第二图像均具有时空序列关系的所有图像;根据确定出的所有图像对应的时间信息和空间信息,生成所述待确定目标的预测路径。
- 根据权利要求6所述的方法,其中,所述根据确定出的所有图像对应的时间信息和空间信息,生成所述待确定目标的预测路径,包括:根据确定出的所有图像对应的时间信息和空间信息,生成以所述第一图像为首节点且以所述第二图像为尾节点的一条预测路径,其中,所述预测路径除对应所述首节点和所述尾节点外,还对应至少一个中间节点。
- 根据权利要求7所述的方法,其中,所述通过链状马尔可夫随机场模型MRF从获取的图像集中确定出包含有所述待确定目标的信息、且与所述第一图像和所述第二图像均具有时空序列关系的所有图像,包括:以所述第一图像的空间信息对应的位置为起始位置,以所述第二图像的空间信息对应的位置为终止位置,获取从所述起始位置至所述终止位置之间的所有摄像设备的位置信息;根据所有摄像设备的位置信息所指示的位置之间的关系,以所述起始位置对应的摄像设备为起点,以所述终止位置对应的摄像设备为终点,生成至少一个设备路径,其中,每个设备路径除包括所述起点的摄像设备和所述终点的摄像设备外,还包括至少一个其它摄像设备的信息;针对每一个设备路径,以所述第一图像的时间信息对应的时间为起始时间,以所述第二图像的时间信息对应的时间为终止时间,从当前路径上的每个其它摄像设备拍摄的图像中,确定出与当前摄像设备相邻的前一摄像设备拍摄的包含所述待确定目标的信息的图像具有设定时间序列关系、且包含所述待确定目标的信息的图像。
- 根据权利要求8所述的方法,其中,所述根据确定出的所有图像对应的时间信息和空间信息,生成以所述第一图像为首节点且以所述第二图像为尾节点的一条预测路径,包括:针对每一个设备路径,根据确定出的所述图像的时间序列关系生成相连的具有时空序列关系的多个中间节点;根据所述首节点、所述尾节点、和所述中间节点,生成与当前设备路径对应的具有时空序列关系的图像路径;从每一个设备路径对应的图像路径中,确定出以所述第一图像为首节点且以所述第二图像为尾节点的最大概率图像路径作为所述待确定目标的预测路径。
- 根据权利要求9所述的方法,其中,所述从每一个设备路径对应的图像路径中,确定出以所述第一图像为首节点且以所述第二图像为尾节点的最大概率图像路径作为所述待确定目标的预测路径,包括:针对每一个设备路径对应的图像路径,获取所述图像路径中每两个相邻节点的图像之间具有同一个待确定目标的信息的概率;根据所述图像路径中每两个相邻节点的图像之间具有同一个待确定目 标的信息的概率计算所述图像路径作为所述待确定目标的预测路径的概率;根据每一个图像路径作为所述待确定目标的预测路径的概率确定最大概率图像路径作为所述待确定目标的预测路径。
- 根据权利要求1-10任一项所述的方法,其中,所述对所述预测路径进行有效性判断,基于判断结果,确定所述第一图像及所述第二图像中的待确定目标是否为同一待确定目标,包括:通过神经网络,对所述预测路径进行有效性判断,基于判断结果,确定所述第一图像和第二图像中的待确定目标是否为同一个待确定目标。
- 根据权利要求11所述的方法,其中,所述通过神经网络,对所述预测路径进行有效性判断,基于判断结果,确定所述第一图像和第二图像中的待确定目标是否为同一个待确定目标,包括:根据所述预测路径中的相邻的图像的时间信息,获取相邻的图像的时间差异;根据相邻的图像的空间信息,获取相邻的图像的空间差异;根据相邻的图像中的待确定目标的特征信息,获取相邻的图像中的待确定目标的特征差异;将获得的所述预测路径中相邻的图像的时间差异、空间差异和特征差异输入长短时记忆网络LSTM,获得所述预测路径的待确定目标的识别概率;根据所述预测路径的待确定目标的识别概率,确定所述第一图像和第二图像中的待确定目标是否为同一个待确定目标。
- 根据权利要求12所述的方法,其中,根据相邻的图像中的待确定目标的特征信息,获取相邻的图像中的待确定目标的特征差异,包括:通过Siamese-CNN分别获取相邻的图像中的待确定目标的特征信息;根据分别获取的所述特征信息,获取相邻的图像中的待确定目标的特征差异。
- 一种目标识别装置,包括:获取模块,配置为获取第一图像及第二图像,所述第一图像及所述第二图像中均包含待确定目标;生成模块,配置为基于所述第一图像及所述第二图像,生成预测路径,所述预测路径的两端分别对应所述第一图像及所述第二图像;第一确定模块,配置为对所述预测路径进行有效性判断,基于判断结果,确定所述第一图像及所述第二图像中的待确定目标是否为同一待确定目标。
- 根据权利要求14所述的装置,其中,所述待确定目标为交通工具。
- 根据权利要求14所述的装置,其中,所述装置还包括:第二确定模块,配置为根据所述第一图像的时间信息、空间信息、图像特征信息,以及,所述第二图像的时间信息、空间信息、图像特征信息, 确定所述第一图像及所述第二图像中分别包含的待确定目标的初步相同概率值;所述生成模块,包括:第一生成子模块,配置为当所述初步相同概率值大于预设值时,基于所述第一图像及所述第二图像,生成所述预测路径。
- 根据权利要求16所述的装置,其中,所述第二确定模块,包括:第一确定子模块,配置为将所述第一图像和所述第二图像,以及,所述第一图像和所述第二图像之间的时间信息的差异和空间信息的差异输入孪生卷积神经网络Siamese-CNN,获得所述第一图像和第二图像中的待确定目标的初步相同概率值。
- 根据权利要求14-17任一项所述的装置,其中,所述生成模块,包括:第二生成子模块,配置为根据所述第一图像的特征信息、所述第一图像的时间信息、所述第一图像的空间信息、所述第二图像的特征信息、所述第二图像的时间信息、和所述第二图像的空间信息,通过概率模型生成所述待确定目标的预测路径。
- 根据权利要求18所述的装置,其中,所述第二生成子模块,包括:第一确定单元,配置为通过链状马尔可夫随机场模型MRF从获取的图像集中确定出包含有所述待确定目标的信息、且与所述第一图像和所述第二图像均具有时空序列关系的所有图像;第一生成单元,配置为根据确定出的所有图像对应的时间信息和空间信息,生成所述待确定目标的预测路径。
- 根据权利要求19所述的装置,其中,所述第一生成单元,包括:第二生成单元,配置为根据确定出的所有图像对应的时间信息和空间信息,生成以所述第一图像为首节点且以所述第二图像为尾节点的一条预测路径,其中,所述预测路径除对应所述首节点和所述尾节点外,还对应至少一个中间节点。
- 根据权利要求20所述的装置,其中,所述第一确定单元,配置为:以所述第一图像的空间信息对应的位置为起始位置,以所述第二图像的空间信息对应的位置为终止位置,获取从所述起始位置至所述终止位置之间的所有摄像设备的位置信息;根据所有摄像设备的位置信息所指示的位置之间的关系,以所述起始位置对应的摄像设备为起点,以所述终止位置对应的摄像设备为终点,生成至少一个设备路径,其中,每个设备路径除包括所述起点的摄像设备和所述终点的摄像设备外,还包括至少一个其它摄像设备的信息;针对每一个设备路径,以所述第一图像的时间信息对应的时间为起始时间,以所述第二图像的时间信息对应的时间为终止时间,从当前路径上的每个其它摄像设备拍摄的图像中,确定出与当前摄像设备相邻的前一摄 像设备拍摄的包含所述待确定目标的信息的图像具有设定时间序列关系、且包含所述待确定目标的信息的图像。
- 根据权利要求21所述的装置,其中,所述第二生成单元,配置为:针对每一个设备路径,根据确定出的所述图像的时间序列关系生成相连的具有时空序列关系的多个中间节点;根据所述首节点、所述尾节点、和所述中间节点,生成与当前设备路径对应的具有时空序列关系的图像路径;从每一个设备路径对应的图像路径中,确定出以所述第一图像为首节点且以所述第二图像为尾节点的最大概率图像路径作为所述待确定目标的预测路径。
- 根据权利要求22所述的装置,其中,所述第二生成单元,还配置为:针对每一个设备路径对应的图像路径,获取所述图像路径中每两个相邻节点的图像之间具有同一个待确定目标的信息的概率;根据所述图像路径中每两个相邻节点的图像之间具有同一个待确定目标的信息的概率计算所述图像路径作为所述待确定目标的预测路径的概率;根据每一个图像路径作为所述待确定目标的预测路径的概率确定最大概率图像路径作为所述待确定目标的预测路径。
- 根据权利要求14-23任一项所述的装置,其中,所述第一确定模块,包括:第二确定子模块,配置为通过神经网络,对所述预测路径进行有效性判断,基于判断结果,确定所述第一图像和第二图像中的待确定目标是否为同一个待确定目标。
- 根据权利要求24所述的装置,其中,所述第二确定子模块,包括:第一获取单元,配置为根据所述预测路径中的相邻的图像的时间信息,获取相邻的图像的时间差异;根据相邻的图像的空间信息,获取相邻的图像的空间差异;根据相邻的图像中的待确定目标的特征信息,获取相邻的图像中的待确定目标的特征差异;第二获取单元,配置为将获得的所述预测路径中相邻的图像的时间差异、空间差异和特征差异输入长短时记忆网络LSTM,获得所述预测路径的待确定目标的识别概率;第二确定单元,配置为根据所述预测路径的待确定目标的识别概率,确定所述第一图像和第二图像中的待确定目标是否为同一个待确定目标。
- 根据权利要求25所述的装置,其中,所述第一获取单元,配置为:通过Siamese-CNN分别获取相邻的图像中的待确定目标的特征信息;根据分别获取的所述特征信息,获取相邻的图像中的待确定目标的特征差异。
- 一种计算机可读存储介质,其上存储有计算机程序指令,其中,所述程序指令被处理器执行时实现权利要求1至13中任意一项权利要求所述的目标识别方法的步骤。
- 一种电子设备,包括:处理器、存储器、通信元件和通信总线,所述处理器、所述存储器和所述通信元件通过所述通信总线完成相互间的通信;所述存储器用于存放至少一可执行指令,所述可执行指令使所述处理器执行如权利要求1至13中任意一项权利要求所述的目标识别方法的步骤。
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
SG11201911625YA SG11201911625YA (en) | 2017-07-28 | 2018-07-27 | Target recognition method and apparatus, storage medium, and electronic device |
JP2019557616A JP6893564B2 (ja) | 2017-07-28 | 2018-07-27 | ターゲット識別方法、装置、記憶媒体および電子機器 |
KR1020197031657A KR102339323B1 (ko) | 2017-07-28 | 2018-07-27 | 타겟 인식 방법, 장치, 저장 매체 및 전자 기기 |
US16/565,069 US11200682B2 (en) | 2017-07-28 | 2019-09-09 | Target recognition method and apparatus, storage medium, and electronic device |
US17/452,776 US20220051417A1 (en) | 2017-07-28 | 2021-10-29 | Target recognition method and appartus, storage medium, and electronic device |
US17/453,487 US20220058812A1 (en) | 2017-07-28 | 2021-11-03 | Target recognition method and appartus, storage medium, and electronic device |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710633604.3 | 2017-07-28 | ||
CN201710633604.3A CN108229292A (zh) | 2017-07-28 | 2017-07-28 | 目标识别方法、装置、存储介质和电子设备 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/565,069 Continuation US11200682B2 (en) | 2017-07-28 | 2019-09-09 | Target recognition method and apparatus, storage medium, and electronic device |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2019020103A1 true WO2019020103A1 (zh) | 2019-01-31 |
Family
ID=62654256
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2018/097374 WO2019020103A1 (zh) | 2017-07-28 | 2018-07-27 | 目标识别方法、装置、存储介质和电子设备 |
Country Status (6)
Country | Link |
---|---|
US (3) | US11200682B2 (zh) |
JP (1) | JP6893564B2 (zh) |
KR (1) | KR102339323B1 (zh) |
CN (1) | CN108229292A (zh) |
SG (1) | SG11201911625YA (zh) |
WO (1) | WO2019020103A1 (zh) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2021017891A1 (zh) * | 2019-07-31 | 2021-02-04 | 腾讯科技(深圳)有限公司 | 对象跟踪方法和装置、存储介质及电子设备 |
CN112766301A (zh) * | 2020-12-11 | 2021-05-07 | 南京富岛信息工程有限公司 | 一种采油机示功图相似性判断方法 |
JP2022539250A (ja) * | 2019-07-03 | 2022-09-07 | ウェイモ エルエルシー | アンカー軌道を使用したエージェント軌道予測 |
Families Citing this family (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108229292A (zh) * | 2017-07-28 | 2018-06-29 | 北京市商汤科技开发有限公司 | 目标识别方法、装置、存储介质和电子设备 |
CN108921811B (zh) * | 2018-04-03 | 2020-06-30 | 阿里巴巴集团控股有限公司 | 检测物品损伤的方法和装置、物品损伤检测器 |
CN109508787A (zh) * | 2018-10-16 | 2019-03-22 | 深圳大学 | 用于超声位移估计的神经网络模型训练方法及系统 |
CN111160067A (zh) * | 2018-11-07 | 2020-05-15 | 北京奇虎科技有限公司 | 危险识别方法、装置、电子设备及计算机可读存储介质 |
CN109740479A (zh) * | 2018-12-25 | 2019-05-10 | 苏州科达科技股份有限公司 | 一种车辆重识别方法、装置、设备及可读存储介质 |
CN109726684B (zh) * | 2018-12-29 | 2021-02-19 | 百度在线网络技术(北京)有限公司 | 一种地标元素获取方法和地标元素获取系统 |
US10373323B1 (en) * | 2019-01-29 | 2019-08-06 | StradVision, Inc. | Method and device for merging object detection information detected by each of object detectors corresponding to each camera nearby for the purpose of collaborative driving by using V2X-enabled applications, sensor fusion via multiple vehicles |
CN110490906A (zh) * | 2019-08-20 | 2019-11-22 | 南京邮电大学 | 一种基于孪生卷积网络和长短期记忆网络的实时视觉目标跟踪方法 |
CN110728330A (zh) * | 2019-10-23 | 2020-01-24 | 腾讯科技(深圳)有限公司 | 基于人工智能的对象识别方法、装置、设备及存储介质 |
CN110991413B (zh) * | 2019-12-20 | 2020-12-15 | 西南交通大学 | 一种基于ReID的跑步检测方法 |
CN113627260A (zh) * | 2021-07-12 | 2021-11-09 | 科大讯飞股份有限公司 | 识别手写汉字的笔顺的方法、系统和计算设备 |
CN113673412B (zh) * | 2021-08-17 | 2023-09-26 | 驭势(上海)汽车科技有限公司 | 关键目标物的识别方法、装置、计算机设备及存储介质 |
CN113688776B (zh) * | 2021-09-06 | 2023-10-20 | 北京航空航天大学 | 一种用于跨视场目标重识别的时空约束模型构建方法 |
CN114338974A (zh) * | 2021-12-02 | 2022-04-12 | 深圳市领航卫士安全技术有限公司 | 多通道的活动路径确定方法、装置、设备及存储介质 |
CN114783003B (zh) * | 2022-06-23 | 2022-09-20 | 之江实验室 | 一种基于局部特征注意力的行人重识别方法和装置 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105095362A (zh) * | 2015-06-25 | 2015-11-25 | 深圳码隆科技有限公司 | 一种基于目标对象的图像显示方法和装置 |
CN106326837A (zh) * | 2016-08-09 | 2017-01-11 | 北京旷视科技有限公司 | 对象追踪方法和装置 |
CN106650660A (zh) * | 2016-12-19 | 2017-05-10 | 深圳市华尊科技股份有限公司 | 一种车型识别方法及终端 |
CN108229292A (zh) * | 2017-07-28 | 2018-06-29 | 北京市商汤科技开发有限公司 | 目标识别方法、装置、存储介质和电子设备 |
Family Cites Families (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001208844A (ja) * | 2000-01-25 | 2001-08-03 | Sumitomo Electric Ind Ltd | 光学式車両感知装置および光学式車両感知方法 |
US7623674B2 (en) * | 2003-11-05 | 2009-11-24 | Cognex Technology And Investment Corporation | Method and system for enhanced portal security through stereoscopy |
JP4447309B2 (ja) * | 2003-12-25 | 2010-04-07 | 財団法人生産技術研究奨励会 | 交差点交通量計測装置および交差点交通量計測方法 |
US9002060B2 (en) * | 2012-06-28 | 2015-04-07 | International Business Machines Corporation | Object retrieval in video data using complementary detectors |
JP5941987B2 (ja) * | 2012-08-08 | 2016-06-29 | 株式会社日立製作所 | 交通量予測装置および方法 |
US9436895B1 (en) * | 2015-04-03 | 2016-09-06 | Mitsubishi Electric Research Laboratories, Inc. | Method for determining similarity of objects represented in images |
JP6439571B2 (ja) * | 2015-04-28 | 2018-12-19 | オムロン株式会社 | 交通情報収集装置、交通情報収集方法、および交通情報収集プログラム |
JP6433877B2 (ja) * | 2015-10-27 | 2018-12-05 | 日本電信電話株式会社 | 目的地予測装置、目的地予測方法、及び目的地予測プログラム |
JP6521835B2 (ja) * | 2015-10-27 | 2019-05-29 | 日本電信電話株式会社 | 移動経路予測装置、移動経路予測方法、及び移動経路予測プログラム |
WO2017122086A1 (en) * | 2016-01-11 | 2017-07-20 | Mobileye Vision Technologies Ltd. | Systems and methods for augmenting upright object detection |
CN106778517A (zh) * | 2016-11-25 | 2017-05-31 | 河南高速公路驻信段改扩建工程有限公司 | 一种监控视频序列图像车辆再识别的方法 |
-
2017
- 2017-07-28 CN CN201710633604.3A patent/CN108229292A/zh active Pending
-
2018
- 2018-07-27 WO PCT/CN2018/097374 patent/WO2019020103A1/zh active Application Filing
- 2018-07-27 KR KR1020197031657A patent/KR102339323B1/ko active IP Right Grant
- 2018-07-27 SG SG11201911625YA patent/SG11201911625YA/en unknown
- 2018-07-27 JP JP2019557616A patent/JP6893564B2/ja active Active
-
2019
- 2019-09-09 US US16/565,069 patent/US11200682B2/en active Active
-
2021
- 2021-10-29 US US17/452,776 patent/US20220051417A1/en not_active Abandoned
- 2021-11-03 US US17/453,487 patent/US20220058812A1/en not_active Abandoned
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105095362A (zh) * | 2015-06-25 | 2015-11-25 | 深圳码隆科技有限公司 | 一种基于目标对象的图像显示方法和装置 |
CN106326837A (zh) * | 2016-08-09 | 2017-01-11 | 北京旷视科技有限公司 | 对象追踪方法和装置 |
CN106650660A (zh) * | 2016-12-19 | 2017-05-10 | 深圳市华尊科技股份有限公司 | 一种车型识别方法及终端 |
CN108229292A (zh) * | 2017-07-28 | 2018-06-29 | 北京市商汤科技开发有限公司 | 目标识别方法、装置、存储介质和电子设备 |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2022539250A (ja) * | 2019-07-03 | 2022-09-07 | ウェイモ エルエルシー | アンカー軌道を使用したエージェント軌道予測 |
JP7459224B2 (ja) | 2019-07-03 | 2024-04-01 | ウェイモ エルエルシー | アンカー軌道を使用したエージェント軌道予測 |
WO2021017891A1 (zh) * | 2019-07-31 | 2021-02-04 | 腾讯科技(深圳)有限公司 | 对象跟踪方法和装置、存储介质及电子设备 |
CN112766301A (zh) * | 2020-12-11 | 2021-05-07 | 南京富岛信息工程有限公司 | 一种采油机示功图相似性判断方法 |
CN112766301B (zh) * | 2020-12-11 | 2024-04-12 | 南京富岛信息工程有限公司 | 一种采油机示功图相似性判断方法 |
Also Published As
Publication number | Publication date |
---|---|
US20200005090A1 (en) | 2020-01-02 |
CN108229292A (zh) | 2018-06-29 |
SG11201911625YA (en) | 2020-01-30 |
US20220058812A1 (en) | 2022-02-24 |
US11200682B2 (en) | 2021-12-14 |
JP2020519989A (ja) | 2020-07-02 |
KR20190128724A (ko) | 2019-11-18 |
US20220051417A1 (en) | 2022-02-17 |
KR102339323B1 (ko) | 2021-12-14 |
JP6893564B2 (ja) | 2021-06-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2019020103A1 (zh) | 目标识别方法、装置、存储介质和电子设备 | |
US11643076B2 (en) | Forward collision control method and apparatus, electronic device, program, and medium | |
US11182592B2 (en) | Target object recognition method and apparatus, storage medium, and electronic device | |
US10599958B2 (en) | Method and system for classifying an object-of-interest using an artificial neural network | |
CN107851318B (zh) | 用于对象跟踪的系统和方法 | |
CN109035304B (zh) | 目标跟踪方法、介质、计算设备和装置 | |
US11600008B2 (en) | Human-tracking methods, systems, and storage media | |
US9760800B2 (en) | Method and system to detect objects using block based histogram of oriented gradients | |
TW202034270A (zh) | 車輛事故的鑑定方法及裝置、電子設備 | |
WO2020047854A1 (en) | Detecting objects in video frames using similarity detectors | |
US10671887B2 (en) | Best image crop selection | |
CN110660102B (zh) | 基于人工智能的说话人识别方法及装置、系统 | |
CN108229494B (zh) | 网络训练方法、处理方法、装置、存储介质和电子设备 | |
WO2016145591A1 (en) | Moving object detection based on motion blur | |
WO2015186347A1 (ja) | 検出システム、検出方法及びプログラム記憶媒体 | |
CN114663871A (zh) | 图像识别方法、训练方法、装置、系统及存储介质 | |
US10599946B2 (en) | System and method for detecting change using ontology based saliency | |
CN111079621A (zh) | 检测对象的方法、装置、电子设备和存储介质 | |
Qi et al. | Cascaded cast shadow detection method in surveillance scenes | |
JPWO2018179119A1 (ja) | 映像解析装置、映像解析方法およびプログラム | |
CN111310595A (zh) | 用于生成信息的方法和装置 | |
CN112819859B (zh) | 一种应用于智慧安防的多目标跟踪方法及装置 | |
US9842406B2 (en) | System and method for determining colors of foreground, and computer readable recording medium therefor | |
CN112347853A (zh) | 一种基于视频的车牌数据脱敏方法、存储介质及服务器 | |
CN113129333A (zh) | 多目标实时跟踪方法及其系统和电子设备 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 18838049 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2019557616 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20197031657 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18838049 Country of ref document: EP Kind code of ref document: A1 |