CN112380305B - High-precision map data crowdsourcing method and device based on space-time similarity - Google Patents

High-precision map data crowdsourcing method and device based on space-time similarity Download PDF

Info

Publication number
CN112380305B
CN112380305B CN202011227747.2A CN202011227747A CN112380305B CN 112380305 B CN112380305 B CN 112380305B CN 202011227747 A CN202011227747 A CN 202011227747A CN 112380305 B CN112380305 B CN 112380305B
Authority
CN
China
Prior art keywords
similarity
metadata
space
time
calculating
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011227747.2A
Other languages
Chinese (zh)
Other versions
CN112380305A (en
Inventor
黄睿
唐洁
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
South China University of Technology SCUT
Original Assignee
South China University of Technology SCUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South China University of Technology SCUT filed Critical South China University of Technology SCUT
Priority to CN202011227747.2A priority Critical patent/CN112380305B/en
Publication of CN112380305A publication Critical patent/CN112380305A/en
Application granted granted Critical
Publication of CN112380305B publication Critical patent/CN112380305B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/29Geographical information databases
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/23Updating
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/7867Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title and artist information, manually generated time, location and usage information, user ratings
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/787Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using geographical or spatial information, e.g. location
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5027Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/02Protocols based on web technology, e.g. hypertext transfer protocol [HTTP]
    • H04L67/025Protocols based on web technology, e.g. hypertext transfer protocol [HTTP] for remote control or remote monitoring of applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1001Protocols in which an application is distributed across nodes in the network for accessing one among a plurality of replicated servers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/12Protocols specially adapted for proprietary or special-purpose networking environments, e.g. medical networks, sensor networks, networks in vehicles or remote metering networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2209/00Indexing scheme relating to G06F9/00
    • G06F2209/50Indexing scheme relating to G06F9/50
    • G06F2209/502Proximity

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Software Systems (AREA)
  • Library & Information Science (AREA)
  • Multimedia (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Remote Sensing (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Image Analysis (AREA)

Abstract

A high-precision map data crowdsourcing method and device based on space-time similarity are disclosed, wherein the method comprises the following steps: crowdsourcing collected video data and extracting video metadata; calculating the time similarity among the metadata; calculating the spatial similarity among the metadata; calculating the space-time similarity between the video metadata according to the time and space similarity; and uploading the required video according to the space-time similarity, the bandwidth environment and the coverage required by the low points. The method and the device reduce the redundancy of data in the high-precision map so as to reduce the delay of transmission and complete map generation.

Description

High-precision map data crowdsourcing method and device based on space-time similarity
Technical Field
The invention relates to the technical field of data processing, in particular to a high-precision map data crowdsourcing method and device based on space-time similarity.
Background
Along with the gradual increase of population in recent years, science and technology is developed gradually, and along with the continuous development of urban traffic, people slowly feel that ordinary car can no longer satisfy present life needs now, consequently, people look the eye that develops towards unmanned driving direction, compare and drive in traditional car, unmanned car can bring huge profit for the car producer, and the human cost is showing to be reduced, and unmanned car can strengthen the security of vehicle road greatly moreover.
Compared with the traditional navigation map, the high-precision map serving automatic driving has higher requirements in all aspects, and can provide support for a decision layer by matching with a sensor and an algorithm, so that various requirements of unmanned driving are better met. The high-precision map is a 2D grid with high precision and fine definition, the main data is point cloud data generated by a laser radar, and the point cloud is derived from conversion of data of types such as video images and the like acquired by a vehicle through data acquisition. The high-precision map is one of key capabilities of realizing automatic driving, and the high-precision map can be an effective supplement to an automatic driving existing sensor and provides more reliable sensing capability for a vehicle. The high-precision map service system is important for unmanned driving, because the high-precision map service is a machine, the map has to reach certain accuracy to ensure that the safety of vehicles and people can be ensured under the unsupervised driving, and the vehicles can be ensured to avoid various obstacles and select the most suitable road according to the actual situation, which is completely impossible by the traditional electronic map, and the unmanned driving can be enabled only through a large amount of driving auxiliary information contained in the high-precision map.
Although the high-precision map is popular in research at present and various automobile manufacturers have unsophisticated achievements, in the research process, many unsolved problems still exist to restrict the research progress, and due to the nature of the high-precision map, the integral map data volume is huge, and the requirement of instantaneity required by the high-precision map cannot be met only by simply carrying out one-time transmission under the existing network conditions. For unmanned driving, the information provided by the map must be accurate and time-efficient, at this time, if the source of the picture or video is single, the transmission may cause delay due to bandwidth limitation, and the accuracy may not meet the requirement of unmanned driving due to the limitation of the angle and shooting location, so the high-accuracy map is updated in a crowd-sourced manner in the research, and the accuracy and the time-efficient performance of the map are guaranteed.
Disclosure of Invention
The application provides a high-precision map data crowdsourcing method and device based on space-time similarity, and aims to solve the problems that redundancy of data in a high-precision map is reduced, and transmission and delay of generation of a complete map are reduced.
According to a first aspect, an embodiment provides a high-precision map data crowdsourcing method based on spatiotemporal similarity, which includes:
crowdsourcing collected video data and extracting video metadata;
calculating the time similarity among the metadata;
calculating the spatial similarity among the metadata;
calculating the space-time similarity among the video metadata according to the time and space similarity;
and uploading the required video according to the space-time similarity, the bandwidth environment and the coverage required by the low points.
In some embodiments, the step of calculating spatial similarity between metadata comprises:
simplifying the representation of two different video frame metadata;
decomposing the rigid motion of the two simplified video metadata into translation and rotation and respectively calculating the similarity of the two motions;
calculating a rotated coverage intersection and a translated coverage intersection;
and obtaining the similarity between the two metadata according to the rotated coverage intersection and the translated coverage intersection.
In some embodiments, the computing the coverage intersection of translations is computing a similarity of translations, including: calculating from the vertical direction and the horizontal direction to respectively obtain the similarity in the vertical direction and the similarity in the horizontal direction; and performing weighted calculation on the similarity in the vertical direction and the similarity in the horizontal direction to obtain the translation similarity.
In some embodiments, two different video frame metadata f 1 、f 2 Simplified representation f 1 =(l 11 ) And f 2 =(l 22 ),
Wherein, theta 1 And theta 2 Representing the angular extent of two different video frame metadata, l 1 And l 2 Representing the shooting locations of two different video frame metadata;
the rotated coverage intersection is represented as:
Figure GDA0003638965560000031
wherein,
Figure GDA0003638965560000032
indicating the size of the field of view of the camera lens,
Figure GDA0003638965560000033
a view difference representing metadata of two different video frames;
the coverage intersection of the translations is represented as:
Figure GDA0003638965560000034
wherein Ssec is the similarity in the vertical direction,
Figure GDA0003638965560000035
r denotes the effective range of the camera, δ l Is represented by 1 And l 2 The distance between the two or more of the two or more,
Figure GDA0003638965560000036
δ θ =min(|θ 21 |,2π-|θ 21 |),δ θ representing the difference in orientation of two different video frame metadata,
the similarity in the horizontal direction can be calculated by the following equation:
Figure GDA0003638965560000037
two different video frame metadata f 1 And f 2 The spatial similarity between them is expressed as:
Ssem (f 1,f2)=Ssem R ×Ssem T
in some embodiments, the calculating the temporal similarity between the metadata comprises:
calculating the time measurement relation correlation degree:
Figure GDA0003638965560000038
calculating the time similarity:
Figure GDA0003638965560000043
the method of claim 1, wherein the calculating the spatiotemporal similarity between the video metadata comprises: calculating the measurement relation of the space-time similarity and calculating the topological relation of the space-time similarity,
the metric relation for calculating the space-time similarity is to calculate the integral of the area S along the lj change curve:
Figure GDA0003638965560000041
the topological relation for calculating the space-time similarity is to obtain the space similarity Ssem between two different video frame metadata by decomposing the operation of the topological relation and the orientation relation on each frame (f1,f2) And then, associating with the time similarity to obtain a topological relation of the space-time similarity:
Stsem TA(f1,f2) =Tsem×Ssem (f1,f2)
the spatio-temporal similarity is expressed as:
Figure GDA0003638965560000042
in the formula: stsem is the total spatiotemporal similarity; w ts ,W tta Are respectively space and time associated weight values and satisfy W ts +W tta =1, large similarity of two video metadataA small range is between 0 and 1.
According to a third aspect, an embodiment provides a high-precision map data crowdsourcing device based on space-time similarity, comprising:
an acquisition module: the method comprises the steps of collecting video data through crowdsourcing, and extracting video metadata;
a time module: for calculating the time similarity between metadata;
a space module: the spatial similarity between metadata is calculated;
a calculation module: the video processing device is used for calculating the space-time similarity among the video metadata according to the time and space similarities;
an uploading module: the method is used for uploading the required video according to the space-time similarity, the bandwidth environment and the coverage required by the low points.
According to a third aspect, an embodiment provides a high-precision map data crowdsourcing device based on space-time similarity, including:
a memory for storing a program;
a processor for implementing the method as described in the first aspect by executing the program stored by the memory.
According to a fourth aspect, there is provided in an embodiment a computer readable storage medium comprising a program executable by a processor to implement the method of the first aspect.
According to the embodiment, the high-precision map data crowdsourcing method and system based on the space-time similarity are provided, crowdsourcing data acquisition and edge calculation are combined, and meanwhile the operating pressure of a cloud system is remarkably reduced.
Drawings
FIG. 1 is a flow chart of a high-precision map data crowdsourcing method based on spatiotemporal similarity provided by the application;
FIG. 2 is a system operational scenario diagram of an embodiment;
FIG. 3 is a flow chart of a spatio-temporal similarity computation method according to an embodiment.
Detailed Description
The present invention will be described in further detail with reference to the following detailed description and accompanying drawings. Wherein like elements in different embodiments have been given like element numbers associated therewith. In the following description, numerous specific details are set forth in order to provide a better understanding of the present application. However, those skilled in the art will readily recognize that some of the features may be omitted or replaced with other elements, materials, methods in different instances. In some instances, certain operations related to the present application have not been shown or described in this specification in order not to obscure the core of the present application with unnecessary detail, and it is not necessary for those skilled in the art to describe these operations in detail, so that they may be fully understood from the description in the specification and the general knowledge in the art.
Furthermore, the features, operations, or characteristics described in the specification may be combined in any suitable manner to form various embodiments. Also, the various steps or actions in the method descriptions may be transposed or transposed in order, as will be apparent to one of ordinary skill in the art. Thus, the various sequences in the specification and drawings are for the purpose of describing certain embodiments only and are not intended to imply a required sequence unless otherwise indicated where such sequence must be followed.
The application provides a high-precision map data crowdsourcing method based on space-time similarity, which comprises the following steps of:
s100: the vehicle collects video data and extracts information to form six-element group for each video
Figure GDA0003638965560000061
S200: calculating spatial similarity Ssec and temporal similarity Tsec, and executing steps S310-S380 when calculating the spatial similarity Ssec; when the temporal similarity Tsem is calculated, steps S410-S420 are performed.
The method for calculating the spatial similarity Ssem comprises the following steps:
s310: metadata f for two different video frames 1 And f 2 Simplified representation f 1 =(l 11 ) And f 2 =(l 22 ) The effective range of the camera and the visual field size of the camera lens are fixed as r and
Figure GDA0003638965560000062
. Each having an angular range theta 1 And theta 2 。δ l Is represented by 1 And l 2 Distance between, δ θ Indicates a difference in direction of
Figure GDA0003638965560000063
S320: decomposing rigid motion of two video metadata into translation and rotation and respectively calculating the similarity of the two motions;
s330: a rotated coverage intersection is computed, defining the coverage intersection as the similarity between two metadata.
Figure GDA0003638965560000064
S340; calculating the translational coverage intersection, and calculating from the vertical direction and the horizontal direction;
s350: similarity Ssem in vertical direction Can be obtained by the following formula
Figure GDA0003638965560000065
S360: the similarity in the horizontal direction can be calculated by the following equation:
Figure GDA0003638965560000066
wherein
Figure GDA0003638965560000067
S370: the translational similarity is obtained by a weighting equation:
Figure GDA0003638965560000071
S380:f 1 and f 2 The similarity between them can be obtained by the equation: ssem (f1,f2) =Ssem R ×Ssem T
The calculation of the temporal similarity Tsem comprises the following steps:
s410: the time metric relationship correlation is calculated as follows:
Figure GDA0003638965560000072
Figure GDA0003638965560000073
s420: calculating from functional expressions
Figure GDA0003638965560000074
Calculating the spatio-temporal similarity according to the temporal similarity Tsem and the spatial similarity Ssem comprises the following steps:
s510: calculating integral of space-time similarity measurement relation, i.e. area S along lj change curve
Figure GDA0003638965560000075
S520: the topological relation of the space-time similarity can be calculated and can be decomposed into the operation of the topological and orientation relation on each frame to directly obtain Ssem (f1,f2) And then correlated with the time similarity, i.e. Stsem TA(f1,f2) =Tsem×Ssem (f1,f2)
S530: the calculation of the total spatio-temporal similarity may be obtained
Figure GDA0003638965560000076
In the formula: stsem is the total spatio-temporal similarity; w is a group of ts ,W tta Are respectively space and time associated weight values and satisfy W ts +W tta =1, the size of the similarity of two video metadata ranges between 0 and 1.
The method of the application uploads the required video according to the space-time similarity, the bandwidth environment and the coverage required by the low points. Due to the limitations of local storage capacity, bandwidth and the like, the video clip collected by the data collection vehicle has a fixed time length limitation, and the longest available video length T is provided by the application md The length of the acquisition period is set, and when the cloud server issues a task each time, all video metadata in the period participate in operation to upload. Because the cloud server also has a requirement on the coverage degree of each place, for some places, one-time collection may not meet the published requirement, and if the coverage degree selected in the current period is not enough, what influence is caused on the selection in the next period needs to be considered.
The maximum diversity is regarded as the minimum redundancy, the video is selected on the basis of space-time similarity, the diversity of the finally uploaded video is guaranteed to be maximum, if the requirement cannot be met by one-time uploading, multiple rounds of acquisition are needed, and the requirement of coverage is finally met by multiple selections.
The maximum problem of diversity based on spatio-temporal similarity is described as follows:
suppose that the number d of the car participates in the collection 1 ,d 2 ,...,d i The bandwidth allocated by the cloud is B, and the coverage requirement is S cov A group of video segments collected in a cycle
Figure GDA0003638965560000081
Each section V i All have a start time ts (V) i ) End time te (V) i ) And size s (V) i ). Given an upload time limit t, each device can upload its video in V form, i.e., s (V), within the time limit i ) Not more than B, and U V i The space-time similarity is minimum, namely the coverage diversity is maximum.
The biggest problem of diversity based on spatio-temporal similarity is the NP-hard problem.
We reduce the maximum problem of diversity based on spatio-temporal similarity to a known NP-hard problem, namely the 0-1 knapsack problem. In the 0-1 knapsack problem, there is a set of items, each with a value and weight. Given a weight-constrained backpack, the problem is to find a subset whose total weight does not exceed the weight constraint and whose total value is maximized, and for any instance of the 0-1 backpack problem, we can construct an example as follows. We randomly select a mobile device d i Wherein the network bandwidth limit B is set as the weight limit of the backpack and the number of videos n uploaded by the mobile device i Is set as the number of items. We set the initial spatio-temporal similarity Stsem to 0, for the j-th segment V j Its sum of similarity to the selected video metadata ∑ Stsem (V) j ) Set to the value of the jth item, V is a subset of items that satisfy the weight constraint and minimize the total value of the items. Therefore, V is also a solution to the 0-1 knapsack problem.
For the approximate knapsack problem, we can solve with an algorithm based on dynamic programming. Specifically, the algorithm first randomly selects video metadata for a vehicle as the object of the first selection queue. Once video metadata is selected, it will not be deleted, nor will it continue to be selected at a later time. And selecting videos corresponding to the metadata one by one according to the similarity between the new video metadata and the metadata in the selected video metadata set. If no more videos can be selected (i.e., either all videos have been selected or the bandwidth cap has been reached), the process stops.
And (3) defining dynamic planning: in the solution problem, various possible local solutions are listed for each step of decision, then according to a certain judgment condition, the local solutions which can not obtain the optimal solution certainly are abandoned, and the optimal solution is screened in each step to ensure that the global optimal solution is obtained.
The 0-1 knapsack problem in this context can be seen as a special integer programming problem:
Figure GDA0003638965560000091
wherein:
Figure GDA0003638965560000092
suppose that
Figure GDA0003638965560000093
The backpack has a capacity of B, and the selectable articles are
Figure GDA0003638965560000094
Time 0-1 optimum for knapsack problem. While
Figure GDA0003638965560000095
Can select the article as
Figure GDA0003638965560000096
Time 0-1 backpack problem. At this time:
Figure GDA0003638965560000097
extrapolating back this recursion, deducing from the putting in of the ith video, if no video is put in, the value of the backpack is not changed, still m (V) i B), if put in, the backpack capacity is reduced to (B) i -w i s(V i ) Value increased to m (V) i-1 ,B-w i s(V i ))+Stsem(V i );
Figure GDA0003638965560000098
If the most primitive recursion method is used, the time complexity of the algorithm is
Figure GDA0003638965560000099
Wherein n is i For all videoIn the case of crowd-sourced acquisition of data, the size of the number of videos is very large, so the time complexity is correspondingly very large, and many subproblems repeat calculations, which causes considerable redundancy. The algorithm in the text simplifies recursion, adopts a bottom-up solving mode to compare the values obtained by selection and non-selection, only compares the result with the previous result, and simplifies the time complexity into O (n) i * B) Wherein n is i For all video quantities, B is the allocated bandwidth size.
For the convenience of subsequent operation, the space-time similarity of different metadata is converted into the coverage U of the metadata cov The corresponding calculation is as follows:
defined by the video metadata, the area covered by a frame of video can be defined by the camera's effective range r in the metadata j And field of view of camera lens
Figure GDA0003638965560000101
Determining, from the formula:
Figure GDA0003638965560000102
the area covered by the complete video is
Figure GDA0003638965560000103
Wherein l is a camera l on the collecting vehicle j The trajectory of the run. The effective area covered by two different metadata is calculated by
Figure GDA0003638965560000104
If the coverage area of the video selected this time is less than S cov Temporarily keeping the data in the edge server, waiting for the next acquisition, recalculating the video coverage, continuing the process until the coverage requirement is met, and uploading the video to the cloudAnd a server.
The above steps and flows are only for illustrating the technical idea of the present invention, and are not intended to limit the present invention, and any modification and improvement made on the technical scheme, technical idea, introduction method proposed by the present invention are within the protection scope of the present invention.
In summary, the present invention uses a new method to define temporal and spatial similarities, and defines the spatio-temporal similarity between different video metadata based on the definition of the two. The spatiotemporal similarity is computed using only definitions of metadata, without using video content features.
Correspondingly, the application also provides a high-precision map data crowdsourcing device based on space-time similarity, which comprises:
an acquisition module: the method comprises the steps of collecting video data through crowdsourcing, and extracting video metadata;
a time module: for calculating the time similarity between metadata;
a space module: the method is used for calculating the spatial similarity among the metadata;
a calculation module: the video processing device is used for calculating the space-time similarity among the video metadata according to the time and space similarities;
an uploading module: the method is used for uploading the required video according to the space-time similarity, the bandwidth environment and the coverage required by the low points.
The application also provides a high-precision map data crowdsourcing device based on space-time similarity, which comprises:
a memory for storing a program;
a processor for implementing the above method by executing the program stored in the memory.
Accordingly, the present application provides a computer-readable storage medium, characterized by comprising a program executable by a processor to implement the method as described above.
Those skilled in the art will appreciate that all or part of the functions of the methods in the above embodiments may be implemented by hardware, or may be implemented by a computer program. When all or part of the functions of the above embodiments are implemented by a computer program, the program may be stored in a computer-readable storage medium, and the storage medium may include: a read only memory, a random access memory, a magnetic disk, an optical disk, a hard disk, etc., and the program is executed by a computer to realize the above functions. For example, the program may be stored in a memory of the device, and when the program in the memory is executed by the processor, all or part of the functions described above may be implemented. In addition, when all or part of the functions in the above embodiments are implemented by a computer program, the program may be stored in a storage medium such as a server, another computer, a magnetic disk, an optical disk, a flash disk, or a removable hard disk, and may be downloaded or copied to a memory of a local device, or may be version-updated in a system of the local device, and when the program in the memory is executed by a processor, all or part of the functions in the above embodiments may be implemented.
The present invention has been described in terms of specific examples, which are provided to aid understanding of the invention and are not intended to be limiting. For a person skilled in the art to which the invention pertains, several simple deductions, modifications or substitutions may be made according to the idea of the invention.

Claims (6)

1. A high-precision map data crowdsourcing method based on space-time similarity is characterized by comprising the following steps:
crowdsourcing collecting video data and extracting video metadata;
calculating the time similarity among the metadata;
calculating the spatial similarity among the metadata;
calculating the space-time similarity between the video metadata according to the time and space similarity;
uploading the required video according to the space-time similarity, the bandwidth environment and the coverage required by the low point;
the step of calculating the spatial similarity between the metadata comprises the following steps:
simplifying the representation of two different video frame metadata;
decomposing the rigid motion of the two simplified video metadata into translation and rotation and respectively calculating the similarity of the two motions;
calculating a rotated coverage intersection and a translated coverage intersection;
obtaining the similarity between the two metadata according to the rotated coverage intersection and the translated coverage intersection;
calculating from the vertical direction and the horizontal direction to respectively obtain the similarity in the vertical direction and the similarity in the horizontal direction; carrying out weighted calculation on the similarity in the vertical direction and the similarity in the horizontal direction to obtain translation similarity;
two different video frame metadata f 1 、f 2 Simplified representation f 1 =(l 11 ) And f 2 =(l 22 ),
Wherein, theta 1 And theta 2 Representing the angular extent of two different video frame metadata, l 1 And l 2 Representing the shooting locations of two different video frame metadata;
the rotated coverage intersection is represented as:
Figure FDA0003988870140000011
wherein,
Figure FDA0003988870140000012
is the field of view of the camera lens, phi represents the field of view size of the camera lens,
Figure FDA0003988870140000013
a view difference representing metadata of two different video frames;
the coverage intersection of the translations is represented as:
Figure FDA0003988870140000014
wherein, ssem Is prepared by reacting withThe similarity of the vertical direction of the lens,
Figure FDA0003988870140000021
r denotes the effective range of the camera, δ l Is represented by 1 And l 2 In between the distance between the first and second electrodes is less than the predetermined distance,
Figure FDA0003988870140000022
Ssem || for similarity in the horizontal direction, δ θ =min(|θ 21 |,2π-|θ 21 |),δ θ A direction difference representing metadata of two different video frames;
similarity Ssem in horizontal direction || Calculated from the following equation:
Figure FDA0003988870140000023
when f is 2 In a translational direction of
Figure FDA0003988870140000024
And a translation distance of delta l Then, the translational similarity is obtained through a weighting equation:
Figure FDA0003988870140000025
two different video frame metadata f 1 And f 2 The spatial similarity between them is expressed as:
Ssem(f1,f2)=Ssem R(f1,f2) +Ssem T(f1,f2)
2. the method of claim 1, wherein the calculating of the temporal similarity between the metadata step comprises:
calculating the time measurement relation correlation degree:
Figure FDA0003988870140000026
calculating the time similarity:
Figure FDA0003988870140000027
3. the method of claim 1, wherein the calculating the spatiotemporal similarity between the video metadata comprises: calculating the measurement relation of the space-time similarity and calculating the topological relation of the space-time similarity,
the metric relation for calculating the space-time similarity is to calculate the integral of the area S along the lj change curve:
Figure FDA0003988870140000028
the topological relation for calculating the space-time similarity is to obtain the space similarity Ssem between two different video frame metadata by decomposing the operation of the topological relation and the orientation relation on each frame (f1,f2) And then, associating with the time similarity to obtain a topological relation of the space-time similarity:
Stsem TA(f1,f2) =Tsem×Ssem (f1,f2)
the spatio-temporal similarity is expressed as:
Stsem=W ts Stsem s(f1,f2) +W tta Stsem TA(f1,f2)
in the formula: stsem is the total spatio-temporal similarity; w is a group of ts ,W tta Are respectively space and time associated weight values, and satisfy W ts +W tta =1, the size of the similarity of two video metadata ranges between 0 and 1.
4. A high-precision map data crowdsourcing device based on space-time similarity, wherein the method of any one of claims 1-3 is applied, and comprises the following steps:
an acquisition module: the method comprises the steps of collecting video data in a crowdsourcing mode and extracting video metadata;
a time module: for calculating the time similarity between metadata;
a space module: the spatial similarity between metadata is calculated;
a calculation module: the video processing device is used for calculating the space-time similarity among the video metadata according to the time and space similarities;
an uploading module: the method is used for uploading the required video according to the space-time similarity, the bandwidth environment and the coverage required by the low points.
5. A high-precision map data crowdsourcing device based on space-time similarity is characterized by comprising the following components:
a memory for storing a program;
a processor for implementing the method of any one of claims 1-3 by executing a program stored by the memory.
6. A computer-readable storage medium, characterized by comprising a program executable by a processor to implement the method of any one of claims 1-3.
CN202011227747.2A 2020-11-06 2020-11-06 High-precision map data crowdsourcing method and device based on space-time similarity Active CN112380305B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011227747.2A CN112380305B (en) 2020-11-06 2020-11-06 High-precision map data crowdsourcing method and device based on space-time similarity

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011227747.2A CN112380305B (en) 2020-11-06 2020-11-06 High-precision map data crowdsourcing method and device based on space-time similarity

Publications (2)

Publication Number Publication Date
CN112380305A CN112380305A (en) 2021-02-19
CN112380305B true CN112380305B (en) 2023-01-17

Family

ID=74579502

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011227747.2A Active CN112380305B (en) 2020-11-06 2020-11-06 High-precision map data crowdsourcing method and device based on space-time similarity

Country Status (1)

Country Link
CN (1) CN112380305B (en)

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108847121A (en) * 2018-07-04 2018-11-20 深圳地平线机器人科技有限公司 The method and apparatus for constructing high-precision map
CN111488421B (en) * 2020-04-27 2024-04-16 立得空间信息技术股份有限公司 Data fusion method of traditional map and high-precision map

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
地理空间数据本质特征语义相关度计算模型;赵红伟等;《地理研究》;20160131;第60-61、63-65页 *
基于时空相似性的跨域目标关联分析方法;成磊峰等;《电讯技术》;20180917(第04期);第62-66页 *
视觉群智感知应用综述;翟书颖等;《计算机科学》;20190630;第46卷(第6A期);第11-12、14页 *
面向无人驾驶的边缘高精地图服务;唐洁等;《中兴通讯技术》;20190605;第25卷(第3期);第58-59、62-57页 *

Also Published As

Publication number Publication date
CN112380305A (en) 2021-02-19

Similar Documents

Publication Publication Date Title
Hausler et al. Multi-process fusion: Visual place recognition using multiple image processing methods
CN110322446B (en) Domain self-adaptive semantic segmentation method based on similarity space alignment
Sun et al. Fine-grained vehicle type classification using lightweight convolutional neural network with feature optimization and joint learning strategy
US11670071B2 (en) Fine-grained image recognition
Bendali-Braham et al. Recent trends in crowd analysis: A review
CN106909924B (en) Remote sensing image rapid retrieval method based on depth significance
Michieli et al. Adversarial learning and self-teaching techniques for domain adaptation in semantic segmentation
US20160125272A1 (en) Object recognizer and detector for two-dimensional images using bayesian network based classifier
JP2019511020A (en) Method and system for estimating arrival time
JP2010529529A (en) Specific subject detection device, learning device and learning method thereof
Reinders et al. Learning convolutional neural networks for object detection with very little training data
CN110781262A (en) Semantic map construction method based on visual SLAM
dos Santos Rosa et al. Sparse-to-continuous: Enhancing monocular depth estimation using occupancy maps
Araslanov et al. Actor-critic instance segmentation
CN111797970A (en) Method and apparatus for training neural network
CN115071762A (en) Pedestrian trajectory prediction method, model and storage medium oriented to urban scene
Sikirić et al. Traffic scene classification on a representation budget
Wang et al. Small vehicle classification in the wild using generative adversarial network
Radwan Leveraging sparse and dense features for reliable state estimation in urban environments
CN114550091A (en) Unsupervised pedestrian re-identification method and unsupervised pedestrian re-identification device based on local features
Wang et al. Cross-domain learning using optimized pseudo labels: toward adaptive car detection in different weather conditions and urban cities
CN112380305B (en) High-precision map data crowdsourcing method and device based on space-time similarity
Jia et al. Novel hybrid neural network for dense depth estimation using on-board monocular images
Tanner et al. Large-scale outdoor scene reconstruction and correction with vision
Zhang et al. Boosting the speed of real-time multi-object trackers

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant