WO2020048487A1 - Image data processing method and system - Google Patents

Image data processing method and system Download PDF

Info

Publication number
WO2020048487A1
WO2020048487A1 PCT/CN2019/104415 CN2019104415W WO2020048487A1 WO 2020048487 A1 WO2020048487 A1 WO 2020048487A1 CN 2019104415 W CN2019104415 W CN 2019104415W WO 2020048487 A1 WO2020048487 A1 WO 2020048487A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
layer
lane line
line
traffic
Prior art date
Application number
PCT/CN2019/104415
Other languages
French (fr)
Chinese (zh)
Inventor
李友增
李国镇
李佩伦
赵震
Original Assignee
北京嘀嘀无限科技发展有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京嘀嘀无限科技发展有限公司 filed Critical 北京嘀嘀无限科技发展有限公司
Publication of WO2020048487A1 publication Critical patent/WO2020048487A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/588Recognition of the road, e.g. of lane markings; Recognition of the vehicle driving pattern in relation to the road
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
    • G06V20/582Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of traffic signs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/09Recognition of logos

Definitions

  • the present application relates to the field of intelligent assisted driving or unmanned driving, and in particular, to an image data processing method and system.
  • Cars have become more and more important means of transportation in people's lives. In order to better improve the safety and comfort of drivers during driving, people have made many achievements in car intelligence.
  • the lane detection of vehicle-mounted equipment is mainly used in unmanned driving and lane keeping systems. It mainly detects lane lanes in front of the vehicle based on the images collected by the camera, and the system can automatically adjust the steering wheel steering when the vehicle deviates. Return the vehicle to the center of the lane.
  • lane detection of vehicle equipment is gradually applied to the construction of high-precision maps. Therefore, it is necessary to propose an image data processing method and system to improve the accuracy of road traffic marking recognition.
  • Embodiments of the present application provide an image data processing method, system, device, and computer-readable storage medium. Specifically include the following methods.
  • the present application discloses a method for processing image data.
  • the method includes obtaining an image including a road; obtaining at least one layer of the image based on the image and a trained traffic marking recognition model; the layer including at least one traffic marking in the image, and Reflecting the area of the at least one traffic line in the overhead image; processing the at least one layer to obtain position coordinates of the at least one traffic line in the image.
  • obtaining at least one layer of the image based on the image and the trained traffic marking recognition model further includes: obtaining multiple sub-images based on the image; the multiple Each of the sub-images includes a part of the image, and at least two sub-images collectively include a part of the image; for each sub-image, based on the sub-image and the trained traffic marking recognition model, Obtain at least one layer of the sub-image; at least one layer of the sub-image includes at least a portion of at least one traffic line in the sub-image, and at least a portion of the at least one traffic line reflects that A region in the sub-image; and based on at least one layer of at least two sub-images, at least one layer of the image is jointly determined.
  • the trained traffic marking recognition model is a trained MASK-RCNN model.
  • the image size of the layer is the same as the image size of the overhead image, and the layer is a binary image.
  • the processing the at least one layer to obtain the position coordinates of the at least one traffic line in the image further includes: performing an erosion treatment and an expansion treatment on the at least one layer. Or a combination of one or more of the smoothing processes.
  • the image includes a top-down image of a road.
  • the obtaining a top-view image including a road further includes: obtaining a road video taken by a vehicle-mounted device; obtaining multiple images based on the road video; obtaining image data of a same position in each image; and stitching the Image data of the same position in a plurality of images is used to obtain a top-view image of the road.
  • the image data of the same position in each image includes image data of at least one line that is the same in each image.
  • the layer also reflects a category of the at least one traffic line.
  • the layer responsive to the image is a lane line layer
  • processing the at least one layer to obtain the position coordinates of the at least one traffic marking in the image further includes: based on The pixel values of the lane line layer and the pixel points in the lane line layer determine the position coordinates of the left edge point and the right edge point of at least one lane line in the lane line layer; based on the at least one Position coordinates of a left edge point and a right edge point of a lane line in the lane line layer, determining position coordinates of a center line of the at least one lane line in the lane line layer; based on a preset lane line A width and a position coordinate of a center line of the at least one lane line determine a position coordinate of the at least one lane line in the image.
  • the layer in response to the image is a lane line layer
  • processing the at least one layer to obtain the position coordinates of the at least one traffic marking in the image further includes: judging Whether the distance in the horizontal direction between the end of at least one lane line and the end of another lane line in the lane line layer is smaller than the first threshold, and whether the distance in the vertical direction is smaller than the second threshold; When the distance between the end point of at least one lane line and the end point of another lane line in the horizontal direction is less than the first threshold value, and the distance in the vertical direction is less than the second threshold value, the The certain end point is spliced with the certain end point of the other lane line.
  • the layer in response to the image is a lane line layer
  • processing the at least one layer to obtain the position coordinates of the at least one traffic marking in the image further includes: judging Whether the length of a lane line in the lane line layer is less than a third threshold; and in response to the length of the certain lane line being less than a third threshold, removing the lane line from the image.
  • the processing the at least one layer to obtain position coordinates of the at least one traffic line in the image further includes: based on the at least one layer and pixels in the layer The pixel value of the point, determining the maximum coordinate value and the minimum coordinate value of the traffic marking line in the horizontal direction in the layer, and the maximum coordinate value and the minimum coordinate value in the vertical direction; based on the traffic marking line in the horizontal direction The maximum coordinate value and the minimum coordinate value on the vertical direction and the maximum coordinate value and the minimum coordinate value in the vertical direction determine the position coordinates of the traffic marking in the overhead image.
  • the present application discloses an image data processing system.
  • the system includes at least one memory for storing computer instructions; at least one processor in communication with the memory, wherein when the at least one processor executes the computer instructions, the at least one processor causes the system to execute : Obtaining an image containing a road; obtaining at least one layer of the image based on the image and a trained traffic marking recognition model; the layer including at least one traffic marking in the image, and reflecting the The area of the at least one traffic line in the image is described; the at least one layer is processed to obtain position coordinates of the at least one traffic line in the image.
  • the present application discloses an image data processing system.
  • the system includes: an image acquisition module for acquiring an image including a road; a layer acquisition module for acquiring at least one layer of the image based on the image and a trained traffic marking recognition model; the The layer includes at least one traffic line in the image; a layer processing module is configured to process the at least one layer to obtain position coordinates of the at least one traffic line in the image.
  • the layer acquisition module further includes: a sub-image acquisition unit configured to acquire multiple sub-images based on the image; the multiple sub-images each include a part of the image, and at least two The sub-images collectively include a portion of the overhead image; the sub-layer acquisition unit is configured to obtain at least at least one of the sub-images for each sub-image based on the sub-image and the trained traffic marking recognition model One layer; at least one layer of the sub-image includes at least a portion of at least one traffic line in the sub-image, and an area in the sub-image that reflects at least a portion of the at least one traffic line ; The joint determination unit is configured to jointly determine at least one layer of the images based on at least one layer of the at least two sub-images.
  • the trained traffic marking recognition model is a trained MASK-RCNN model.
  • the image size of the layer is the same as the image size of the image, and the layer is a binary image.
  • the layer processing module is further configured to perform one or more of an etching process, an expansion process, or a smoothing process on the at least one layer.
  • the image includes a top-down image of a road.
  • the image acquisition module further includes: a video acquisition unit for acquiring a road video taken by the vehicle-mounted device; an image acquisition unit for acquiring multiple images based on the road video; an image data extraction unit for acquiring each image The image data at the same position in the image; the image data stitching unit is used to stitch the image data at the same position in the multiple images to obtain a top view image of the road.
  • the image data of the same position in each image includes image data of at least one line that is the same in each image.
  • the layer also reflects a category of the at least one traffic line.
  • the layer in response to the overhead image is a lane line layer
  • the layer processing module is further configured to: based on the pixel values of the lane line layer and the pixel values in the lane line layer, Determining position coordinates of the left edge point and the right edge point of at least one lane line in the lane line layer; based on the positions of the left edge point and the right edge point of the at least one lane line in the lane line layer Coordinates, determining position coordinates of a center line of the at least one lane line in the lane line layer; determining the at least one based on a preset width of the lane line and position coordinates of the center line of the at least one lane line The position coordinates of the lane line in the image.
  • the layer in response to the overhead image is a lane line layer
  • the layer processing module is further configured to determine a certain end point of at least one lane line and a certain lane line of another lane line in the lane line layer. Whether the distance of one endpoint in the horizontal direction is smaller than the first threshold and whether the distance in the vertical direction is smaller than the second threshold; in response to a certain endpoint of the at least one lane line and a certain endpoint of the other lane line being horizontal When the distance in the direction is less than the first threshold and the distance in the vertical direction is less than the second threshold, the certain end point of the at least one road and the certain end point of the other lane line are stitched .
  • the layer in response to the overhead image is a lane line layer
  • the layer processing module is further configured to: determine whether a length of a lane line in the lane line layer is less than a third threshold; The length of a certain lane line is less than the third threshold, and the lane line is removed from the image.
  • the layer processing module is further configured to determine, based on the at least one layer and pixel values of pixels in the layer, a maximum coordinate value of a traffic marking line in the layer in a horizontal direction And minimum coordinate values, and the maximum and minimum coordinate values in the vertical direction; based on the maximum and minimum coordinate values in the horizontal direction of the traffic markings, and the maximum and minimum coordinate values in the vertical direction The coordinate value determines a position coordinate of the traffic marking line in the image.
  • the present application discloses a computer-readable storage medium.
  • the storage medium stores computer instructions, and when the computer reads the computer instructions in the storage medium, the computer executes the image data processing method.
  • FIG. 1 is a schematic diagram of an on-demand service system according to some embodiments of the present application.
  • FIG. 2 is a block diagram of an exemplary computing device for a dedicated system for implementing the technical solutions of the present application.
  • FIG. 3 is a block diagram of an exemplary mobile device of a dedicated system for implementing the technical solution of the present application.
  • FIG. 4 is a flowchart of an exemplary image data processing method according to some embodiments of the present application.
  • FIG. 5 is a flowchart of a method for acquiring a top-view image of a road according to some embodiments of the present application.
  • FIG. 6 is a top view image of an exemplary road according to some embodiments of the present application.
  • FIG. 7 is a flowchart of a method for obtaining at least one layer of an image according to some embodiments of the present application.
  • FIG. 8 is an original image before processing a lane line layer according to some embodiments of the present application.
  • FIG. 9 is a result diagram of the lane line layer being etched according to some embodiments of the present application.
  • FIG. 10 is a result diagram of the lane line layer after being etched and expanded after being etched according to some embodiments of the present application.
  • FIG. 11 is a method for determining a position coordinate of a lane line in a top-view image according to some embodiments of the present application.
  • FIG. 12 is a result diagram after a center line of a lane line is determined according to some embodiments of the present application.
  • FIG. 13 is a result diagram of extracting a center lane line center line after the lane line layer is eroded and expanded according to some embodiments of the present application.
  • FIG. 14 is a flowchart of an exemplary method for processing a multi-segment lane line according to some embodiments of the present application.
  • FIG. 15 is a flowchart of a method for determining position coordinates of at least one traffic line in a top-view image according to some embodiments of the present application.
  • FIG. 16 is a schematic diagram of processing a triangle landmark according to some embodiments of the present application.
  • FIG. 17 is a block diagram of an image data processing apparatus according to some embodiments of the present application.
  • FIG. 18 is a block diagram of a top-view image acquisition 1710 according to some embodiments of the present application.
  • FIG. 19 is a block diagram of a layer acquisition module 1720 according to some embodiments of the present application.
  • a flowchart is used in the present application to explain the operations performed by the system according to the embodiments of the present application. It should be understood that the preceding or following operations are not necessarily performed precisely in sequence. Instead, the various steps can be processed in reverse order or simultaneously. At the same time, you can add other operations to these processes, or remove a step or steps from these processes.
  • the embodiments of the present application can be applied to different transportation systems.
  • Different transportation systems include, but are not limited to, one or a combination of several types of land, ocean, aviation, and aerospace.
  • taxis, special cars, downwind cars, buses, driving cars, trains, trains, high-speed rail, driverless vehicles, delivery / courier delivery, etc. apply a management and / or distribution transportation system.
  • This application can accurately detect road images and determine the position coordinates of traffic markings in the images, so as to apply to the construction of high-precision maps.
  • the present application may also provide services for unmanned driving, for example, to achieve high-precision navigation based on the results of the processing of this application, so that unmanned vehicles are driven according to traffic markings.
  • Application scenarios of different embodiments of the present application include, but are not limited to, one or a combination of a webpage, a browser plug-in, a client, a custom system, an enterprise internal analysis system, an artificial intelligence robot, and the like. It should be understood that the application scenarios of the system and method of the present application are merely some examples or embodiments of the present application. For those of ordinary skill in the art, without paying creative labor, they can also refer to these drawings. Apply this application to other similar scenarios.
  • the traffic markings may include at least two types of markings.
  • the first type is lane lines, such as lane lines 601 and 602 shown in FIG. 6, and the second type is non-lane line markings (also known as landmarks or landmark lines), such as U-turns shown in FIG. 6 Landmark 603 and go straight landmark 604.
  • the lane line may include a two-way two-lane road surface center line to separate the opposite traffic flow, indicating that under the principle of ensuring safety, vehicles are allowed to cross overtaking, usually instructing motorists to drive to the right, and the line is a yellow dotted line.
  • the lane line may include a lane dividing line to separate the traffic flow in the same direction, which means that under the principle of ensuring safety, vehicles are allowed to overtake the line or change lanes, and the line is white.
  • the lane line may include a lane edge line, which is used to indicate the edge of a motor vehicle lane or to divide a boundary between a motor vehicle and a non-motor vehicle lane.
  • the line is a solid white line.
  • the lane line may include a highway distance confirmation marking line to provide a reference for the driver of the vehicle to maintain a safe driving distance.
  • the line is a thick white parallel solid line.
  • Lane lanes may include overtaking lanes, meaning that vehicles are strictly prohibited from overtaking or overpassing lanes, and are used to divide roads with two or more motorways in the up and down direction without a central divider. For example, the center Yellow double solid line, center yellow solid line, three lane marking line, lane changing is forbidden. Lane lines can include diversion lines, indicating that the vehicle must travel on a prescribed route, and must not run on or over the line.
  • the line is white.
  • the landmark line can include a left turn to-be-turned area line, which is used to indicate that a turning vehicle can enter the to-be-turned area during a straight ahead period and wait for a left turn. The left turn period ends. Vehicles are prohibited from staying in the area to be turned.
  • the line is a white dotted line.
  • the landmark line may include a left turn guide line, which indicates a boundary between a left turn motorized vehicle and a non-motorized vehicle. Motor vehicles travel to the left of the line, and non-motor vehicles travel to the right of the line.
  • the lines are white dashed lines.
  • the landmark line may include a pedestrian crossing line, which indicates that a pedestrian is allowed to cross the lane, and the line is a white parallel thick solid line (orthogonal, oblique).
  • the landmarks can include highway entrance and exit markings, which are used to provide safe intersections for vehicles entering or exiting the ramp, and to reduce collisions with protruding edges, including horizontal markings at entrances and exits, and markings in triangular areas.
  • the landmark line may include a parking space line indicating the location where the vehicle is parked, and the line is a solid white line.
  • the landmark line can include a harbour-style stop line, which indicates that public buses pass through a dedicated separation approach and stop location.
  • the landmark line may include a toll island mark, indicating the location of the toll island, and providing clear markings for vehicles entering the toll lane.
  • the landmark line may include a guide arrow, which indicates the driving direction of the vehicle, and is mainly used in the guide lane at the intersection, near the exit ramp and to guide channelized traffic.
  • the landmark line may include a road surface text line, which is a mark that indicates or restricts a vehicle's travel using road surface text, for example, a maximum speed, a large car, a small car, a passing lane.
  • the landmark line may include a forbidden line, for example, a roadside parking line is prohibited, and a line indicating parking on the roadside is prohibited.
  • the landmark line may include a stop line, for example, a stop line at a signal light intersection, which is a solid white line, indicating a parking position where the vehicle is waiting to be released.
  • the landmark line can include a parking yield line, which means that vehicles must stop at this intersection or slow down to give way to the main road.
  • the landmark line can include a deceleration yield line, which means that vehicles must decelerate or stop at this intersection, allowing vehicles on the main road to go ahead.
  • the landmark line may include a non-motor vehicle no-car zone marking line to notify the cyclist of the prohibited entry area at the intersection.
  • the landmark line may include a mesh line to inform the driver that temporary parking is not allowed at the intersection.
  • the landmark line can include a center circle to distinguish between large and small turning vehicles.
  • the landmark line may include a dedicated lane line to indicate a dedicated lane restricted to a certain vehicle type, and other vehicles and pedestrians may not enter.
  • the landmark line may include a U-turn prohibition mark.
  • the landmark line may include a warning line and a lane width gradient line to warn the driver of the vehicle that the road width is reduced or the number of lanes is reduced. Drive with caution and prohibit overtaking.
  • Landmark lines can include deceleration lines to warn drivers of vehicles that they should slow down.
  • Passenger”, “Passenger User”, “Passenger Device”, “Driver”, “Driver User”, “Driver Device”, “Client”, “Client Device”, “ “Client user”, “user”, “user terminal”, etc. are interchangeable and refer to the party who needs or subscribes to the service, which can be an individual or a tool.
  • the "user” described in this application may be a party that requires or subscribes to a service, or a party that provides or assists in providing a service.
  • FIG. 1 is a schematic diagram of an on-demand service system according to some embodiments of the present application.
  • the image data processing system 100 may include a server 110, a network 120, a user terminal 130, a vehicle-mounted device 140, and a memory 150.
  • the server 110 may be local or remote.
  • the server 110 may process information and / or data.
  • the server 110 may be a system for analyzing and processing the collected information to generate an analysis result.
  • the server 110 may be a terminal device, a server, or a server group.
  • the server group may be centralized, such as a data center.
  • the server group may also be distributed, such as a distributed system.
  • the server 110 may be implemented on a cloud platform.
  • the cloud platform may include a private cloud, a public cloud, a hybrid cloud, a community cloud, a distributed cloud, an internal cloud, a multi-layer cloud, or the like, or any combination thereof.
  • the server 110 may execute on the computing device 200 described in FIG. 2 that includes one or more components.
  • the server 110 may include a processing engine 112.
  • the processing engine 112 may process information and / or data to perform one or more functions described in this application.
  • the processing engine 112 may acquire an image containing a road.
  • the processing engine 112 may obtain a trained traffic marking recognition model.
  • the processing engine 112 may obtain at least one layer of the overhead image based on the image and the trained traffic landmark recognition model; the layer includes at least one traffic landmark in the overhead image, and at least one An area of the at least one traffic line in the overhead image is reflected.
  • the processing engine 112 may process the at least one layer to obtain position coordinates of the at least one traffic marking line in the overhead image.
  • the processing engine 112 may include one or more processing engines (eg, a single-core processing engine or a multi-core processor).
  • the processing engine 112 may include one or more hardware processors, such as a central processing unit (CPU), an application-specific integrated circuit (ASIC), an application-specific instruction set processor (ASIP), an image processing unit (GPU), a physical Computing Processing Unit (PPU), Digital Signal Processor (DSP), Field Programmable Gate Array (FPGA), Programmable Logic Device (PLD), Controller, Microcontroller Unit, Reduced Instruction Set Computer (RISC), Microprocessor Device or any combination thereof.
  • CPU central processing unit
  • ASIC application-specific integrated circuit
  • ASIP application-specific instruction set processor
  • GPU graphics processing unit
  • PPU physical Computing Processing Unit
  • DSP Digital Signal Processor
  • FPGA Field Programmable Gate Array
  • PLD Programmable Logic Device
  • Controller Microcontroller Unit, Reduced Instruction Set Computer (RISC), Microprocessor Device or any combination thereof.
  • RISC Reduced Instruction Set Computer
  • the user terminal 130 may be a passenger terminal or a driver terminal, and may also refer to an individual, a tool, or other entity that issues a service order.
  • the user terminal 130 may take a road image through a camera carried by the user terminal 130 and upload the road image to the server 110.
  • the user terminal 130 includes, but is not limited to, one or more of a mobile device 130-1, a built-in vehicle 130-2, a notebook computer 130-3, a desktop computer 130-4, and the like.
  • the mobile device 130-1 may include a wearable device, a smart mobile device, a virtual reality device, an augmented reality device, or the like, or any combination thereof.
  • the wearable device may include a bracelet, footwear, glasses, helmet, watch, clothing, backpack, smart accessories, etc. or any combination thereof.
  • the mobile device may include a mobile phone, a personal digital assistant (PDA), a gaming device, a navigation device, a point of sale (POS) device, a laptop computer, a desktop computer, etc., or any combination thereof.
  • the virtual reality device and / or the augmented reality device may include a virtual reality helmet, virtual reality glasses, a virtual reality eye mask, an augmented reality helmet, an augmented reality glasses, an augmented reality eye mask, etc., or any combination thereof.
  • the virtual reality device and / or augmented reality device may include Google Glass TM , RiftCon TM , Fragments TM , Gear VR TM, and the like.
  • the client 130 may process information and / or data.
  • the user terminal can be used to analyze and process the collected information to generate a system analysis result.
  • the user end 130 may perform preprocessing such as filtering, denoising, and correcting distortion on the acquired image.
  • the client 130 may also receive the processing result of the image by the server 110 and output the processing result to the user.
  • the in-vehicle device 140 may be a mobile device 140-1, a navigation device 140-2, a notebook computer 140-3, and a driving recorder 140-4.
  • the in-vehicle device 140 may include, but is not limited to, a driving recorder, a vehicle camera, a vehicle camera, a vehicle recorder, a vehicle recording device, a driving recorder, a car rear-view mirror, a driving recorder, a driving recorder, and the like having a camera function.
  • the in-vehicle device 140 may capture a road image to complete a road image acquisition process.
  • the server 110 may acquire the above-mentioned road image and process it.
  • the server 110 may directly access and access data information stored in the memory 150, and may also directly access and access information of the client 130 through the network 120.
  • the memory 150 may generally refer to a device having a storage function.
  • the memory 150 is mainly used to store data collected from the user terminal 130 and various data generated during the operation of the image data processing system 100.
  • the storage device 150 may store images acquired from the user terminal 130 and / or the in-vehicle device 140.
  • the storage device 150 may store instructions executed by the processing engine 112 or used for image data processing.
  • the storage device 150 may store training samples for training a traffic marking recognition model. Specifically, the training sample may be multiple original images and their label information.
  • the original image may include a top-view image of the road or an image in a non-top view of the road, and the original image may also include a processed top-view image of the image in a non-top view of the road.
  • the annotation information reflects the traffic signs in each original image The area of the line, and / or the category.
  • the memory 150 may be local or remote.
  • the connection or communication between the system database and other modules of the system can be wired or wireless.
  • the storage device 150 may include mass storage, removable memory, volatile read-write memory, read-only memory (ROM), etc., or any combination thereof.
  • Exemplary mass storage may include magnetic disks, optical disks, solid-state disks, and the like.
  • Exemplary removable memories may include flash drives, floppy disks, optical disks, memory cards, compact disks, magnetic tapes, and the like.
  • An exemplary volatile read-only memory may include random access memory (RAM).
  • Exemplary RAMs may include dynamic RAM (DRAM), double-rate synchronous dynamic RAM (DDR SDRAM), static RAM (SRAM), gate fluid RAM (T-RAM), zero-capacity RAM (Z-RAM), and the like.
  • Exemplary ROMs may include mask ROM (MROM), programmable ROM (PROM), erasable programmable ROM (EPROM), electronically erasable programmable ROM (EEPROM), compact disc ROM (CD-ROM), and digital General-purpose disk ROM, etc.
  • the storage device 150 may be implemented on a cloud platform.
  • the cloud platform may include a private cloud, a public cloud, a hybrid cloud, a community cloud, a distributed cloud, an internal cloud, a multi-layer cloud, or the like, or any combination thereof.
  • the storage device 150 may be connected to the network 120 to communicate with one or more components in the image data processing system 100 (eg, the server 110, the in-vehicle device 140, etc.). One or more components in the image data processing system 100 may access data or instructions stored in the storage device 150 via the network 120. In some embodiments, the storage device 150 may be directly connected to or in communication with one or more components in the image data processing system 100 (eg, the server 110, the in-vehicle device 140). In some embodiments, the storage device 150 may be part of the server 110.
  • the network 120 may provide a channel for information and / or data exchange.
  • one or more components in the image data processing system 100 may send information to other components in the image data processing system 100 through the network 120 And / or data.
  • the processing engine 112 may obtain a top-view image of the road from the storage device 150 and / or the vehicle-mounted device 140 through the network 120.
  • the network 120 may be a single network or a combination of multiple networks.
  • the network 120 may include, but is not limited to, a cable network, a wired network, a fiber optic network, a telecommunications network, an internal network, the Internet, a local area network (LAN), a wide area network (WAN), a wireless local area network (WLAN), and a metropolitan area network (MAN). , Public switched telephone network (PSTN), Bluetooth network, ZigBee network, near field communication (NFC) network, etc. or any combination thereof.
  • PSTN Public switched telephone network
  • Bluetooth network ZigBee network
  • NFC near field communication
  • the network 120 may include one or more network access points, such as a wired or wireless access point, a base station (such as 120-1, 120-2), or a network exchange point. Through the above access points, the data source is connected to the network 120 and passed through Send information over the network.
  • the image data processing system 100 is provided for illustrative purposes only and is not intended to limit the scope of the application.
  • the image data processing system 100 may further include a database, an information source, and the like.
  • the image data processing system 100 may be implemented on other devices to implement similar or different functions. However, these amendments and changes will not depart from the scope of this application.
  • FIG. 2 is a block diagram of an exemplary computing device 200 for a dedicated system for implementing the technical solutions of the present application.
  • the computing device 200 may include a processor 210, a memory 220, an input / output interface 230, and a communication communication port 240.
  • the processor 210 may execute calculation instructions (program code) and perform functions of the image data processing system 100 described in this application.
  • the calculation instructions may include programs, objects, components, data structures, procedures, modules, and functions (the functions refer to specific functions described in this application).
  • the processor 210 may process image or text data obtained from any other component of the image data processing system 100.
  • the processor 210 may include a microcontroller, a microprocessor, a reduced instruction set computer (RISC), an application specific integrated circuit (ASIC), an application specific instruction set processor (ASIP), a central processing unit (CPU) , Graphics processing unit (GPU), physical processing unit (PPU), microcontroller unit, digital signal processor (DSP), field programmable gate array (FPGA), advanced RISC machine (ARM), programmable logic device, and capable of Any circuit, processor, etc., or any combination thereof that performs one or more functions.
  • RISC reduced instruction set computer
  • ASIC application specific integrated circuit
  • ASIP application specific instruction set processor
  • CPU central processing unit
  • GPU Graphics processing unit
  • PPU physical processing unit
  • DSP digital signal processor
  • FPGA field programmable gate array
  • ARM advanced RISC machine
  • the memory 220 may store data / information obtained from any other components of the image data processing system 100.
  • the memory 220 may include mass storage, removable memory, volatile read and write memory, read-only memory (ROM), or the like, or any combination thereof.
  • Exemplary mass storage may include magnetic disks, optical disks, solid-state drives, and the like.
  • Removable storage may include flash drives, floppy disks, optical disks, memory cards, compact disks, and magnetic tapes.
  • Volatile read and write memory may include random access memory (RAM).
  • RAM can include dynamic RAM (DRAM), double-rate synchronous dynamic RAM (DDR, SDRAM), static RAM (SRAM), thyristor RAM (T-RAM), zero capacitance (Z-RAM), and so on.
  • ROM can include mask ROM (MROM), programmable ROM (PROM), erasable programmable ROM (PEROM), electrically erasable programmable ROM (EEPROM), compact disc ROM (CD-ROM), and digital versatile disc ROM Wait.
  • MROM mask ROM
  • PROM programmable ROM
  • PEROM erasable programmable ROM
  • EEPROM electrically erasable programmable ROM
  • CD-ROM compact disc ROM
  • digital versatile disc ROM Wait digital versatile disc ROM Wait.
  • An input / output (I / O) interface 230 may be used to input or output signals, data, or information.
  • the input / output interface 230 may enable a user to communicate with the image data processing system 100.
  • the input / output interface 230 may include an input device and an output device.
  • Exemplary input devices may include a keyboard, a mouse, a touch screen, a microphone, and the like, or any combination thereof.
  • Exemplary output devices may include display devices, speakers, printers, projectors, etc., or any combination thereof.
  • Exemplary display devices may include a liquid crystal display (LCD), a light emitting diode (LED) based display, a flat panel display, a curved display, a television device, a cathode ray tube (CRT), etc., or any combination thereof.
  • the communication port 240 may be connected to a network for data communication.
  • the connection may be a wired connection, a wireless connection, or a combination of both.
  • the wired connection may include a cable, an optical cable or a telephone line, or any combination thereof.
  • the wireless connection may include Bluetooth, Wi-Fi, WiMax, WLAN, ZigBee, a mobile network (for example, 3G, 4G, or 5G, etc.), etc., or any combination thereof.
  • the communication port 240 may be a standardized port, such as RS232, RS485, and the like.
  • the communication port 240 may be a specially designed port.
  • FIG. 3 is a block diagram of an exemplary mobile device 300 for implementing a dedicated system of the technical solution of the present application. According to some embodiments of the present application, the user terminal 130 may be implemented thereon.
  • the mobile device 300 may include a communication platform 310, a display 320, a graphics processor (GPU) 330, a central processing unit (CPU) 340, an input / output interface 350, a memory 360, a memory 370, and the like.
  • the CPU 340 may include an interface circuit and a processing circuit similar to the processor 210.
  • any other suitable components including but not limited to a system bus or controller (not shown), may also be included within the mobile device 300.
  • the operating system 361 eg, iOS, Android, Windows Phone, etc.
  • application programs 362 may be loaded from the memory 370 into the memory 360 for execution by the CPU 340.
  • the application programs 362 may include a browser or an application program for receiving imaging, graphics processing, audio, or other related information from the image data processing system 100.
  • the user's interaction with the information flow can be achieved via the input / output 350 and provided to the processing engine 112 and / or other components of the image data processing system 100 via the network 120.
  • a computing device or mobile device may be used as a hardware platform for one or more components described in this application.
  • the hardware elements, operating systems, and programming languages of these computers or mobile devices are conventional in nature, and those skilled in the art can adapt these technologies to the on-demand service system described in this application after being familiar with these technologies.
  • a computer with user interface elements can be used to implement a personal computer (PC) or other type of workstation or terminal device, and the computer can also act as a server if properly programmed. It is believed that those skilled in the art may also be familiar with the structure, programs, or general operations of this type of computer equipment. Therefore, no additional explanation is described for the drawings.
  • FIG. 4 is a flowchart of an exemplary image data processing method according to some embodiments of the present application.
  • the image data processing method 400 is executed by a device having processing and computing capabilities, such as the server 110 or the computing device 200.
  • Step 401 Acquire an image including a road.
  • the image of the road may include a top view image of the road.
  • a bird's-eye view image of a road may refer to a road image taken from an aerial view of the road.
  • the road image may also include road images taken from other angles.
  • Obtaining road images can be obtained in many ways. For example, road images can be obtained by aerial photography, road images can also be obtained by surveillance cameras installed on the road, and road images collected by the user or on-board equipment (for example, driving recorder) can be used to obtain road images. image.
  • the road image captured by the user terminal or the in-vehicle device includes a road image not viewed from another angle.
  • a user sits in a vehicle to obtain a road image through a user terminal
  • another example is a road image captured by a vehicle-mounted device placed in a rearview mirror of a vehicle through a vehicle's front windshield.
  • the road image at a non-top view angle may be processed and converted into a corresponding top view image.
  • Step 402 Obtain a trained traffic marking recognition model.
  • the server 110 may obtain the trained traffic marking recognition model through the network 120.
  • the trained traffic marking recognition model can identify the type of traffic markings in the road image and / or the location in the image, such as the position of lane lines in the road image, and the locations of various landmarks in the road image. .
  • the trained traffic marking recognition model can process a road image, output the processed image, and label different types of traffic markings in different forms on the image.
  • the user can circle each traffic line with a dotted frame and / or mark different categories of traffic lines with text.
  • a trained traffic marking recognition model can output an image and use different pixel values to represent different types of traffic markings on the image. For example, the user number "1" represents a lane line, and the number "2 "Means a landmark (for example, a sidewalk).
  • the preferred trained traffic lane recognition model in this embodiment is a trained MASK-RCNN model.
  • the input when training the MASK-RCNN model is multiple original images and their label information.
  • the original image may include a top-view image of a road or an image taken at a non-top-down angle of the road, and the original image may also include a processed top-view image of an image taken at a non-top-down angle of the road.
  • the labeling information may be layer information labeled with various traffic marking categories.
  • the layer can be an image the same size as the original image.
  • one original image corresponds to multiple layers, and each layer contains a type or a traffic line.
  • one original image may correspond to only one layer, and the layer includes all traffic markings in the original image.
  • the pixel values of the layer may be limited to a limited number of values.
  • the layer may be a binary image, and the pixel value of the traffic markings in the layer is "1".
  • the pixel values of the remaining pixels are "0"; for example, by marking different types of traffic markings on the original image, different pixel values can be used to represent different types of traffic markings on the same image (for example, the pixel value " 1 "represents lane lines, pixel value” 2 "represents sidewalk, pixel value” 3 "represents straight), and pixel values of the remaining pixels are" 4 ".
  • the server 110 can obtain layer information corresponding to different types of traffic markings. Based on the original image and the layers of its various traffic markings, the MASK-RCNN model learns to make the predicted output traffic markings and the various traffic markings consistent. MASK-RCNN model.
  • the trained traffic marking recognition model may also be a trained traffic marking model based on a convolutional neural network, a deep learning-based traffic marking model, a MATLAB-based traffic marking recognition model, Traffic recognition model based on intelligent cognition, traffic feature recognition model based on shape features, etc. Different types of original images (such as road top images, road images taken at non-top angles, or processed top images) are input into the MASK-RCNN model.
  • the trained MASK-RCNN model can The road image is processed to obtain one or more layers of the corresponding image, and the layer may include traffic markings obtained by model recognition.
  • Step 403 Obtain at least one layer of the image based on the image and the trained traffic marking recognition model; the layer includes at least one traffic marking in the image, and reflects at least the An area of at least one traffic line in the image.
  • the server 110 can obtain at least one layer of the image. Specifically, at least one layer corresponding to different image types may be output according to an input image type (for example, a road top image, an image in a non-top view angle of the road, or a processed top image). In some embodiments, the output layers have the same traffic markings at the same location in the image corresponding to the traffic markings.
  • a layer can include one or more traffic lines. In some embodiments, a layer may include only one type of traffic line (for example, only lane lines), or a plurality of types of traffic line (for example, lane lines and landmarks).
  • a layer may include only one type of landmark (for example, a sidewalk), or may include multiple types of landmarks (for example, a sidewalk, a U-turn, and a slow down).
  • one layer may only include lane lines, and each other layer may include only one type of landmark (for example, one layer may only include U-turns, one layer may include only sidewalks, and one layer may include only right turns).
  • the layer can be a binary image (the pixel value of each pixel is 0 or 1).
  • a pixel value of 1 indicates that the point is a traffic marking, and a pixel value of 0 indicates that the point is non-traffic. Marking.
  • the layer can also reflect the category of at least one traffic line.
  • a layer only contains information about lane lines, which is also called a lane line layer, where a pixel value of 1 indicates that the point is a lane line.
  • a certain layer only includes information of a landmark (for example, a sidewalk), which is also referred to as a landmark layer, and a pixel value of 1 indicates that the point is a landmark.
  • the layer may have more types of pixel values.
  • the MASK-RCNN model can output mask layer information of different types of traffic markings, and can obtain pixel-level location information of traffic markings.
  • the mask layer information of each type of traffic markings is two-dimensional 0/1 raster data (the grid can include one or more pixels), and the pixel value is 1 in the area with traffic markings, and the value is 0 otherwise.
  • the traffic markings on the output layer may be framed by a wire frame, and the area within the wire frame is the area of the traffic markings in the image.
  • the wireframe can be one or more combinations of rectangular, circular, oval, polygonal, or irregular wireframes. Different wireframes can have different colors.
  • Step 404 Process the at least one layer to obtain position coordinates of the at least one traffic line in the image.
  • the layer reflects the position of at least one traffic line in the image.
  • the output layers have the same traffic markings at the same location in the image corresponding to the traffic markings.
  • the server 110 can obtain more precise position coordinates of at least one traffic line in the image.
  • the processing methods include, but are not limited to, image denoising, image transformation, edge detection, edge thinning, image analysis, image compression, image enhancement, image blur processing, image interpolation, binary transformation, threshold transformation, Fourier transformation, Discrete cosine transform, etc.
  • image denoising image transformation, edge detection, edge thinning, image analysis, image compression, image enhancement, image blur processing, image interpolation, binary transformation, threshold transformation, Fourier transformation, Discrete cosine transform, etc.
  • FIG. 5 illustrates steps in which the server 110 acquires a top-view image of a road in some embodiments.
  • Step 501 Obtain a road video taken by a vehicle-mounted device.
  • Vehicle-mounted devices include, but are not limited to, vehicle-mounted recorders, vehicle cameras, vehicle cameras, vehicle-mounted recorders, vehicle-mounted recording devices, vehicle-mounted recorders, connected car rear-view mirrors, vehicle-mounted recorders, and vehicle-recording devices.
  • the server 110 may obtain a road video shot by the vehicle-mounted device through the network 120.
  • Step 502 Acquire multiple images based on the road video.
  • the server 110 may divide the road video into multiple frames of data to obtain multiple images. For example, the server 110 may extract several consecutive frames of images, obtain an image every one frame, or obtain an image at multiple frames.
  • Step 503 Obtain image data at the same position in each image.
  • the server 110 may extract image data of the same line in each image. In still other embodiments, the server 110 may extract image data of the same multiple lines in each image. For example, the server 110 may extract three lines (ie, the first line, the second line, and the third line) of image data from the top to the bottom of each image. In some embodiments, the server 110 may extract one or more lines of data with less distortion from each image. For example, you can extract one or more lines of data at the top or bottom of the image, or one or more lines with less distortion in the middle part of the image. In some embodiments, the server 110 may extract image data from other fixed positions in each image. The fixed position is not limited to image data of a fixed line or multiple lines, but may also be local image data, such as a fixed line. Or some columns in the image data in multiple rows.
  • Step 504 stitch the image data of the same position among the multiple images to obtain a top-view image of the road.
  • the server 110 can stitch the image data of the same line in each image in chronological order to form the final overhead image of the road.
  • the server 110 may form a plurality of top-view images of roads by stitching based on extracting image data of the same multiple lines in each image. For example, after the server 110 extracts three rows of image data from line 100, line 320, and line 480 from the top of each image, the server 110 may stitch the image data based on the line 100 image data of each image to form a line.
  • a bird's-eye view image, the image data of the 320th line of each image are stitched to form a second bird's-eye view image, and the image data of the 480th line of each image are stitched to form a third bird's-eye view image.
  • the server 110 After the server 110 obtains the three overhead images, it can finally determine a suitable overhead image by comparing the quality (e.g., sharpness) of the three images.
  • on-vehicle equipment is installed in the car to obtain road images through the front windshield.
  • the roads in the images may appear near wide, narrow and narrow visual effects. After the process shown in FIG. 5, the roads have different images.
  • the widths on the lines can be approximately equal, which can be understood as converting the road images acquired by the in-vehicle equipment into images with similar shooting effects from a bird's-eye view.
  • FIG. 7 shows that in other embodiments, the server 110 obtains at least one layer of the image based on the image and the trained traffic marking recognition model, including:
  • Step 701 Obtain multiple sub-images based on the images; the multiple sub-images each include a portion of the image, and at least two sub-images collectively include a portion of the image.
  • the image may include a top-view image of the road, and may also include an image in a non-top-view perspective of the road or an image in a non-top-view perspective after processing.
  • FIG. 4 and related descriptions, and details are not described herein.
  • Step 702 For each sub-image, obtain at least one layer of the sub-image based on the sub-image and the trained traffic marking recognition model; the layer includes at least one of the sub-images At least a portion of a traffic line, and at least an area in the sub-image that reflects at least a portion of the at least one traffic line.
  • Step 703 Jointly determine at least one layer of the images based on at least one layer of the at least two sub-images.
  • the server 110 may capture sub-images of multiple images through a window smaller than the image.
  • the pixel size of an image is 1024 ⁇ 1024
  • the size of the window is 64 ⁇ 64.
  • the image can be captured to obtain 1024 ⁇ 1024 sub-images each having a size of 64 ⁇ 64.
  • the pixel value of the uncovered image in the window is replaced with a pixel value of 0.
  • At least one layer of each sub-image can be identified by inputting each sub-image into a trained traffic marking recognition model (for example, a MASK-RCNN model), where each layer can include only one type of traffic Marking.
  • a trained traffic marking recognition model for example, a MASK-RCNN model
  • the pixel value of the pixel point of the layer is 0 or 1, where 1 indicates that the pixel point is the traffic marking.
  • a pixel value of 1 indicates that the point is a lane line.
  • a pixel value of 1 at a pixel point indicates that the point is a sidewalk.
  • the server 110 It can be determined that the pixel value of the point in the layer of the traffic marking corresponding to the image is finally 1.
  • a traffic marking recognition model for example, a MASK-RCNN model
  • 100 sub-image lane line layers are output.
  • the 100 sub-image lane line layers include the same A point in the image. .
  • the server 110 may determine that the pixel value of point A in the image lane line layer is 1, otherwise, Determine that the final pixel value of point A in the final lane line layer is 0. Similarly, the server 110 may determine the final pixel value of each pixel point in each layer of the image. Finally, the server 110 may determine at least one layer of the image. For example, the server 110 may finally determine a lane line layer and a layer corresponding to each landmark (for example, a sidewalk layer, a U-turn layer, a left turn layer, etc.).
  • the sub-region of the image can be identified multiple times, and the results of the multiple recognition can be combined (e.g., fused) to determine the pixel value at the corresponding pixel position in the image layer, which can effectively reduce the recognition error and improve the image recognition. Precision.
  • the server 110 may include, but is not limited to, performing an etching process, an expansion process, or a smoothing process on the processing of the layer.
  • Figure 8 is the original image before processing the lane line layer.
  • White represents the lane line and black represents the non-lane line.
  • the server 110 may perform morphological processing on the lane line layer, mainly using a relatively small rectangular kernel (for example, kernel (2, 2), kernel (3, 3), Core (5,5)).
  • Corrosion operation means that when the pixel value of a pixel in the core is not all 1, the pixel value of the pixel in the core is changed from 1 to 0 (that is, from white to black). When the value is all 1, no change is made. This operation can largely filter out invalid (or isolated) noise.
  • FIG. 9 is a result diagram of the lane line layer after being corroded. As can be seen from FIG. 9, after the erosion, the lane line after the erosion is narrower than the original lane line.
  • FIG. 10 is a result diagram of the lane line layer after being corroded and then subjected to expansion processing.
  • Part of the expansion can offset the loss of precision introduced in the corrosion operation, and part of it can solve the situation of lane line disconnection caused by partial occlusion (as shown in Figure 901, the two lane lines are disconnected. After expansion, Figure 10 is obtained. The two lane lines in the middle 1001 section are connected).
  • FIG. 11 illustrates a method in which the server 110 may perform the following operations on the lane line layer to obtain the position coordinates of the image in response to the image layer being a lane line layer.
  • the server 110 may perform one or more of an etching process, an expansion process, or a smoothing process on the lane line layer.
  • Step 1101 Determine the position coordinates of the left edge point and the right edge point of at least one lane line in the lane line layer based on the pixel values of the pixel points in the lane line layer and the lane line layer.
  • the server 110 may perform left and right lane line edge extraction on the layer from top to bottom. For example, the server 110 may read one line of data at a time to determine the first left edge point and the first right edge point of the first lane line in the line, where the first left edge point is a pixel value from 0 to 1 The first pixel point (from black to white), and the first right edge point is the first pixel point whose pixel value changes from 1 to 0 (from white to black). The server 110 continues to scan the line of data to obtain the second left edge point and the second right edge point.
  • the server 110 scans data of different lines, and compares the position of a pixel point (for example, the left edge point) of the lane line of the line with the position of a pixel point (for example, the left edge point) of the lane line of the previous line. If it is less than a certain threshold, it is regarded as the same lane line, so that the second line of information is combined with the first line of information; and so on, combining multiple lines of information, and finally connecting to obtain multiple lane lines.
  • Step 1102 Determine position coordinates of a center line of the at least one lane line in the lane line layer based on position coordinates of a left edge point and a right edge point of the at least one lane line in the lane line layer.
  • the server 110 may perform a certain operation on the coordinates of the left edge point and the right edge point of each lane line in the lane line layer, and use the operation result as the coordinates of the center line point of the lane line. For example, the coordinates of the left edge point and the right edge point of each lane line in the lane line layer are averaged, and the result is taken as the coordinate position of the lane line center line.
  • the server 110 forms the center line of the lane line by connecting the midpoint of the left edge point and the right edge point of the lane line.
  • FIG. 12 is a result diagram after the center line of the lane line is determined for the lane line layer.
  • Step 1103 Determine position coordinates of the at least one lane line in the image based on a preset width of the lane line and position coordinates of a center line of the at least one lane line.
  • the server 110 may determine the pixel width of the lane line in the lane line layer based on a preset width of the lane line (for example, 10 mm, 20 mm, 30 mm, 50 mm, 100 mm, 200 mm, etc.). For example, the server 110 may determine that the pixel width of the lane lines in the lane line layer is 30 pixels. In some embodiments, the server 110 may widen a certain pixel area from the center line of the lane line to the left and right to determine the position coordinates of the lane line in the image. For example, the server 110 may use the center line of the lane line as a reference to widen 15 pixels to the left and right sides to determine the position coordinates of the lane line in the image.
  • a preset width of the lane line for example, 10 mm, 20 mm, 30 mm, 50 mm, 100 mm, 200 mm, etc.
  • the server 110 may determine that the pixel width of the lane lines in the lane line layer is 30 pixels.
  • the server 110 may determine whether the distance in the horizontal direction between the end of the centerline of one lane line and the head of the centerline of another lane line is less than the first threshold, and whether the distance in the vertical direction is Less than the second threshold.
  • the first threshold may be set to 15 pixels
  • the second threshold may be set to 100 pixels.
  • the server 110 may The ends of the at least one lane line and the head of the center line of the other lane line are connected by a linear interpolation method. Further, the server 110 may determine a position coordinate of the at least one lane line in the bird's-eye view image by widening a number of pixels to the left and right sides by using the spliced center line as a reference.
  • the server 110 may perform the morphological processing (including corrosion and expansion) on the lane line layer and then perform the operation in FIG. 11, as shown in FIG. 13, extracting the center lane after the lane line layer is corroded and expanded. Result graph of line centerline.
  • a detected lane line may be divided into several segments due to pedestrian occlusion or other occlusion reasons.
  • the server 110 may perform splicing, filtering, and smoothing on the lane line.
  • FIG. 14 shows a flowchart of an exemplary method for processing a multi-segment lane line.
  • Step 1401 it is determined whether a distance between a certain end point of at least one lane line and another end point of another lane line in the lane line layer in a horizontal direction is less than a first threshold value, and whether the distance in a vertical direction is less than a second threshold value .
  • the server 110 may determine the position coordinates of the left edge point and the right edge point of a certain end of each lane line in the lane line layer by the method of step 1101. Based on the position coordinates of the left edge point and the right edge point of a certain end of each lane line in the lane line layer, the server 110 may determine the endpoint of each lane line in the lane line layer. It should be noted that the end is the area where the lane line starts or ends in the vertical direction of the lane line layer, such as the top row of pixels or the bottom row of pixels. Width, the end can include multiple pixels in a row. The end point is a point in the end area, which is used to characterize the end of the lane line.
  • the server 110 may determine a distance in a horizontal direction (also referred to as a horizontal distance) and a distance in a vertical direction (also referred to as a vertical direction) between a certain end point of each lane line and another end point of another lane line in the lane line layer. Straight distance).
  • the server 110 may calculate a horizontal distance and a vertical distance between one end point of a certain lane line end (which may include multiple pixels) and one end point of another lane line head end (which may include multiple pixels). For example, the server 110 may calculate a horizontal distance and a vertical distance between the end of the centerline of a certain lane line and the beginning of the centerline of another lane line.
  • the server 110 may calculate a horizontal distance and a vertical distance between a left edge point at the end of a certain lane line and a left edge point at the head of another lane line. Further, it is determined whether a distance between a certain end point of at least one lane line and another end point of another lane line in the lane line layer in a horizontal direction is less than a first threshold value, and whether a distance in a vertical direction is less than a second threshold value.
  • step 1402 in response to a distance between a certain end point of the at least one lane line and a certain end point of the other lane line in the horizontal direction being smaller than the first threshold and the distance in the vertical direction being smaller than the second threshold, The one end point of the at least one lane line and the one end point of the other lane line are spliced.
  • the first threshold may be set to 15 pixels
  • the second threshold may be set to 100 pixels.
  • the server 110 may combine pixels in a range of pixels formed by the left edge point and the right edge point of the end of the at least one lane line and the left edge point and the right edge point of the head end of the other lane line. The value becomes 1 to form a spliced segment of the two lane lines.
  • the server 110 may connect the end of the center line of the at least one lane line and the head of the center line of the other lane line by a linear interpolation method. Further, the server 110 can widen a number of pixels to the left and right sides by using the spliced center line as a reference to realize the splicing of two lane lines.
  • the server 110 may connect the left edge point of the end of the at least one lane line and the left edge point of the head of the other lane line by a linear interpolation method. Further, the server 110 can realize the splicing of the two lane lines by using the left edge line after splicing as a reference to be several pixels wide to the right.
  • the server 110 may connect a right edge point at the end of the at least one lane line and a right edge point at the head of the other lane line by a linear interpolation method. Further, the server 110 may widen a number of pixels to the left by using the spliced right edge line as a reference to realize splicing of two lane lines.
  • Step 1403 Determine whether the length of a lane line in the lane line layer is less than a third threshold.
  • the server 110 may determine the positions of the head and end of a lane line in the lane line layer.
  • the server 110 may determine the length of the lane line based on the position of the head and end of the lane line in the lane line layer. Further, the server 110 may determine whether the length of a certain lane line is less than a third threshold.
  • the third threshold may be a value set in advance by the server 110, and for example, may be a distance of 100 pixels.
  • Step 1404 in response to the length of a certain lane line being less than a third threshold, removing the lane line from the image.
  • the server 110 may change the pixel value of the pixel point of the lane line in the lane line layer from 1 to 0 (that is, from white to black).
  • Step 1405 Perform window smoothing processing on the processed lane line layer.
  • the short-term bumps of the vehicle may cause a very small number of lane lines to have “spikes” after being extracted.
  • the server 110 may Line layers are smoothed. Smoothing algorithms include, but are not limited to, mean filtering, median filtering, Gaussian filtering, bilateral filtering, and so on.
  • the processing sequence of the method 1400 for splicing, filtering, and smoothing lane lines is only exemplary, and the application does not limit the processing sequence.
  • the server 110 may perform filtering processing on the lane lines before performing stitching and smoothing processing.
  • certain steps in method 1400 may be used independently.
  • steps 1401 and 1402 can be used independently to perform lane line splicing.
  • steps 1403 to 1404 can be used in combination to remove other traces or objects on the ground that interfere with traffic markings recognition.
  • 1405 can be used independently to smooth the traffic markings. Such deformations are all within the protection scope of this application.
  • the method 1500 is a flowchart of another method for processing a layer to obtain position coordinates of at least one traffic line in an image according to some embodiments of the present application.
  • the method 1500 can be used to process an image as a landmark layer.
  • the server 110 may perform the following operations on the landmark layer to obtain the position coordinates of the landmark in the image.
  • Step 1501 Based on the at least one layer and the pixel values of the pixels in the layer, determine a maximum coordinate value and a minimum coordinate value of the traffic markings in the layer in the horizontal direction and a maximum value in the vertical direction. Coordinate value and minimum coordinate value.
  • the server 110 may determine that the maximum coordinate value of the triangular landmark in the horizontal direction of the layer is x2, the minimum coordinate value is x1, and the maximum coordinate value in the vertical direction y2 and the minimum coordinate value y1.
  • Step 1502 Determine the position coordinates of the landmark in the image based on the maximum coordinate value and the minimum coordinate value in the horizontal direction of the traffic marking line and the maximum coordinate value and the minimum coordinate value in the vertical direction.
  • the server 110 may determine a rectangular area of the triangular landmark.
  • the server 110 may determine the rectangular area ABCD of the triangular landmark based on the coordinates A (x1, y1), coordinates B (x2, y2), coordinates C (x1, y2), and coordinates D (x2, y2).
  • the server 110 may frame a region containing a landmark in a rectangular frame manner, and then mark the guide line such as going straight, turning left, or a zebra crossing.
  • the server 110 can also identify the rectangular area where the landmark is located by using the method 1500.
  • the server 110 frames the area containing the landmark in a rectangular frame manner, and then marks the guide line such as going straight, turning left, or a zebra crossing. It should be noted that the method 1500 is also applicable to the processing of lane lines.
  • FIG. 17 is a block diagram of an image data processing apparatus according to some embodiments of the present application.
  • the image data processing apparatus 1700 may be implemented by the server 110.
  • the image data processing apparatus 1700 may also be referred to as an image data processing system.
  • the image data processing device 1700 includes an image acquisition module 1710, a layer acquisition module 1720, and a layer processing module 1730.
  • the image acquisition module 1710 is configured to acquire an image including a road.
  • a layer acquisition module 1720 is configured to obtain at least one layer of the image based on the image and a trained traffic marking recognition model; the layer includes at least one traffic marking in the image, and reflects at least one A region of the at least one traffic line in the image.
  • a layer processing module 1730 is configured to process the at least one layer to obtain position coordinates of the at least one traffic line in the image.
  • the trained traffic marking recognition model is a trained MASK-RCNN model.
  • the image size of the layer is the same as the image size of the image, and the layer is a binary image.
  • the layer also reflects a category of the at least one traffic line.
  • the layer processing module 1730 is further configured to perform one or more combinations of an etching process, an expansion process, or a smoothing process on the at least one layer.
  • the image includes a top-down image of the road.
  • the layer processing module 1730 is further configured to determine at least one of the lane line layers and the pixel values of the pixels in the lane line layer. Position coordinates of a left edge point and a right edge point of a lane line in the lane line layer; based on position coordinates of the left edge point and the right edge point of the at least one lane line in the lane line layer, determine Position coordinates of a center line of the at least one lane line in the lane line layer; determining a position of the at least one lane line based on a preset width of the lane line and position coordinates of the center line of the at least one lane line Position coordinates in the image.
  • the layer processing module 1730 in response to the image layer being a lane line layer, is further configured to determine whether an end point of at least one lane line and an end point of another lane line in the lane line layer are at Whether the distance in the horizontal direction is less than the first threshold value, and whether the distance in the vertical direction is less than the second threshold value; in response to a distance in the horizontal direction between an end point of the at least one lane line and an end point of the other lane line When the distance is smaller than the first threshold and the distance in the vertical direction is smaller than the second threshold, the certain end point of the at least one road and the certain end point of the other lane line are stitched.
  • the layer in response to the image is a lane line layer
  • the layer processing module 1730 is further configured to determine whether a length of a lane line in the lane line layer is less than a third threshold; in response to the certain line The length of the lane line is less than the third threshold, and the lane line is removed from the image.
  • the layer processing module 1730 is further configured to determine, based on the at least one layer and the pixel values of the pixels in the layer, the maximum coordinate value and the minimum coordinate value of the traffic markings in the layer in the horizontal direction. Coordinate values, and the maximum and minimum coordinate values in the vertical direction; based on the maximum and minimum coordinate values in the horizontal direction of the traffic markings, and the maximum and minimum coordinate values in the vertical direction To determine the position coordinates of the traffic marking in the image.
  • FIG. 18 is a block diagram of a top-view image acquisition 1710 according to some embodiments of the present application.
  • the overhead image acquisition module 1710 includes a video acquisition unit 1810, an image acquisition unit 19820, an image data extraction unit 1830, and an image data splicing unit 1840.
  • the video acquiring unit 1810 is configured to acquire a road video captured by a vehicle-mounted device.
  • the image acquisition unit 1820 is configured to acquire multiple images based on the road video.
  • the image data extraction unit 1830 is configured to acquire image data of a same position in each image.
  • the image data splicing unit 1840 is configured to splice image data of the same position among the multiple images to obtain a top-view image of the road.
  • the image data at the same position in each image includes at least one line that is the same in each image.
  • FIG. 19 is a block diagram of a layer acquisition module 1720 according to some embodiments of the present application.
  • the layer acquisition module 1720 includes a sub-image acquisition unit 1910, a sub-layer acquisition unit 1920, and a joint determination unit 1930.
  • the sub-image obtaining unit 1910 is configured to obtain multiple sub-images based on the image; the multiple sub-images each include a part of the image, and at least two sub-images collectively include a certain part of the image;
  • the sub-layer obtaining unit 1920 is configured to obtain, for each sub-image, at least one layer of the sub-image based on the sub-image and the trained traffic marking recognition model; the layer includes the sub-image At least a portion of at least one traffic line in the image, and at least a region reflecting at least a portion of the at least one traffic line in the sub-image;
  • the joint determining unit 1930 is configured to jointly determine at least one layer of the images based on at least one layer of the at least two sub-images.
  • the present application discloses an image data processing system.
  • the system includes at least one processor and at least one memory; the at least one memory is configured to store computer instructions; the at least one processor is configured to execute at least a part of the computer instructions to implement any image data processing method.
  • the present disclosure discloses a computer-readable storage medium.
  • the storage medium stores computer instructions, and when the computer instructions are executed by a processor, any image data processing method is implemented.
  • each module may be software modules implemented by computer instructions.
  • the above description of the candidate selection display and determination system and its modules is for convenience of description only, and cannot limit the application to the scope of the illustrated embodiments. It can be understood that for those skilled in the art, after understanding the principle of the system, it is possible to arbitrarily combine the various modules or form a subsystem to connect with other modules without departing from this principle. For example, each module can share a storage module, and each module can also have its own storage module. Such deformations are all within the protection scope of this application.
  • the beneficial effects that the embodiments of the present application may bring include, but are not limited to: (1) different types of traffic markings can be identified through a traffic marking recognition model; (2) splicing processing, filtering processing, and smoothing of the extracted lane lines The processing solves the problem that the traffic markings are partially occluded and discontinuous to a certain extent; (3) extracting the same position of multiple images in the road video to stitch the top view image of the road to a certain extent solves the problem of image distortion; ( 4)
  • the Mask-RCNN model can identify the pixel-level positions of different types of traffic markings in the layer, meeting the requirements for higher-precision traffic marking recognition. It should be noted that different embodiments may have different beneficial effects. In different embodiments, the possible beneficial effects may be any one or a combination of the foregoing, or any other beneficial effects that may be obtained.
  • aspects of this application can be illustrated and described through several patentable categories or situations, including any new and useful processes, machines, products, or combinations of materials, or their Any new and useful improvements. Accordingly, various aspects of the present application can be executed entirely by hardware, can be executed entirely by software (including firmware, resident software, microcode, etc.), and can also be executed by a combination of hardware and software.
  • the above hardware or software can be called “data block”, “module”, “engine”, “unit”, “component” or “system”.
  • aspects of the present application may manifest as a computer product located in one or more computer-readable media, the product including a computer-readable program code.
  • Computer storage media may contain a transmitted data signal containing a computer program code, such as on baseband or as part of a carrier wave.
  • the propagation signal may have multiple manifestations, including electromagnetic form, optical form, etc., or a suitable combination form.
  • a computer storage medium may be any computer-readable medium other than a computer-readable storage medium, which may be connected to an instruction execution system, apparatus, or device to enable communication, propagation, or transmission of a program for use.
  • Program code on a computer storage medium may be transmitted through any suitable medium, including radio, cable, fiber optic cable, RF, or similar media, or any combination of the foregoing.
  • the computer program code required for the operation of each part of this application can be written in any one or more programming languages, including object-oriented programming languages such as Java, Scala, Smalltalk, Eiffel, JADE, Emerald, C ++, C #, VB.NET, Python Etc., conventional programming languages such as C, Visual Basic, Fortran 2003, Perl, COBOL 2002, PHP, ABAP, dynamic programming languages such as Python, Ruby and Groovy, or other programming languages.
  • the program code may run entirely on the user's computer, or as a stand-alone software package on the user's computer, or partly on the user's computer, partly on a remote computer, or entirely on the remote computer or server.
  • the remote computer can be connected to the user's computer through any network form, such as a local area network (LAN) or wide area network (WAN), or to an external computer (for example, via the Internet), or in a cloud computing environment, or as a service Uses such as software as a service (SaaS).
  • LAN local area network
  • WAN wide area network
  • SaaS software as a service
  • numbers describing the number of components and attributes are used. It should be understood that, for such numbers used in the description of the embodiments, the modifiers "about”, “approximately” or “substantially” Modification. Unless stated otherwise, “about”, “approximately” or “substantially” indicates that the number allows for a variation of ⁇ 20%. Accordingly, in some embodiments, the numerical parameters used in the specification and claims are approximate values, and the approximate values may be changed according to the characteristics required by individual embodiments. In some embodiments, the numerical parameter should take the specified significant digits into account and adopt a general digits retention method. Although the numerical ranges and parameters used to confirm the breadth of the range in some embodiments of this application are approximate values, in specific embodiments, the setting of such values is as accurate as possible within the feasible range.
  • this application may be provided as a method, a system, or a computer program product. Therefore, this application may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Moreover, this application may take the form of a computer program product implemented on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, etc.) containing computer-usable program code.
  • computer-usable storage media including, but not limited to, disk storage, CD-ROM, optical storage, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Traffic Control Systems (AREA)

Abstract

Disclosed are an image data processing method and system. The method comprises: obtaining an image containing a road; obtaining at least one image layer of the image on the basis of the image and a trained traffic marking identification model, the image layer comprising at least one traffic marking in the image and a region reflecting at least one landmark in an overhead image; and processing the at least one image layer to obtain a position coordinate of the at least one traffic marking in the image. By means of the traffic marking identification model, the image layers of different categories of traffic markings can be identified, and the image layer is further processed and then can be used for constructing a high-precision map.

Description

一种图像数据处理方法及系统Image data processing method and system
优先权信息Priority information
本申请要求于2018年9月5日提交的中国申请号为201811031784.9的优先权,其全部内容通过引用的方式并入本文。This application claims priority from Chinese application number 201811031784.9, filed on September 5, 2018, the entire contents of which are incorporated herein by reference.
技术领域Technical field
本申请涉及智能辅助驾驶或无人驾驶领域,尤其涉及一种图像数据处理方法及系统。The present application relates to the field of intelligent assisted driving or unmanned driving, and in particular, to an image data processing method and system.
背景技术Background technique
汽车目前已成为人们生活中越来越重要的交通工具,为了更好提高汽车行驶过程中的安全性和驾驶员的舒适性,人们在汽车智能化做出了很多成就。通常车载设备的车道线检测主要在无人驾驶,车道保持系统等应用较多,其主要根据摄像头采集的图像检测车辆前方的车道线,并且当车辆发生跑偏时,系统可以自动调节方向盘转向,使车辆重新回到车道中央行驶。随着汽车智能化的发展,车载设备的车道线检测逐渐运用到高精度地图的构建。因此,有必要提出一种图像数据处理方法及系统以提高道路交通标线识别的精度。Cars have become more and more important means of transportation in people's lives. In order to better improve the safety and comfort of drivers during driving, people have made many achievements in car intelligence. Generally, the lane detection of vehicle-mounted equipment is mainly used in unmanned driving and lane keeping systems. It mainly detects lane lanes in front of the vehicle based on the images collected by the camera, and the system can automatically adjust the steering wheel steering when the vehicle deviates. Return the vehicle to the center of the lane. With the development of automobile intelligence, lane detection of vehicle equipment is gradually applied to the construction of high-precision maps. Therefore, it is necessary to propose an image data processing method and system to improve the accuracy of road traffic marking recognition.
发明内容Summary of the Invention
本申请的实施例提供了一种图像数据处理方法、系统、装置和计算机可读存储介质。具体包括以下几个方法。Embodiments of the present application provide an image data processing method, system, device, and computer-readable storage medium. Specifically include the following methods.
第一方面,本申请披露了一种图像数据处理方法。该方法包括获取包含道路的图像;基于所述图像和经训练的交通标线识别模型,获得所述图像的至少一个图层;所述图层包括所述图像中的至少一个交通标线,以及反映所述至少一个交通标线在所述俯视图像中的区域;对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标。In a first aspect, the present application discloses a method for processing image data. The method includes obtaining an image including a road; obtaining at least one layer of the image based on the image and a trained traffic marking recognition model; the layer including at least one traffic marking in the image, and Reflecting the area of the at least one traffic line in the overhead image; processing the at least one layer to obtain position coordinates of the at least one traffic line in the image.
在一些实施例中,所述基于所述图像和所述经训练的交通标线识别模型,获得所述图像的至少一个图层还包括:基于所述图像,获取多张子图像;所述多张子图像分别包括所述图像的一部分,且至少两张子图像共同包括所述图像中的某一部分;针对每张子图像,基于所述子图像和所述经训练的交通标线识别模型,获得所述子图像的至少一个图层;所述子图像的至少一个图层包括所述子图像中的至少一个交通标线的至少一部分,以及至少反映所述至少一个交通标线的至少一部分在所述子图像中的区域;基于至少两张子图像的至少一个图层,联合确定出所述图像的至少一个图层。In some embodiments, obtaining at least one layer of the image based on the image and the trained traffic marking recognition model further includes: obtaining multiple sub-images based on the image; the multiple Each of the sub-images includes a part of the image, and at least two sub-images collectively include a part of the image; for each sub-image, based on the sub-image and the trained traffic marking recognition model, Obtain at least one layer of the sub-image; at least one layer of the sub-image includes at least a portion of at least one traffic line in the sub-image, and at least a portion of the at least one traffic line reflects that A region in the sub-image; and based on at least one layer of at least two sub-images, at least one layer of the image is jointly determined.
在一些实施例中,所述经训练的交通标线识别模型为经训练的MASK-RCNN模型。In some embodiments, the trained traffic marking recognition model is a trained MASK-RCNN model.
在一些实施例中,所述图层的图像大小与所述俯视图像的图像大小相同,且所述图层为二值化图像。In some embodiments, the image size of the layer is the same as the image size of the overhead image, and the layer is a binary image.
在一些实施例中,所述对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标还包括:对所述至少一个图层进行腐蚀处理、膨胀处理或平滑处理中的一种或多种的组合。In some embodiments, the processing the at least one layer to obtain the position coordinates of the at least one traffic line in the image further includes: performing an erosion treatment and an expansion treatment on the at least one layer. Or a combination of one or more of the smoothing processes.
在一些实施例中,所述图像包括道路的俯视图像。In some embodiments, the image includes a top-down image of a road.
在一些实施例中,所述获取包含道路的俯视图像还包括:获取车载设备拍摄的道路视频;基于所述道路视频,获取多张图像;获取每张图像中同一位置的图像数据;拼接所述多张图像中同一位置的图像数据,获得所述道路的俯视图像。In some embodiments, the obtaining a top-view image including a road further includes: obtaining a road video taken by a vehicle-mounted device; obtaining multiple images based on the road video; obtaining image data of a same position in each image; and stitching the Image data of the same position in a plurality of images is used to obtain a top-view image of the road.
在一些实施例中,所述每张图像中同一位置的图像数据包括每张图像中相同的至少一行的图像数据。In some embodiments, the image data of the same position in each image includes image data of at least one line that is the same in each image.
在一些实施例中,所述图层还反映所述至少一个交通标线的类别。In some embodiments, the layer also reflects a category of the at least one traffic line.
在一些实施例中,响应于图像的图层为车道线图层,所述对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标还包括:基于所述车道线图层和所述车道线图层中像素点的像素值,确定至少一条车道线的左边缘点和右边缘点在所述车道线图层中的位置坐标;基于所述至少一条车道线的左边缘点和右边缘点在所述车道线图层中的位置坐标,确定所述至少一条车道线的中心线在所述车道线图层中的位置坐标;基于车道线的预设宽度以及所述至少一条车道线的中心线的位置坐标,确定所述至少一条车道线在所述图像中的位置坐标。In some embodiments, the layer responsive to the image is a lane line layer, and processing the at least one layer to obtain the position coordinates of the at least one traffic marking in the image further includes: based on The pixel values of the lane line layer and the pixel points in the lane line layer determine the position coordinates of the left edge point and the right edge point of at least one lane line in the lane line layer; based on the at least one Position coordinates of a left edge point and a right edge point of a lane line in the lane line layer, determining position coordinates of a center line of the at least one lane line in the lane line layer; based on a preset lane line A width and a position coordinate of a center line of the at least one lane line determine a position coordinate of the at least one lane line in the image.
在一些实施例中,响应于图像的图层为车道线图层,所述对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标还包括:判断车道线图层中至少一条车道线的某一端点和另一条车道线的某一端点在水平方向的距离是否小于第一阈值,并且在竖直方向上的距离是否小于第二阈值;响应于所述至少一条车道线的某一端点和另一条车道线的某一端点在水平方向的距离小于第一阈值,并且在竖直方向上的距离小于第二阈值时,则将所述至少一条道路的所述某一端点和所述另一条车道线的所述某一端点进行拼接。In some embodiments, the layer in response to the image is a lane line layer, and processing the at least one layer to obtain the position coordinates of the at least one traffic marking in the image further includes: judging Whether the distance in the horizontal direction between the end of at least one lane line and the end of another lane line in the lane line layer is smaller than the first threshold, and whether the distance in the vertical direction is smaller than the second threshold; When the distance between the end point of at least one lane line and the end point of another lane line in the horizontal direction is less than the first threshold value, and the distance in the vertical direction is less than the second threshold value, the The certain end point is spliced with the certain end point of the other lane line.
在一些实施例中,响应于图像的图层为车道线图层,所述对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标还包括:判断车道线图层中某条车道线的长度是否小于第三阈值;响应于所述某条车道线的长度小于第三阈值,将所述车道线从所述图像中去除。In some embodiments, the layer in response to the image is a lane line layer, and processing the at least one layer to obtain the position coordinates of the at least one traffic marking in the image further includes: judging Whether the length of a lane line in the lane line layer is less than a third threshold; and in response to the length of the certain lane line being less than a third threshold, removing the lane line from the image.
在一些实施例中,所述对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标还包括:基于所述至少一个图层和该图层中像素点的像素值,确定该图层中交通标线在水平方向上的最大坐标值和最小坐标值,以及在竖直方向上的最大坐标值和最小坐标值;基于所述交通标线在水平方向上的最大坐标值和最小坐标值以及在竖直方向上的最大坐标值和最小坐标值,确定该交通标线在所述俯视图像中的位置坐标。In some embodiments, the processing the at least one layer to obtain position coordinates of the at least one traffic line in the image further includes: based on the at least one layer and pixels in the layer The pixel value of the point, determining the maximum coordinate value and the minimum coordinate value of the traffic marking line in the horizontal direction in the layer, and the maximum coordinate value and the minimum coordinate value in the vertical direction; based on the traffic marking line in the horizontal direction The maximum coordinate value and the minimum coordinate value on the vertical direction and the maximum coordinate value and the minimum coordinate value in the vertical direction determine the position coordinates of the traffic marking in the overhead image.
第二方面,本申请披露了一种图像数据处理系统。该系统包括用于存储计算机指令的至少一个存储器;与所述存储器通讯的至少一个处理器,其中当所述至少一个处理器执行所述计算机指令时,所述至少一个处理器使所述系统执行:获取包含道路的图像;基于所述图像和经训练的交通标线识别模型,获得所述图像的至少一个图层;所述图层包括所述图像中的至少一个交通标线,以及反映所述至少一个交通标线在所述图像中的区域;对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标。In a second aspect, the present application discloses an image data processing system. The system includes at least one memory for storing computer instructions; at least one processor in communication with the memory, wherein when the at least one processor executes the computer instructions, the at least one processor causes the system to execute : Obtaining an image containing a road; obtaining at least one layer of the image based on the image and a trained traffic marking recognition model; the layer including at least one traffic marking in the image, and reflecting the The area of the at least one traffic line in the image is described; the at least one layer is processed to obtain position coordinates of the at least one traffic line in the image.
第三方面,本申请披露了一种图像数据处理系统。所述系统包括:图像获取模块,用于获取包含道路的图像;图层获取模块,用于基于所述图像和经训练的交通标线识别模型,获得所述图像的至少一个图层;所述图层包括所述图像中的至少一个交通标线;图层处理模块,用于对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标。In a third aspect, the present application discloses an image data processing system. The system includes: an image acquisition module for acquiring an image including a road; a layer acquisition module for acquiring at least one layer of the image based on the image and a trained traffic marking recognition model; the The layer includes at least one traffic line in the image; a layer processing module is configured to process the at least one layer to obtain position coordinates of the at least one traffic line in the image.
在一些实施例中,所述图层获取模块还包括:子图像获取单元用于基于所述图像,获取多张子图像;所述多张子图像分别包括所述图像的一部分,且至少两张子图像共同包括所述俯视图像中的某一部分;子图层获取单元用于针对每张子图像,基于所述子图像和所述经训练的交通标线识别模型,获得所述子图像的至少一个图层;所述子图像的至少一个图层包括所述子图像中的至少一个交通标线的至少一部分,以及至少反映所述至少一个交通标线的至少一部分在所述子图像中的区域;联合确定单元用于基于至少两张子图像的至少一个图层,联合确定出所述图像的至少一个图层。In some embodiments, the layer acquisition module further includes: a sub-image acquisition unit configured to acquire multiple sub-images based on the image; the multiple sub-images each include a part of the image, and at least two The sub-images collectively include a portion of the overhead image; the sub-layer acquisition unit is configured to obtain at least at least one of the sub-images for each sub-image based on the sub-image and the trained traffic marking recognition model One layer; at least one layer of the sub-image includes at least a portion of at least one traffic line in the sub-image, and an area in the sub-image that reflects at least a portion of the at least one traffic line ; The joint determination unit is configured to jointly determine at least one layer of the images based on at least one layer of the at least two sub-images.
在一些实施例中,所述经训练的交通标线识别模型为经训练的MASK-RCNN模型。In some embodiments, the trained traffic marking recognition model is a trained MASK-RCNN model.
在一些实施例中,所述图层的图像大小与所述图像的图像大小相同,且所述图层为二值化图像。In some embodiments, the image size of the layer is the same as the image size of the image, and the layer is a binary image.
在一些实施例中,图层处理模块还用于:对所述至少一个图层进行腐蚀处理、膨胀处理或平滑处理中的一种或多种处理。In some embodiments, the layer processing module is further configured to perform one or more of an etching process, an expansion process, or a smoothing process on the at least one layer.
在一些实施例中,所述图像包括道路的俯视图像。In some embodiments, the image includes a top-down image of a road.
在一些实施例中,图像获取模块还包括:视频获取单元用于获取车载设备拍摄的道路 视频;图像获取单元用于基于所述道路视频,获取多张图像;图像数据提取单元用于获取每张图像中同一位置的图像数据;图像数据拼接单元用于拼接所述多张图像中同一位置的图像数据,获得所述道路的俯视图像。In some embodiments, the image acquisition module further includes: a video acquisition unit for acquiring a road video taken by the vehicle-mounted device; an image acquisition unit for acquiring multiple images based on the road video; an image data extraction unit for acquiring each image The image data at the same position in the image; the image data stitching unit is used to stitch the image data at the same position in the multiple images to obtain a top view image of the road.
在一些实施例中,所述每张图像中同一位置的图像数据包括每张图像中相同的至少一行的图像数据。In some embodiments, the image data of the same position in each image includes image data of at least one line that is the same in each image.
在一些实施例中,所述图层还反映所述至少一个交通标线的类别。In some embodiments, the layer also reflects a category of the at least one traffic line.
在一些实施例中,响应于俯视图像的图层为车道线图层,所述图层处理模块还用于:基于所述车道线图层和所述车道线图层中像素点的像素值,确定至少一条车道线的左边缘点和右边缘点在所述车道线图层中的位置坐标;基于所述至少一条车道线的左边缘点和右边缘点在所述车道线图层中的位置坐标,确定所述至少一条车道线的中心线在所述车道线图层中的位置坐标;基于车道线的预设宽度以及所述至少一条车道线的中心线的位置坐标,确定所述至少一条车道线在所述图像中的位置坐标。In some embodiments, the layer in response to the overhead image is a lane line layer, and the layer processing module is further configured to: based on the pixel values of the lane line layer and the pixel values in the lane line layer, Determining position coordinates of the left edge point and the right edge point of at least one lane line in the lane line layer; based on the positions of the left edge point and the right edge point of the at least one lane line in the lane line layer Coordinates, determining position coordinates of a center line of the at least one lane line in the lane line layer; determining the at least one based on a preset width of the lane line and position coordinates of the center line of the at least one lane line The position coordinates of the lane line in the image.
在一些实施例中,响应于俯视图像的图层为车道线图层,所述图层处理模块还用于:判断车道线图层中至少一条车道线的某一端点和另一条车道线的某一端点在水平方向的距离是否小于第一阈值,并且在竖直方向上的距离是否小于第二阈值;响应于所述至少一条车道线的某一端点和另一条车道线的某一端点在水平方向的距离小于第一阈值,并且在竖直方向上的距离小于第二阈值时,则将所述至少一条道路的所述某一端点和所述另一条车道线的所述某一端点进行拼接。In some embodiments, the layer in response to the overhead image is a lane line layer, and the layer processing module is further configured to determine a certain end point of at least one lane line and a certain lane line of another lane line in the lane line layer. Whether the distance of one endpoint in the horizontal direction is smaller than the first threshold and whether the distance in the vertical direction is smaller than the second threshold; in response to a certain endpoint of the at least one lane line and a certain endpoint of the other lane line being horizontal When the distance in the direction is less than the first threshold and the distance in the vertical direction is less than the second threshold, the certain end point of the at least one road and the certain end point of the other lane line are stitched .
在一些实施例中,响应于俯视图像的图层为车道线图层,所述图层处理模块还用于:判断车道线图层中某条车道线的长度是否小于第三阈值;响应于所述某条车道线的长度小于第三阈值,将所述车道线从所述图像中去除。In some embodiments, the layer in response to the overhead image is a lane line layer, and the layer processing module is further configured to: determine whether a length of a lane line in the lane line layer is less than a third threshold; The length of a certain lane line is less than the third threshold, and the lane line is removed from the image.
在一些实施例中,所述图层处理模块还用于:基于所述至少一个图层和该图层中像素点的像素值,确定该图层中交通标线在水平方向上的最大坐标值和最小坐标值,以及在竖直方向上的最大坐标值和最小坐标值;基于所述交通标线在水平方向上的最大坐标值和最小坐标值以及在竖直方向上的最大坐标值和最小坐标值,确定该交通标线在所述图像中的位置坐标。In some embodiments, the layer processing module is further configured to determine, based on the at least one layer and pixel values of pixels in the layer, a maximum coordinate value of a traffic marking line in the layer in a horizontal direction And minimum coordinate values, and the maximum and minimum coordinate values in the vertical direction; based on the maximum and minimum coordinate values in the horizontal direction of the traffic markings, and the maximum and minimum coordinate values in the vertical direction The coordinate value determines a position coordinate of the traffic marking line in the image.
第四方面,本申请披露了一种计算机可读存储介质。所述存储介质存储计算机指令,当计算机读取存储介质中的计算机指令后,计算机执行所述图像数据处理方法。In a fourth aspect, the present application discloses a computer-readable storage medium. The storage medium stores computer instructions, and when the computer reads the computer instructions in the storage medium, the computer executes the image data processing method.
本申请的一部分附加特性可以在下面的描述中进行说明。通过对以下描述和相应附图的检查或者对实施例的生产或操作的了解,本申请的一部分附加特性对于本领域技术人员是 显而易见的。本申请的特性可以通过对以下描述的具体实施例的各种方面的方法、手段和组合的实践或使用得以实现和达到。Some additional features of this application can be explained in the following description. Some of the additional features of this application will be apparent to those skilled in the art from a review of the following description and corresponding drawings, or an understanding of the production or operation of the embodiments. The features of the present application can be achieved and achieved through the practice or use of the methods, means, and combinations of the various aspects of the specific embodiments described below.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
图1是根据本申请的一些实施例所示的一种按需服务系统的示意图。FIG. 1 is a schematic diagram of an on-demand service system according to some embodiments of the present application.
图2是用于实现本申请技术方案的专用系统的示例性计算设备的框图。FIG. 2 is a block diagram of an exemplary computing device for a dedicated system for implementing the technical solutions of the present application.
图3是用于实现本申请技术方案的专用系统的示例性移动设备的框图。FIG. 3 is a block diagram of an exemplary mobile device of a dedicated system for implementing the technical solution of the present application.
图4为根据本申请的一些实施例所示的示例性图像数据处理方法的流程图。FIG. 4 is a flowchart of an exemplary image data processing method according to some embodiments of the present application.
图5为根据本申请的一些实施例所示的获取道路的俯视图像的方法的流程图FIG. 5 is a flowchart of a method for acquiring a top-view image of a road according to some embodiments of the present application.
图6为根据本申请的一些实施例所示的一种示例性的道路的俯视图像。FIG. 6 is a top view image of an exemplary road according to some embodiments of the present application.
图7为根据本申请的一些实施例所示的获得图像的至少一个图层的方法的流程图。FIG. 7 is a flowchart of a method for obtaining at least one layer of an image according to some embodiments of the present application.
图8为根据本申请的一些实施例所示的对车道线图层处理前的原图。FIG. 8 is an original image before processing a lane line layer according to some embodiments of the present application.
图9为根据本申请的一些实施例所示的对车道线图层进行腐蚀后的结果图。FIG. 9 is a result diagram of the lane line layer being etched according to some embodiments of the present application.
图10为根据本申请的一些实施例所示的对车道线图层进行腐蚀后又经过膨胀处理后的结果图。FIG. 10 is a result diagram of the lane line layer after being etched and expanded after being etched according to some embodiments of the present application.
图11为根据本申请的一些实施例所示的确定车道线在俯视图像中的位置坐标的方法。FIG. 11 is a method for determining a position coordinate of a lane line in a top-view image according to some embodiments of the present application.
图12为根据本申请的一些实施例所示的确定车道线的中心线后的结果图。FIG. 12 is a result diagram after a center line of a lane line is determined according to some embodiments of the present application.
图13为根据本申请的一些实施例所示的车道线图层经过腐蚀、膨胀后提取中心车道线中心线的结果图。FIG. 13 is a result diagram of extracting a center lane line center line after the lane line layer is eroded and expanded according to some embodiments of the present application.
图14为根据本申请的一些实施例所示的一种示例性的处理多段车道线的方法的流程。FIG. 14 is a flowchart of an exemplary method for processing a multi-segment lane line according to some embodiments of the present application.
图15为根据本申请的一些实施例所示的确定至少一个交通标线在俯视图像中的位置坐标的方法的流程图。FIG. 15 is a flowchart of a method for determining position coordinates of at least one traffic line in a top-view image according to some embodiments of the present application.
图16为根据本申请的一些实施例所示的处理三角形地标的示意图。FIG. 16 is a schematic diagram of processing a triangle landmark according to some embodiments of the present application.
图17为根据本申请的一些实施例所示的一种图像数据处理装置的框图。FIG. 17 is a block diagram of an image data processing apparatus according to some embodiments of the present application.
图18为根据本申请的一些实施例所示的一种俯视图像获取1710的框图。FIG. 18 is a block diagram of a top-view image acquisition 1710 according to some embodiments of the present application.
图19为根据本申请的一些实施例所示的一种图层获取模块1720的框图。FIG. 19 is a block diagram of a layer acquisition module 1720 according to some embodiments of the present application.
具体实施方式detailed description
为了更清楚地说明本申请的实施例的技术方案,下面将对实施例描述中所需要使用的附图作简单的介绍。显而易见地,下面描述中的附图仅仅是本申请的一些示例或实施例,对于本领域的普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图将本申 请应用于其它类似情景。除非从语言环境中显而易见或另做说明,图中相同标号代表相同结构或操作。In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings used in the description of the embodiments will be briefly introduced below. Obviously, the drawings in the following description are just some examples or embodiments of the present application. For those of ordinary skill in the art, the present application can also be applied to these drawings without creative efforts. Other similar scenarios. Unless obvious from a locale or otherwise stated, the same reference numerals in the figures represent the same structure or operation.
如本申请和权利要求书中所示,除非上下文明确提示例外情形,“一”、“一个”、“一种”和/或“该”等词并非特指单数,也可包括复数。一般说来,术语“包括”与“包含”仅提示包括已明确标识的步骤和元素,而这些步骤和元素不构成一个排它性的罗列,方法或者设备也可能包含其它的步骤或元素。As shown in this application and the claims, the words "a", "an", "an" and / or "the" do not specifically refer to the singular, but may include the plural unless the context clearly indicates an exception. Generally speaking, the terms "including" and "comprising" are only meant to include clearly identified steps and elements, and these steps and elements do not constitute an exclusive list, and the method or equipment may also include other steps or elements.
虽然本申请对根据本申请的实施例的系统中的某些模块或单元做出了各种引用,然而,任何数量的不同模块或单元可以被使用并运行在客户端和/或服务器上。所述模块仅是说明性的,并且所述系统和方法的不同方面可以使用不同模块。Although this application makes various references to certain modules or units in the system according to the embodiments of this application, any number of different modules or units may be used and run on the client and / or server. The modules are merely illustrative, and different aspects of the systems and methods may use different modules.
本申请中使用了流程图用来说明根据本申请的实施例的系统所执行的操作。应当理解的是,前面或下面操作不一定按照顺序来精确地执行。相反,可以按照倒序或同时处理各种步骤。同时,也可以将其他操作添加到这些过程中,或从这些过程移除某一步或数步操作。A flowchart is used in the present application to explain the operations performed by the system according to the embodiments of the present application. It should be understood that the preceding or following operations are not necessarily performed precisely in sequence. Instead, the various steps can be processed in reverse order or simultaneously. At the same time, you can add other operations to these processes, or remove a step or steps from these processes.
本申请的实施例可以应用于不同的运输系统,不同的运输系统包括但不限于陆地、海洋、航空、航天等中的一种或几种的组合。例如,出租车、专车、顺风车、巴士、代驾、火车、动车、高铁、无人驾驶的交通工具、收/送快递等应用了管理和/或分配的运输系统。本申请可以对道路图像进行精确检测,确定图像中的交通标线位置坐标,以应用于构建高精度地图。在一些实施例中,本申请还可以为无人驾驶提供服务,例如基于本申请处理的结果实现高精度导航,使无人驾驶的交通工具按照交通标线行驶。本申请的不同实施例应用场景包括但不限于网页、浏览器插件、客户端、定制系统、企业内部分析系统、人工智能机器人等中的一种或几种的组合。应当理解的是,本申请的系统及方法的应用场景仅仅是本申请的一些示例或实施例,对于本领域的普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图将本申请应用于其它类似情景。The embodiments of the present application can be applied to different transportation systems. Different transportation systems include, but are not limited to, one or a combination of several types of land, ocean, aviation, and aerospace. For example, taxis, special cars, downwind cars, buses, driving cars, trains, trains, high-speed rail, driverless vehicles, delivery / courier delivery, etc. apply a management and / or distribution transportation system. This application can accurately detect road images and determine the position coordinates of traffic markings in the images, so as to apply to the construction of high-precision maps. In some embodiments, the present application may also provide services for unmanned driving, for example, to achieve high-precision navigation based on the results of the processing of this application, so that unmanned vehicles are driven according to traffic markings. Application scenarios of different embodiments of the present application include, but are not limited to, one or a combination of a webpage, a browser plug-in, a client, a custom system, an enterprise internal analysis system, an artificial intelligence robot, and the like. It should be understood that the application scenarios of the system and method of the present application are merely some examples or embodiments of the present application. For those of ordinary skill in the art, without paying creative labor, they can also refer to these drawings. Apply this application to other similar scenarios.
在一些实施例中,交通标线可以至少包括两类标线。第一类为车道线,如图6中所示的车道线601和车道线602,第二类为非车道线标线(也称为地标或地标标线),如图6中所示的掉头地标603和直行地标604。In some embodiments, the traffic markings may include at least two types of markings. The first type is lane lines, such as lane lines 601 and 602 shown in FIG. 6, and the second type is non-lane line markings (also known as landmarks or landmark lines), such as U-turns shown in FIG. 6 Landmark 603 and go straight landmark 604.
车道线可以包括双向两车道路面中心线,用来分隔对向行驶的交通流,表示在保证安全的原则下,准许车辆跨越超车,通常指示机动车驾驶人靠右行驶,线为黄色虚线。车道线可以包括车行道分界线,用来分隔同向行驶的交通流,表示在保证安全的原则下,准许车辆跨越线超车或变更车道行驶,线为白色。车道线可以包括车行道边缘线,用来指示机动车道的边缘或用来划分机动车与非机动车道的分界,线为白色实线。车道线可以包括高速公路车 距确认标线,用以提供车辆机动车驾驶人保持行车安全距离之参考,线为白色平行粗实线。车道线可以包括禁止超车线,表示严格禁止车辆跨线超车或压线行驶,用以划分上、下行方向各有两条或两条以上机动车道,而没有设置中央分隔带的道路,例如,中心黄色双实线、中心黄色虚实线、三车道标线、禁止变换车道线。车道线可以包括导流线,表示车辆须按规定的路线行驶,不得压线或越线行驶,线为白色。The lane line may include a two-way two-lane road surface center line to separate the opposite traffic flow, indicating that under the principle of ensuring safety, vehicles are allowed to cross overtaking, usually instructing motorists to drive to the right, and the line is a yellow dotted line. The lane line may include a lane dividing line to separate the traffic flow in the same direction, which means that under the principle of ensuring safety, vehicles are allowed to overtake the line or change lanes, and the line is white. The lane line may include a lane edge line, which is used to indicate the edge of a motor vehicle lane or to divide a boundary between a motor vehicle and a non-motor vehicle lane. The line is a solid white line. The lane line may include a highway distance confirmation marking line to provide a reference for the driver of the vehicle to maintain a safe driving distance. The line is a thick white parallel solid line. Lane lanes may include overtaking lanes, meaning that vehicles are strictly prohibited from overtaking or overpassing lanes, and are used to divide roads with two or more motorways in the up and down direction without a central divider. For example, the center Yellow double solid line, center yellow solid line, three lane marking line, lane changing is forbidden. Lane lines can include diversion lines, indicating that the vehicle must travel on a prescribed route, and must not run on or over the line. The line is white.
地标标线可以包括左转弯待转区线,用来指示转弯车辆可在直行时段进入待转区,等待左转。左转时段终止,禁止车辆在待转区停留,线为白色虚线。地标标线可以包括左转弯导向线,表示左转弯的机动车与非机动车之间的分界。机动车在线的左侧行驶,非机动车在线的右侧行驶,线为白色虚线。地标标线可以包括人行横道线,表示准许行人横穿车行道的标线,线为白色平行粗实线(正交、斜交)。地标标线可以包括高速公路出入口标线,是为了驶入或驶出匝道车辆提供安全交汇,减少与突出部缘石的碰撞的标线,包括出入口的横向标线、三角地带的标线。地标标线可以包括停车位标线,表示车辆停放的位置,线为白色实线。地标标线可以包括港湾式停靠站标线,表示公共客车通过专门的分离引道和停靠位置。地标标线可以包括收费岛标线,表示收费岛的位置,为驶入收费车道的车辆提供清晰的标记。地标标线可以包括导向箭头,表示车辆的行驶方向,主要用于交叉道口的导向车道内、出口匝道附近及对渠化交通的引导。地标标线可以包括路面文字标线,是利用路面文字,指示或限制车辆行驶的标记,例如,最高速度、大型车、小型车、超车道。地标标线可以包括禁止标线,例如,禁止路边停放车辆线,表示禁止在路边停车的标线。地标标线可以包括停止线,例如,信号灯路口的停止线,为白色实线,表示车辆等候放行的停车位置。地标标线可以包括停车让行线,表示车辆在此路口必须停车或减速让干道车辆先行。地标标线可以包括减速让行线,表示车辆在此路口必须减速或停车,让干道车辆先行。地标标线可以包括非机动车禁驶区标线,用以告示骑车人在路口禁止驶入的范围。地标标线可以包括网状线,用以告示驾驶人禁止在该交叉路口临时停车。地标标线可以包括中心圈,用以区分车辆大、小转弯车辆不得压线行驶。地标标线可以包括专用车道线,用以指示仅限于某车种行驶的专用车道,其他车辆、行人不得进入。地标标线可以包括禁止掉头标记。地标标线可以包括警告标线车行道宽度渐变标线,用以警告车辆驾驶人了解路宽缩减或车道数减少,应谨慎行车,禁止超车。地标标线可以包括减速标线,用以警告车辆驾驶人前方应减速慢行。The landmark line can include a left turn to-be-turned area line, which is used to indicate that a turning vehicle can enter the to-be-turned area during a straight ahead period and wait for a left turn. The left turn period ends. Vehicles are prohibited from staying in the area to be turned. The line is a white dotted line. The landmark line may include a left turn guide line, which indicates a boundary between a left turn motorized vehicle and a non-motorized vehicle. Motor vehicles travel to the left of the line, and non-motor vehicles travel to the right of the line. The lines are white dashed lines. The landmark line may include a pedestrian crossing line, which indicates that a pedestrian is allowed to cross the lane, and the line is a white parallel thick solid line (orthogonal, oblique). The landmarks can include highway entrance and exit markings, which are used to provide safe intersections for vehicles entering or exiting the ramp, and to reduce collisions with protruding edges, including horizontal markings at entrances and exits, and markings in triangular areas. The landmark line may include a parking space line indicating the location where the vehicle is parked, and the line is a solid white line. The landmark line can include a harbour-style stop line, which indicates that public buses pass through a dedicated separation approach and stop location. The landmark line may include a toll island mark, indicating the location of the toll island, and providing clear markings for vehicles entering the toll lane. The landmark line may include a guide arrow, which indicates the driving direction of the vehicle, and is mainly used in the guide lane at the intersection, near the exit ramp and to guide channelized traffic. The landmark line may include a road surface text line, which is a mark that indicates or restricts a vehicle's travel using road surface text, for example, a maximum speed, a large car, a small car, a passing lane. The landmark line may include a forbidden line, for example, a roadside parking line is prohibited, and a line indicating parking on the roadside is prohibited. The landmark line may include a stop line, for example, a stop line at a signal light intersection, which is a solid white line, indicating a parking position where the vehicle is waiting to be released. The landmark line can include a parking yield line, which means that vehicles must stop at this intersection or slow down to give way to the main road. The landmark line can include a deceleration yield line, which means that vehicles must decelerate or stop at this intersection, allowing vehicles on the main road to go ahead. The landmark line may include a non-motor vehicle no-car zone marking line to notify the cyclist of the prohibited entry area at the intersection. The landmark line may include a mesh line to inform the driver that temporary parking is not allowed at the intersection. The landmark line can include a center circle to distinguish between large and small turning vehicles. The landmark line may include a dedicated lane line to indicate a dedicated lane restricted to a certain vehicle type, and other vehicles and pedestrians may not enter. The landmark line may include a U-turn prohibition mark. The landmark line may include a warning line and a lane width gradient line to warn the driver of the vehicle that the road width is reduced or the number of lanes is reduced. Drive with caution and prohibit overtaking. Landmark lines can include deceleration lines to warn drivers of vehicles that they should slow down.
本申请描述的“乘客端”、“乘客端用户”、“乘客端设备”、“司机端”、“司机端用户”、“司机端设备”、“客户端”、“客户端设备”、“客户端用户”、“用户”、“使用者”、“用户端”等是可以互换的,是指需要或者订购服务的一方,可以是个人,也可以是 工具。另外,本申请描述的“用户”可以是需要或者订购服务的一方,也可以是提供服务或者协助提供服务的一方。"Passenger", "Passenger User", "Passenger Device", "Driver", "Driver User", "Driver Device", "Client", "Client Device", " "Client user", "user", "user", "user terminal", etc. are interchangeable and refer to the party who needs or subscribes to the service, which can be an individual or a tool. In addition, the "user" described in this application may be a party that requires or subscribes to a service, or a party that provides or assists in providing a service.
图1是根据本申请的一些实施例所示的一种按需服务系统的示意图。FIG. 1 is a schematic diagram of an on-demand service system according to some embodiments of the present application.
图像数据处理系统100可以包括服务器110、网络120、用户端130、车载设备140和存储器150。The image data processing system 100 may include a server 110, a network 120, a user terminal 130, a vehicle-mounted device 140, and a memory 150.
服务器110可以是本地的,也可以是远程的。服务器110可以处理信息和/或数据。在一些实施例中,服务器110可以用于对收集的信息进行分析加工以生成分析结果的系统。服务器110可以是一个终端设备,也可以是一个服务器,还可以是服务器群组。所述服务器群组可以是集中式的,例如数据中心。所述服务器群组也可以是分布式的,例如分布式系统。在一些实施例中,服务器110可以在云平台上实施。仅作为示例,所述云平台可以包括私有云、公共云、混合云、社区云、分布云、内部云、多层云等或其任意组合。在一些实施例中,服务器110可以在图2中描述的包含了一个或者多个组件的计算设备200上执行。The server 110 may be local or remote. The server 110 may process information and / or data. In some embodiments, the server 110 may be a system for analyzing and processing the collected information to generate an analysis result. The server 110 may be a terminal device, a server, or a server group. The server group may be centralized, such as a data center. The server group may also be distributed, such as a distributed system. In some embodiments, the server 110 may be implemented on a cloud platform. For example only, the cloud platform may include a private cloud, a public cloud, a hybrid cloud, a community cloud, a distributed cloud, an internal cloud, a multi-layer cloud, or the like, or any combination thereof. In some embodiments, the server 110 may execute on the computing device 200 described in FIG. 2 that includes one or more components.
在一些实施例中,服务器110可以包括处理引擎112。处理引擎112可以处理信息和/或数据以执行本申请中描述的一个或以上的功能。例如,处理引擎112可以获取包含道路的图像。处理引擎112可以获取经训练的交通标线识别模型。处理引擎112可以基于所述图像和所述经训练的交通标线识别模型,获得所述俯视图像的至少一个图层;所述图层包括所述俯视图像中的至少一个交通标线,以及至少反映所述至少一个交通标线在所述俯视图像中的区域。处理引擎112可以对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述俯视图像的位置坐标。在一些实施例中,处理引擎112可以包括一个或多个处理引擎(例如,单核处理引擎或多核处理器)。仅作为示例,处理引擎112可以包括一个或以上硬件处理器,例如中央处理单元(CPU)、特定应用集成电路(ASIC)、特定应用指令集处理器(ASIP)、图像处理单元(GPU)、物理运算处理单元(PPU)、数字信号处理器(DSP)、现场可程序门阵列(FPGA)、可程序逻辑装置(PLD)、控制器、微控制器单元、精简指令集计算机(RISC)、微处理器等或其任意组合。In some embodiments, the server 110 may include a processing engine 112. The processing engine 112 may process information and / or data to perform one or more functions described in this application. For example, the processing engine 112 may acquire an image containing a road. The processing engine 112 may obtain a trained traffic marking recognition model. The processing engine 112 may obtain at least one layer of the overhead image based on the image and the trained traffic landmark recognition model; the layer includes at least one traffic landmark in the overhead image, and at least one An area of the at least one traffic line in the overhead image is reflected. The processing engine 112 may process the at least one layer to obtain position coordinates of the at least one traffic marking line in the overhead image. In some embodiments, the processing engine 112 may include one or more processing engines (eg, a single-core processing engine or a multi-core processor). For example only, the processing engine 112 may include one or more hardware processors, such as a central processing unit (CPU), an application-specific integrated circuit (ASIC), an application-specific instruction set processor (ASIP), an image processing unit (GPU), a physical Computing Processing Unit (PPU), Digital Signal Processor (DSP), Field Programmable Gate Array (FPGA), Programmable Logic Device (PLD), Controller, Microcontroller Unit, Reduced Instruction Set Computer (RISC), Microprocessor Device or any combination thereof.
用户端130可以是乘客端或司机端,也可以指发布服务订单的个人、工具或者其他实体。用户端130可以通过其携带的摄像头拍摄道路图像,并上传给服务器110。在一些实施例中,用户端130包括但不限于移动设备130-1、机动车内置设备130-2、笔记本电脑130-3、台式电脑130-4等中的一种或几种的组合。在一些实施例中,移动设备130-1可以包括可穿戴设备、智能移动设备、虚拟现实设备、增强现实设备等,或其任意组合。在一些实施例中,可穿戴设备可以包括手环、鞋袜、眼镜、头盔、手表、衣物、背包、智能配饰等或其任意组 合。在一些实施例中,移动设备可以包括移动电话、个人数字助理(PDA)、游戏设备、导航设备、销售点(POS)设备、膝上型电脑、台式机等或其任意组合。在一些实施例中,虚拟现实设备和/或增强现实设备可以包括虚拟现实头盔、虚拟现实眼镜、虚拟现实眼罩、增强实境头盔、增强实境眼镜、增强实境眼罩等或其任意组合。例如,虚拟现实设备和/或增强现实设备可以包括Google Glass TM、RiftCon TM、Fragments TM、Gear VR TM等。在一些实施例中,用户端130可以处理信息和/或数据。用户端可以用于对收集的信息进行分析加工以生成分析结果的系统。例如,用户端130可以对采集的图像进行滤波、去噪、校正畸变等预处理。在一些实施例中,用户端130还可以接收服务器110对图像的处理结果,并向用户输出所述处理结果。 The user terminal 130 may be a passenger terminal or a driver terminal, and may also refer to an individual, a tool, or other entity that issues a service order. The user terminal 130 may take a road image through a camera carried by the user terminal 130 and upload the road image to the server 110. In some embodiments, the user terminal 130 includes, but is not limited to, one or more of a mobile device 130-1, a built-in vehicle 130-2, a notebook computer 130-3, a desktop computer 130-4, and the like. In some embodiments, the mobile device 130-1 may include a wearable device, a smart mobile device, a virtual reality device, an augmented reality device, or the like, or any combination thereof. In some embodiments, the wearable device may include a bracelet, footwear, glasses, helmet, watch, clothing, backpack, smart accessories, etc. or any combination thereof. In some embodiments, the mobile device may include a mobile phone, a personal digital assistant (PDA), a gaming device, a navigation device, a point of sale (POS) device, a laptop computer, a desktop computer, etc., or any combination thereof. In some embodiments, the virtual reality device and / or the augmented reality device may include a virtual reality helmet, virtual reality glasses, a virtual reality eye mask, an augmented reality helmet, an augmented reality glasses, an augmented reality eye mask, etc., or any combination thereof. For example, the virtual reality device and / or augmented reality device may include Google Glass , RiftCon , Fragments , Gear VR ™, and the like. In some embodiments, the client 130 may process information and / or data. The user terminal can be used to analyze and process the collected information to generate a system analysis result. For example, the user end 130 may perform preprocessing such as filtering, denoising, and correcting distortion on the acquired image. In some embodiments, the client 130 may also receive the processing result of the image by the server 110 and output the processing result to the user.
车载设备140可以是移动设备140-1、导航设备140-2、笔记本电脑140-3、行车记录仪140-4。例如,车载设备140可以包括但不限于行车记录仪、车载摄像机、车载照相机、车载记录仪、车载记录装置、行车记录器、车联网后视镜、行驶记录仪、行车记录装置等具有摄像功能的装置。在一些实施例中,车载设备140可以拍摄道路图像以完成道路图像的采集过程。服务器110可以获取上述道路图像并对其进行处理。The in-vehicle device 140 may be a mobile device 140-1, a navigation device 140-2, a notebook computer 140-3, and a driving recorder 140-4. For example, the in-vehicle device 140 may include, but is not limited to, a driving recorder, a vehicle camera, a vehicle camera, a vehicle recorder, a vehicle recording device, a driving recorder, a car rear-view mirror, a driving recorder, a driving recorder, and the like having a camera function. Device. In some embodiments, the in-vehicle device 140 may capture a road image to complete a road image acquisition process. The server 110 may acquire the above-mentioned road image and process it.
服务器110可以直接访问存取储存在存储器150的数据信息,也可以直接通过网络120访问存取用户端130的信息。The server 110 may directly access and access data information stored in the memory 150, and may also directly access and access information of the client 130 through the network 120.
存储器150可以泛指具有存储功能的设备。存储器150主要用于存储从用户端130收集的数据和图像数据处理系统100工作中产生的各种数据。例如,存储设备150可以存储从用户端130和/或车载设备140获取的图像。又例如,存储设备150可以存储处理引擎112执行或用于图像数据处理的指令。又例如,存储设备150可以存储用于训练交通标线识别模型的训练样本。具体的,该训练样本可以为多张原始图像及其标注信息。其中,原始图像可以包括道路的俯视图像或道路的非俯视视角下的图像,原始图像还可以包括道路的非俯视视角下的图像经处理后得到的俯视图像,标注信息反映各原始图像中交通标线的区域,和/或类别。The memory 150 may generally refer to a device having a storage function. The memory 150 is mainly used to store data collected from the user terminal 130 and various data generated during the operation of the image data processing system 100. For example, the storage device 150 may store images acquired from the user terminal 130 and / or the in-vehicle device 140. As another example, the storage device 150 may store instructions executed by the processing engine 112 or used for image data processing. As another example, the storage device 150 may store training samples for training a traffic marking recognition model. Specifically, the training sample may be multiple original images and their label information. The original image may include a top-view image of the road or an image in a non-top view of the road, and the original image may also include a processed top-view image of the image in a non-top view of the road. The annotation information reflects the traffic signs in each original image The area of the line, and / or the category.
存储器150可以是本地的,也可以是远程的。系统数据库与系统其他模块间的连接或通信可以是有线的,也可以是无线的。在一些实施例中,存储设备150可以包括大容量存储器、可移动存储器、易失性读写存储器、只读存储器(ROM)等或其任意组合。示例性的大容量储存器可以包括磁盘、光盘、固态磁盘等。示例性可移动存储器可以包括闪存驱动器、软盘、光盘、存储卡、压缩盘、磁带等。示例性的挥发性只读存储器可以包括随机存取内存(RAM)。示例性的RAM可包括动态RAM(DRAM)、双倍速率同步动态RAM(DDR SDRAM)、静态RAM(SRAM)、闸流体RAM(T-RAM)和零电容RAM(Z-RAM)等。 示例性的ROM可以包括掩模ROM(MROM)、可编程ROM(PROM)、可擦除可编程ROM(EPROM)、电子可擦除可编程ROM(EEPROM)、光盘ROM(CD-ROM)和数字通用磁盘ROM等。在一些实施例中,所述存储设备150可以在云平台上实现。仅作为示例,所述云平台可以包括私有云、公共云、混合云、社区云、分布云、内部云、多层云等或其任意组合。The memory 150 may be local or remote. The connection or communication between the system database and other modules of the system can be wired or wireless. In some embodiments, the storage device 150 may include mass storage, removable memory, volatile read-write memory, read-only memory (ROM), etc., or any combination thereof. Exemplary mass storage may include magnetic disks, optical disks, solid-state disks, and the like. Exemplary removable memories may include flash drives, floppy disks, optical disks, memory cards, compact disks, magnetic tapes, and the like. An exemplary volatile read-only memory may include random access memory (RAM). Exemplary RAMs may include dynamic RAM (DRAM), double-rate synchronous dynamic RAM (DDR SDRAM), static RAM (SRAM), gate fluid RAM (T-RAM), zero-capacity RAM (Z-RAM), and the like. Exemplary ROMs may include mask ROM (MROM), programmable ROM (PROM), erasable programmable ROM (EPROM), electronically erasable programmable ROM (EEPROM), compact disc ROM (CD-ROM), and digital General-purpose disk ROM, etc. In some embodiments, the storage device 150 may be implemented on a cloud platform. For example only, the cloud platform may include a private cloud, a public cloud, a hybrid cloud, a community cloud, a distributed cloud, an internal cloud, a multi-layer cloud, or the like, or any combination thereof.
在一些实施例中,存储设备150可以连接到网络120以与图像数据处理系统100(例如,服务器110、车载设备140等)中的一个或以上组件通信。图像数据处理系统100中的一个或以上组件可以经由网络120访问存储设备150中存储的数据或指令。在一些实施例中,存储设备150可以直接连接到图像数据处理系统100(例如,服务器110、车载设备140)中的一个或以上组件或与之通信。在一些实施例中,存储设备150可以是服务器110的一部分。In some embodiments, the storage device 150 may be connected to the network 120 to communicate with one or more components in the image data processing system 100 (eg, the server 110, the in-vehicle device 140, etc.). One or more components in the image data processing system 100 may access data or instructions stored in the storage device 150 via the network 120. In some embodiments, the storage device 150 may be directly connected to or in communication with one or more components in the image data processing system 100 (eg, the server 110, the in-vehicle device 140). In some embodiments, the storage device 150 may be part of the server 110.
网络120可以提供信息和/或数据交换的渠道。在一些实施例中,图像数据处理系统100中的一个或以上组件(例如,服务器110、用户端130、车载设备140和存储器150)可以通过网络120向图像数据处理系统100中的其他组件发送信息和/或数据。例如,处理引擎112可以通过网络120从存储设备150和/或车载设备140获取道路的俯视图像。网络120可以是单一网络,也可以是多种网络组合的。网络120可以包括但不限于电缆网络、有线网络、光纤网络、远程通信网路、内部网络、互联网、局域网络(LAN)、广域网(WAN)、无线局域网络(WLAN)、城域网络(MAN)、公共交换电话网络(PSTN)、蓝牙网络、ZigBee网络、近场通讯(NFC)网络等或其任意组合。网络120可以包括一个或多个网络接入点,如有线或无线接入点、基站(如120-1,120-2)或网络交换点,通过以上接入点使数据源连接网络120并通过网络发送信息。The network 120 may provide a channel for information and / or data exchange. In some embodiments, one or more components in the image data processing system 100 (for example, the server 110, the client 130, the in-vehicle device 140, and the storage 150) may send information to other components in the image data processing system 100 through the network 120 And / or data. For example, the processing engine 112 may obtain a top-view image of the road from the storage device 150 and / or the vehicle-mounted device 140 through the network 120. The network 120 may be a single network or a combination of multiple networks. The network 120 may include, but is not limited to, a cable network, a wired network, a fiber optic network, a telecommunications network, an internal network, the Internet, a local area network (LAN), a wide area network (WAN), a wireless local area network (WLAN), and a metropolitan area network (MAN). , Public switched telephone network (PSTN), Bluetooth network, ZigBee network, near field communication (NFC) network, etc. or any combination thereof. The network 120 may include one or more network access points, such as a wired or wireless access point, a base station (such as 120-1, 120-2), or a network exchange point. Through the above access points, the data source is connected to the network 120 and passed through Send information over the network.
应当注意,图像数据处理系统100仅是为了说明的目的而提供的,并不旨在限制本申请的范围。对于本领域的普通技术人员来说,可以根据本申请的描述,做出各种修正和改变。例如,图像数据处理系统100还可以包括数据库、信息源等。又例如,图像数据处理系统100可以在其他设备上实现,以实现类似或不同的功能。然而,这些修正和改变不会背离本申请的范围。It should be noted that the image data processing system 100 is provided for illustrative purposes only and is not intended to limit the scope of the application. For those of ordinary skill in the art, various modifications and changes can be made according to the description of this application. For example, the image data processing system 100 may further include a database, an information source, and the like. As another example, the image data processing system 100 may be implemented on other devices to implement similar or different functions. However, these amendments and changes will not depart from the scope of this application.
图2是用于实现本申请技术方案的专用系统的示例性计算设备200的框图。FIG. 2 is a block diagram of an exemplary computing device 200 for a dedicated system for implementing the technical solutions of the present application.
如图2所示,计算设备200可以包括处理器210、存储器220、输入/输出接口230和通信通信端口240。As shown in FIG. 2, the computing device 200 may include a processor 210, a memory 220, an input / output interface 230, and a communication communication port 240.
处理器210可以执行计算指令(程序代码)并执行本申请描述的图像数据处理系统100的功能。所述计算指令可以包括程序、对象、组件、数据结构、过程、模块和功能(所述功能指本申请中描述的特定功能)。例如,处理器210可以处理从图像数据处理系统100的 其他任何组件获得的图像或文本数据。在一些实施例中,处理器210可以包括微控制器、微处理器、精简指令集计算机(RISC)、专用集成电路(ASIC)、应用特定指令集处理器(ASIP)、中央处理器(CPU)、图形处理单元(GPU)、物理处理单元(PPU)、微控制器单元、数字信号处理器(DSP)、现场可编程门阵列(FPGA)、高级RISC机(ARM)、可编程逻辑器件以及能够执行一个或多个功能的任何电路和处理器等,或其任意组合。仅为了说明,图2中的计算设备200只描述了一个处理器,但需要注意的是本申请中的计算设备200还可以包括多个处理器。The processor 210 may execute calculation instructions (program code) and perform functions of the image data processing system 100 described in this application. The calculation instructions may include programs, objects, components, data structures, procedures, modules, and functions (the functions refer to specific functions described in this application). For example, the processor 210 may process image or text data obtained from any other component of the image data processing system 100. In some embodiments, the processor 210 may include a microcontroller, a microprocessor, a reduced instruction set computer (RISC), an application specific integrated circuit (ASIC), an application specific instruction set processor (ASIP), a central processing unit (CPU) , Graphics processing unit (GPU), physical processing unit (PPU), microcontroller unit, digital signal processor (DSP), field programmable gate array (FPGA), advanced RISC machine (ARM), programmable logic device, and capable of Any circuit, processor, etc., or any combination thereof that performs one or more functions. For illustration purposes only, the computing device 200 in FIG. 2 describes only one processor, but it should be noted that the computing device 200 in this application may further include multiple processors.
存储器220可以存储从图像数据处理系统100的任何其他组件获得的数据/信息。在一些实施例中,存储器220可以包括大容量存储器、可移动存储器、易失性读取和写入存储器和只读存储器(ROM)等,或其任意组合。示例性大容量存储器可以包括磁盘、光盘和固态驱动器等。可移动存储器可以包括闪存驱动器、软盘、光盘、存储卡、压缩盘和磁带等。易失性读取和写入存储器可以包括随机存取存储器(RAM)。RAM可以包括动态RAM(DRAM)、双倍速率同步动态RAM(DDR SDRAM)、静态RAM(SRAM)、晶闸管RAM(T-RAM)和零电容(Z-RAM)等。ROM可以包括掩模ROM(MROM)、可编程ROM(PROM)、可擦除可编程ROM(PEROM)、电可擦除可编程ROM(EEPROM)、光盘ROM(CD-ROM)和数字通用盘ROM等。The memory 220 may store data / information obtained from any other components of the image data processing system 100. In some embodiments, the memory 220 may include mass storage, removable memory, volatile read and write memory, read-only memory (ROM), or the like, or any combination thereof. Exemplary mass storage may include magnetic disks, optical disks, solid-state drives, and the like. Removable storage may include flash drives, floppy disks, optical disks, memory cards, compact disks, and magnetic tapes. Volatile read and write memory may include random access memory (RAM). RAM can include dynamic RAM (DRAM), double-rate synchronous dynamic RAM (DDR, SDRAM), static RAM (SRAM), thyristor RAM (T-RAM), zero capacitance (Z-RAM), and so on. ROM can include mask ROM (MROM), programmable ROM (PROM), erasable programmable ROM (PEROM), electrically erasable programmable ROM (EEPROM), compact disc ROM (CD-ROM), and digital versatile disc ROM Wait.
输入/输出(I/O)接口230可以用于输入或输出信号、数据或信息。在一些实施例中,输入/输出接口230可以使用户与图像数据处理系统100进行联系。在一些实施例中,输入/输出接口230可以包括输入装置和输出装置。示例性输入装置可以包括键盘、鼠标、触摸屏和麦克风等,或其任意组合。示例性输出设备可以包括显示设备、扬声器、打印机、投影仪等,或其任意组合。示例性显示装置可以包括液晶显示器(LCD)、基于发光二极管(LED)的显示器、平板显示器、曲面显示器、电视设备、阴极射线管(CRT)等,或其任意组合。通信端口240可以连接到网络以便数据通信。所述连接可以是有线连接、无线连接或两者的组合。有线连接可以包括电缆、光缆或电话线等,或其任意组合。无线连接可以包括蓝牙、Wi-Fi、WiMax、WLAN、ZigBee、移动网络(例如,3G、4G或5G等)等,或其任意组合。在一些实施例中,通信端口240可以是标准化端口,如RS232、RS485等。在一些实施例中,通信端口240可以是专门设计的端口。An input / output (I / O) interface 230 may be used to input or output signals, data, or information. In some embodiments, the input / output interface 230 may enable a user to communicate with the image data processing system 100. In some embodiments, the input / output interface 230 may include an input device and an output device. Exemplary input devices may include a keyboard, a mouse, a touch screen, a microphone, and the like, or any combination thereof. Exemplary output devices may include display devices, speakers, printers, projectors, etc., or any combination thereof. Exemplary display devices may include a liquid crystal display (LCD), a light emitting diode (LED) based display, a flat panel display, a curved display, a television device, a cathode ray tube (CRT), etc., or any combination thereof. The communication port 240 may be connected to a network for data communication. The connection may be a wired connection, a wireless connection, or a combination of both. The wired connection may include a cable, an optical cable or a telephone line, or any combination thereof. The wireless connection may include Bluetooth, Wi-Fi, WiMax, WLAN, ZigBee, a mobile network (for example, 3G, 4G, or 5G, etc.), etc., or any combination thereof. In some embodiments, the communication port 240 may be a standardized port, such as RS232, RS485, and the like. In some embodiments, the communication port 240 may be a specially designed port.
图3是用于实现本申请技术方案的专用系统的示例性移动设备300的框图。根据本申请的一些实施例,可以在其上实现用户端130。FIG. 3 is a block diagram of an exemplary mobile device 300 for implementing a dedicated system of the technical solution of the present application. According to some embodiments of the present application, the user terminal 130 may be implemented thereon.
如图3所示,所述移动设备300可以包括通信平台310、显示器320、图形处理器 (GPU)330、中央处理器(CPU)340、输入/输出接口350、内存360、存储器370等。CPU340可以包括接口电路和类似于处理器210的处理电路。在一些实施例中,任何其他合适的组件,包括但不限于系统总线或控制器(未示出),也可包括在移动设备300内。在一些实施例中,操作系统361(如,iOS,Android,Windows Phone等)和应用程序362可以从存储器370加载到内存360中,以便由CPU340执行。应用程序362可以包括浏览器或用于从图像数据处理系统100接收成像、图形处理、音频或其他相关信息的应用程序。用户与信息流的交互可以经由输入/输出350来实现,并经由网络120被提供给处理引擎112和/或图像数据处理系统100的其他组件。As shown in FIG. 3, the mobile device 300 may include a communication platform 310, a display 320, a graphics processor (GPU) 330, a central processing unit (CPU) 340, an input / output interface 350, a memory 360, a memory 370, and the like. The CPU 340 may include an interface circuit and a processing circuit similar to the processor 210. In some embodiments, any other suitable components, including but not limited to a system bus or controller (not shown), may also be included within the mobile device 300. In some embodiments, the operating system 361 (eg, iOS, Android, Windows Phone, etc.) and application programs 362 may be loaded from the memory 370 into the memory 360 for execution by the CPU 340. The application programs 362 may include a browser or an application program for receiving imaging, graphics processing, audio, or other related information from the image data processing system 100. The user's interaction with the information flow can be achieved via the input / output 350 and provided to the processing engine 112 and / or other components of the image data processing system 100 via the network 120.
为了实现在本申请中描述的各种模块、单元及其功能,计算设备或移动设备可以用作本申请所描述的一个或多个组件的硬件平台。这些计算机或移动设备的硬件元件、操作系统和编程语言本质上是常规的,并且本领域技术人员熟悉这些技术后可将这些技术适应于本申请所描述的按需服务系统。具有用户界面元件的计算机可以用于实现个人计算机(PC)或其他类型的工作站或终端设备,如果适当地编程,计算机也可以充当服务器。可以认为本领域技术人员也可以熟悉这种类型计算机设备的结构、程序或一般操作。因此,没有针对附图描述额外的解释。To implement the various modules, units, and functions described in this application, a computing device or mobile device may be used as a hardware platform for one or more components described in this application. The hardware elements, operating systems, and programming languages of these computers or mobile devices are conventional in nature, and those skilled in the art can adapt these technologies to the on-demand service system described in this application after being familiar with these technologies. A computer with user interface elements can be used to implement a personal computer (PC) or other type of workstation or terminal device, and the computer can also act as a server if properly programmed. It is believed that those skilled in the art may also be familiar with the structure, programs, or general operations of this type of computer equipment. Therefore, no additional explanation is described for the drawings.
图4为根据本申请的一些实施例所示的示例性图像数据处理方法的流程图。FIG. 4 is a flowchart of an exemplary image data processing method according to some embodiments of the present application.
在一些实施例中,该图像数据处理方法400由服务器110或计算设备200等具有处理、运算能力的设备执行。In some embodiments, the image data processing method 400 is executed by a device having processing and computing capabilities, such as the server 110 or the computing device 200.
步骤401,获取包含道路的图像。Step 401: Acquire an image including a road.
在一些实施例中,道路的图像可以包括道路的俯视图像。道路的俯视图像可以指从空中俯瞰道路的角度拍摄得到的道路图像。在一些实施例中,道路的图像还可以包括其他角度拍摄的道路图像。获取道路图像可以通过多种方式获取。例如,可以通过航拍的方式获取道路的图像,也可以通过安装在道路上的监控摄像头获取道路的图像,也可以通过处理用户端或车载设备(例如,行车记录仪)采集的道路视频获得道路的图像。In some embodiments, the image of the road may include a top view image of the road. A bird's-eye view image of a road may refer to a road image taken from an aerial view of the road. In some embodiments, the road image may also include road images taken from other angles. Obtaining road images can be obtained in many ways. For example, road images can be obtained by aerial photography, road images can also be obtained by surveillance cameras installed on the road, and road images collected by the user or on-board equipment (for example, driving recorder) can be used to obtain road images. image.
在一些实施例中,用户端或车载设备拍摄的道路图像包括非俯视他角度的道路图像。例如,用户坐在车辆内通过用户端获取道路图像,又例如,设置在车辆内后视镜处车载设备透过车辆前挡风玻璃拍摄的道路图像。在一些实施例中,可以对非俯视角度的道路图像进行处理,转换成相应的俯视图像。基于用户端或车载设备采集的道路视频获取道路的俯视图像的详细描述参见图5及其相关说明。In some embodiments, the road image captured by the user terminal or the in-vehicle device includes a road image not viewed from another angle. For example, a user sits in a vehicle to obtain a road image through a user terminal, and another example is a road image captured by a vehicle-mounted device placed in a rearview mirror of a vehicle through a vehicle's front windshield. In some embodiments, the road image at a non-top view angle may be processed and converted into a corresponding top view image. For a detailed description of acquiring a top view image of a road based on a road video collected by a user terminal or a vehicle-mounted device, refer to FIG. 5 and related descriptions.
步骤402,获取经训练的交通标线识别模型。Step 402: Obtain a trained traffic marking recognition model.
服务器110可以通过网络120获取经训练的交通标线识别模型。该经训练的交通标线识别模型可以识别出道路图像中的交通标线的类别和/或在图像中的位置,例如车道线在道路图像中的位置、各种地标分别在道路图像中的位置。The server 110 may obtain the trained traffic marking recognition model through the network 120. The trained traffic marking recognition model can identify the type of traffic markings in the road image and / or the location in the image, such as the position of lane lines in the road image, and the locations of various landmarks in the road image. .
一些实施例中,经训练的交通标线识别模型可以对一张道路图像进行处理,输出处理后的图像,并且在该图像上用不同的形式标注出不同类型的交通标线。例如,可以用户虚线框圈出各个交通标线和/或用文字标注不同的交通标线的类别。又例如,经训练的交通标线识别模型可以输出一张图像,并且在该图像上用不同的像素值表示不同类别的交通标线,例如,用户数字“1”表示车道线,用数字“2”表示一种地标(例如,人行道)。In some embodiments, the trained traffic marking recognition model can process a road image, output the processed image, and label different types of traffic markings in different forms on the image. For example, the user can circle each traffic line with a dotted frame and / or mark different categories of traffic lines with text. As another example, a trained traffic marking recognition model can output an image and use different pixel values to represent different types of traffic markings on the image. For example, the user number "1" represents a lane line, and the number "2 "Means a landmark (for example, a sidewalk).
仅作为示例说明,本实施例优选的经训练的交通标线识别模型为经训练的MASK-RCNN模型。训练MASK-RCNN模型时的输入为多张原始图像及其标注信息。其中,原始图像可以包括道路的俯视图像或道路的非俯视角度拍摄的图像,原始图像还可以包括道路的非俯视角度拍摄的图像经处理后得到的俯视图像,标注信息反映各原始图像中交通标线的像素坐标或像素位置,和/或类别。标注信息可以是标注了各种交通标线类别的图层信息。图层可以是与原始图像大小相同的图像。在一些实施例中,一张原始图像对应多张图层,每张图层上包含一类或一个交通标线。在一些实施例中,一张原始图像可以仅对应一张图层,图层上包含了原始图像中全部的交通标线。为了便于后续图像处理,在又一些实施例中,图层的像素值可以限定在有限的若干值上,例如图层可以是二值图像,图层中的交通标线像素值为“1”,其余像素点的像素值为“0”;又例如,通过在原始图像上标注不同类别的交通标线,可以在同一张图像上用不同像素值代表不同类别的交通标线(例如,像素值“1”代表车道线,像素值“2”代表人行道,像素值“3”代表直行),其余像素点的像素值为“4”。服务器110可以获得不同类别的交通标线对应的图层信息。基于原始图像及其各种交通标线的图层,MASK-RCNN模型通过学习,使得预测输出的各种交通标线的图层和标注的各种交通标线的图层达到一致,获得经训练的MASK-RCNN模型。在一些实施例中,经训练的交通标线识别模型还可以是经训练的基于卷积神经网络的交通标线模型、基于深度学习的交通标线识别模型、基于MATLAB的交通标线识别模型、基于智能认知的交通标线识别模型、基于形状特征的交通标线识别模型等。将不同类型的原始图像(如,道路的俯视图像、道路的非俯视角度拍摄的图像或经处理后得到的俯视图像)输入到MASK-RCNN模型中,经训练的MASK-RCNN模型可以对相应的道路图像进行处理,以得到对应图像的一个或多个图层,图层上可以包括模型识别得到的交通标线。By way of example only, the preferred trained traffic lane recognition model in this embodiment is a trained MASK-RCNN model. The input when training the MASK-RCNN model is multiple original images and their label information. The original image may include a top-view image of a road or an image taken at a non-top-down angle of the road, and the original image may also include a processed top-view image of an image taken at a non-top-down angle of the road. The pixel coordinates or pixel position of the line, and / or category. The labeling information may be layer information labeled with various traffic marking categories. The layer can be an image the same size as the original image. In some embodiments, one original image corresponds to multiple layers, and each layer contains a type or a traffic line. In some embodiments, one original image may correspond to only one layer, and the layer includes all traffic markings in the original image. In order to facilitate subsequent image processing, in still other embodiments, the pixel values of the layer may be limited to a limited number of values. For example, the layer may be a binary image, and the pixel value of the traffic markings in the layer is "1". The pixel values of the remaining pixels are "0"; for example, by marking different types of traffic markings on the original image, different pixel values can be used to represent different types of traffic markings on the same image (for example, the pixel value " 1 "represents lane lines, pixel value" 2 "represents sidewalk, pixel value" 3 "represents straight), and pixel values of the remaining pixels are" 4 ". The server 110 can obtain layer information corresponding to different types of traffic markings. Based on the original image and the layers of its various traffic markings, the MASK-RCNN model learns to make the predicted output traffic markings and the various traffic markings consistent. MASK-RCNN model. In some embodiments, the trained traffic marking recognition model may also be a trained traffic marking model based on a convolutional neural network, a deep learning-based traffic marking model, a MATLAB-based traffic marking recognition model, Traffic recognition model based on intelligent cognition, traffic feature recognition model based on shape features, etc. Different types of original images (such as road top images, road images taken at non-top angles, or processed top images) are input into the MASK-RCNN model. The trained MASK-RCNN model can The road image is processed to obtain one or more layers of the corresponding image, and the layer may include traffic markings obtained by model recognition.
步骤403,基于所述图像和所述经训练的交通标线识别模型,获得所述图像的至少一 个图层;所述图层包括所述图像中的至少一个交通标线,以及至少反映所述至少一个交通标线在所述图像中的区域。Step 403: Obtain at least one layer of the image based on the image and the trained traffic marking recognition model; the layer includes at least one traffic marking in the image, and reflects at least the An area of at least one traffic line in the image.
通过将图像输入到经训练的交通标线识别模型,服务器110可以获得图像的至少一个图层。具体的,根据输入的图像类型(如,道路的俯视图像、道路的非俯视视角下的图像或经处理后得到的俯视图像),可以相应输出对应不同图像类型的至少一个图层。在一些实施例中,输出的图层在对应于交通标线在图像中的相同位置具有相同的交通标线。图层可以包括一个或多种交通标线。在一些实施例中,一个图层可以只包括一种交通标线(例如,只包括车道线),也可以包括多种交通标线(例如,车道线和地标)。又例如,一个图层可以只包括一种地标(例如,人行道),也可以包括多种地标(例如,人行道、掉头、减速慢行)。又例如,一个图层可以只包括车道线,其他每个图层只包括一种地标(例如,一个图层只包括掉头、一个图层只包括人行道、一个图层只包括右转)。By inputting the image into a trained traffic marking recognition model, the server 110 can obtain at least one layer of the image. Specifically, at least one layer corresponding to different image types may be output according to an input image type (for example, a road top image, an image in a non-top view angle of the road, or a processed top image). In some embodiments, the output layers have the same traffic markings at the same location in the image corresponding to the traffic markings. A layer can include one or more traffic lines. In some embodiments, a layer may include only one type of traffic line (for example, only lane lines), or a plurality of types of traffic line (for example, lane lines and landmarks). As another example, a layer may include only one type of landmark (for example, a sidewalk), or may include multiple types of landmarks (for example, a sidewalk, a U-turn, and a slow down). As another example, one layer may only include lane lines, and each other layer may include only one type of landmark (for example, one layer may only include U-turns, one layer may include only sidewalks, and one layer may include only right turns).
图层可以为二值化图像(每个像素点的像素值为0或1),像素点的像素值为1表示该点为交通标线,像素点的像素值为0表示该点为非交通标线。图层还可以反映至少一个交通标线的类别。例如,某个图层只包括车道线的信息,也称为车道线图层,其中,像素点的像素值为1代表该点为车道线。又例如,某个图层只包括地标(例如,人行道)的信息,也称为地标图层,其中,像素点的像素值为1代表该点为地标。在其他实施例中,图层可以有更多种类的像素值,不同的交通标线用不同的像素值表示,而同一个交通标线的像素值相同;背景像素点采用区别于交通标线像素点的像素值。以MASK-RCNN模型为例,MASK-RCNN模型可以输出不同类别交通标线的mask图层信息,并可以获得交通标线的像素级位置信息。每种交通标线的mask图层信息为二维的0/1栅格数据(栅格可以包括一个或多个像素点),在有交通标线的区域像素值为1,反之为0。The layer can be a binary image (the pixel value of each pixel is 0 or 1). A pixel value of 1 indicates that the point is a traffic marking, and a pixel value of 0 indicates that the point is non-traffic. Marking. The layer can also reflect the category of at least one traffic line. For example, a layer only contains information about lane lines, which is also called a lane line layer, where a pixel value of 1 indicates that the point is a lane line. For another example, a certain layer only includes information of a landmark (for example, a sidewalk), which is also referred to as a landmark layer, and a pixel value of 1 indicates that the point is a landmark. In other embodiments, the layer may have more types of pixel values. Different traffic markings are represented by different pixel values, and the pixel values of the same traffic marking are the same. The background pixels are distinguished from the traffic marking pixels. The pixel value of the point. Taking the MASK-RCNN model as an example, the MASK-RCNN model can output mask layer information of different types of traffic markings, and can obtain pixel-level location information of traffic markings. The mask layer information of each type of traffic markings is two-dimensional 0/1 raster data (the grid can include one or more pixels), and the pixel value is 1 in the area with traffic markings, and the value is 0 otherwise.
在其他实施例中,输出图层上的交通标线可以由线框框出,线框内的区域即交通标线在图像中的区域。线框可以是矩形、圆形、椭圆形、多边形或不规则线框中的一种或多种组合。不同的线框可以具有不同的颜色。In other embodiments, the traffic markings on the output layer may be framed by a wire frame, and the area within the wire frame is the area of the traffic markings in the image. The wireframe can be one or more combinations of rectangular, circular, oval, polygonal, or irregular wireframes. Different wireframes can have different colors.
步骤404,对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标。Step 404: Process the at least one layer to obtain position coordinates of the at least one traffic line in the image.
所述图层反映至少一个交通标线在图像中的位置。在一些实施例中,输出的图层在对应于交通标线在图像中的相同位置具有相同的交通标线。通过对图层的进一步处理,服务器110可以获得至少一个交通标线在图像的更精确的位置坐标。所述处理方式包括但不限于包括图像去噪、图像变换、边缘检测、边缘细化、图像分析、图像压缩、图像增强、图像模糊处 理、图像插值、二值化变换、阈值变换、傅立叶变换、离散余弦变换等。关于图层处理的具体内容可以进一步参见后面的详细描述。The layer reflects the position of at least one traffic line in the image. In some embodiments, the output layers have the same traffic markings at the same location in the image corresponding to the traffic markings. Through further processing of the layer, the server 110 can obtain more precise position coordinates of at least one traffic line in the image. The processing methods include, but are not limited to, image denoising, image transformation, edge detection, edge thinning, image analysis, image compression, image enhancement, image blur processing, image interpolation, binary transformation, threshold transformation, Fourier transformation, Discrete cosine transform, etc. For details about layer processing, please refer to the detailed description later.
仅作为示例说明,图5示出了在一些实施例中,服务器110获取道路的俯视图像的步骤。By way of example only, FIG. 5 illustrates steps in which the server 110 acquires a top-view image of a road in some embodiments.
步骤501,获取车载设备拍摄的道路视频。Step 501: Obtain a road video taken by a vehicle-mounted device.
车载设备在车辆行驶过程中可以连续拍摄道路视频。车载设备包括但不限于行车记录仪、车载摄像机、车载照相机、车载记录仪、车载记录装置、行车记录器、车联网后视镜、行驶记录仪、行车记录装置等具有摄像功能的装置。服务器110可以通过网络120获取车载设备拍摄的道路视频。On-board equipment can continuously shoot road videos while the vehicle is driving. Vehicle-mounted devices include, but are not limited to, vehicle-mounted recorders, vehicle cameras, vehicle cameras, vehicle-mounted recorders, vehicle-mounted recording devices, vehicle-mounted recorders, connected car rear-view mirrors, vehicle-mounted recorders, and vehicle-recording devices. The server 110 may obtain a road video shot by the vehicle-mounted device through the network 120.
步骤502,基于所述道路视频,获取多张图像。Step 502: Acquire multiple images based on the road video.
服务器110可以将所述道路视频分割成多帧数据,获取多张图像。例如,服务器110可以提取连续的若干帧图像,也可以每间隔一帧获取一张图像,也可以间隔多帧获取一张图像。The server 110 may divide the road video into multiple frames of data to obtain multiple images. For example, the server 110 may extract several consecutive frames of images, obtain an image every one frame, or obtain an image at multiple frames.
步骤503,获取每张图像中同一位置的图像数据。Step 503: Obtain image data at the same position in each image.
在一些实施例中,服务器110可以提取每张图像中同一行的图像数据。在又一些实施例中,服务器110可以提取每张图像中相同的多行的图像数据。例如,服务器110可以从每张图像中从最上方往下分别提取三行(即第一行、第二行、第三行)图像数据。在一些实施例中,服务器110可以从每张图像中提取畸变较小的一行或多行数据。例如,可以提取图像中最上方或最下方的一行或多行数据,或者图像中间部分畸变较少的一行或多行。在一些实施例中,服务器110可以从每张图像中其他固定位置提取图像数据,该固定位置不限于图像中的固定某一行或多行的图像数据,也可以是局部图像数据,例如固定某一行或多行中图像数据中的部分列。In some embodiments, the server 110 may extract image data of the same line in each image. In still other embodiments, the server 110 may extract image data of the same multiple lines in each image. For example, the server 110 may extract three lines (ie, the first line, the second line, and the third line) of image data from the top to the bottom of each image. In some embodiments, the server 110 may extract one or more lines of data with less distortion from each image. For example, you can extract one or more lines of data at the top or bottom of the image, or one or more lines with less distortion in the middle part of the image. In some embodiments, the server 110 may extract image data from other fixed positions in each image. The fixed position is not limited to image data of a fixed line or multiple lines, but may also be local image data, such as a fixed line. Or some columns in the image data in multiple rows.
步骤504,拼接所述多张图像中同一位置的图像数据,获得所述道路的俯视图像。Step 504: stitch the image data of the same position among the multiple images to obtain a top-view image of the road.
例如,当服务器110可以提取每张图像中同一行的图像数据后,服务器110可以按照时间顺序拼接每张图像中同一行的图像数据形成最终的道路的俯视图像。在又一些实施例中,服务器110可以基于提取每张图像中相同的多行的图像数据,通过拼接形成多张道路的俯视图像。例如,当服务器110从每张图像中从最上方往下提取第100行、第320行、第480行三行图像数据后,服务器110可以基于每张图像的第100行的图像数据拼接形成第一张俯视图像、基于每张图像的第320行的图像数据拼接形成第二张俯视图像以及基于每张图像的第480行的图像数据拼接形成第三张俯视图像。服务器110在获得三张俯视图像后,可以通过 比较三张图像的质量(例如,清晰度)最终确定一张合适的俯视图像。For example, after the server 110 can extract the image data of the same line in each image, the server 110 can stitch the image data of the same line in each image in chronological order to form the final overhead image of the road. In still other embodiments, the server 110 may form a plurality of top-view images of roads by stitching based on extracting image data of the same multiple lines in each image. For example, after the server 110 extracts three rows of image data from line 100, line 320, and line 480 from the top of each image, the server 110 may stitch the image data based on the line 100 image data of each image to form a line. A bird's-eye view image, the image data of the 320th line of each image are stitched to form a second bird's-eye view image, and the image data of the 480th line of each image are stitched to form a third bird's-eye view image. After the server 110 obtains the three overhead images, it can finally determine a suitable overhead image by comparing the quality (e.g., sharpness) of the three images.
一般来说,车载设备是安装在车内透过前挡风玻璃获取道路图像,图像中的道路可能呈现出近宽远窄的视觉效果,经过图5示出的流程处理后,道路在图像不同行上的宽度可以近似相等,可以理解为是将车载设备获取的道路图像转换为成了与俯视角度拍摄效果近似的图像。Generally, on-vehicle equipment is installed in the car to obtain road images through the front windshield. The roads in the images may appear near wide, narrow and narrow visual effects. After the process shown in FIG. 5, the roads have different images. The widths on the lines can be approximately equal, which can be understood as converting the road images acquired by the in-vehicle equipment into images with similar shooting effects from a bird's-eye view.
仅作为示例说明,图7示出了在其他一些实施例中,服务器110基于所述图像和所述经训练的交通标线识别模型,获得所述图像的至少一个图层的步骤包括:For illustration only, FIG. 7 shows that in other embodiments, the server 110 obtains at least one layer of the image based on the image and the trained traffic marking recognition model, including:
步骤701,基于图像,获取多张子图像;所述多张子图像分别包括所述图像的一部分,且至少两张子图像共同包括所述图像中的某一部分。在一些实施例中,该图像可以包括道路的俯视图像,也可以包括道路的非俯视视角下的图像或非俯视视角下的图像经处理后得到的俯视图像。关于图像的详细描述参见图4及其相关说明,此处不作赘述。Step 701: Obtain multiple sub-images based on the images; the multiple sub-images each include a portion of the image, and at least two sub-images collectively include a portion of the image. In some embodiments, the image may include a top-view image of the road, and may also include an image in a non-top-view perspective of the road or an image in a non-top-view perspective after processing. For a detailed description of the image, refer to FIG. 4 and related descriptions, and details are not described herein.
步骤702,针对每张子图像,基于所述子图像和所述经训练的交通标线识别模型,获得所述子图像的至少一个图层;所述图层包括所述子图像中的至少一个交通标线的至少一部分,以及至少反映所述至少一个交通标线的至少一部分在所述子图像中的区域。Step 702: For each sub-image, obtain at least one layer of the sub-image based on the sub-image and the trained traffic marking recognition model; the layer includes at least one of the sub-images At least a portion of a traffic line, and at least an area in the sub-image that reflects at least a portion of the at least one traffic line.
步骤703,基于至少两张子图像的至少一个图层,联合确定出所述图像的至少一个图层。Step 703: Jointly determine at least one layer of the images based on at least one layer of the at least two sub-images.
仅作为示例说明,服务器110可以通过小于图像的窗口截取多张图像的子图像。例如,图像的像素大小1024×1024,该窗口的大小为64×64。通过将窗口的第32行、第32列的像素点对准图像的每个像素点对图像进行截取,可以获得1024×1024张大小均为64×64的子图像。窗口内未覆盖的图像的像素点的值用像素值0代替。For illustration only, the server 110 may capture sub-images of multiple images through a window smaller than the image. For example, the pixel size of an image is 1024 × 1024, and the size of the window is 64 × 64. By aligning the pixels of the 32nd row and the 32nd column of the window with each pixel of the image, the image can be captured to obtain 1024 × 1024 sub-images each having a size of 64 × 64. The pixel value of the uncovered image in the window is replaced with a pixel value of 0.
通过将每张子图像输入到经训练的交通标线识别模型(例如,MASK-RCNN模型),可以识别出每张子图像的至少一个图层,其中,每个图层可以只包括一种交通标线。对于单个图层而言,该图层的像素点的像素值为0或1,其中,1表示该像素点为该交通标线。例如,对于某个图层为车道线图层而言,其中,像素点的像素值为1表示该点为车道线。又例如,对于某个图层为人行道图层而言,其中,像素点的像素值为1表示该点为人行道。进一步地,对同一种交通标线的多个图层中在俯视图像中为同一位置的像素点而言,当该像素点的像素值为1的统计个数大于某一阈值时,则服务器110可以确定图像对应的该交通标线的图层中该点的像素值最终为1。例如,当100张子图像分别输入到交通标线识别模型(例如,MASK-RCNN模型),输出100个子图像车道线图层,该100个子图像车道线图层包括了图像中的同一个A点。对于100个子图像车道线图层中,对应于图像A点的像素值为1的个数大于50 时,则服务器110可以确定图像A点在图像车道线图层中的像素值为1,否则,确定A点在最终车道线图层中的最终像素值为0。类似地,服务器110可以确定出图像的每个图层中每个像素点的最终像素值。最终,服务器110可以确定出图像的至少一个图层。例如,服务器110最终可以确定出车道线图层、每种地标对应的图层(例如,人行道图层、掉头图层、左转图层等)。通过该方法可以对图像的子区域识别多次,将多次识别的结果进行联合(例如,融合)判断图像的图层中对应像素位置处的像素值,能有效降低识别误差,提高图像的识别精度。At least one layer of each sub-image can be identified by inputting each sub-image into a trained traffic marking recognition model (for example, a MASK-RCNN model), where each layer can include only one type of traffic Marking. For a single layer, the pixel value of the pixel point of the layer is 0 or 1, where 1 indicates that the pixel point is the traffic marking. For example, for a layer that is a lane line layer, a pixel value of 1 indicates that the point is a lane line. For another example, for a layer that is a sidewalk layer, a pixel value of 1 at a pixel point indicates that the point is a sidewalk. Further, for a plurality of layers of the same traffic marking line, the pixels at the same position in the top view image, when the statistical number of the pixel value of the pixel point is greater than a certain threshold, the server 110 It can be determined that the pixel value of the point in the layer of the traffic marking corresponding to the image is finally 1. For example, when 100 sub-images are input to a traffic marking recognition model (for example, a MASK-RCNN model), 100 sub-image lane line layers are output. The 100 sub-image lane line layers include the same A point in the image. . For the 100 sub-image lane line layers, when the number of pixel values corresponding to point A of the image is greater than 50, the server 110 may determine that the pixel value of point A in the image lane line layer is 1, otherwise, Determine that the final pixel value of point A in the final lane line layer is 0. Similarly, the server 110 may determine the final pixel value of each pixel point in each layer of the image. Finally, the server 110 may determine at least one layer of the image. For example, the server 110 may finally determine a lane line layer and a layer corresponding to each landmark (for example, a sidewalk layer, a U-turn layer, a left turn layer, etc.). Through this method, the sub-region of the image can be identified multiple times, and the results of the multiple recognition can be combined (e.g., fused) to determine the pixel value at the corresponding pixel position in the image layer, which can effectively reduce the recognition error and improve the image recognition. Precision.
在一些实施中,服务器110可以对图层的处理保括但不限于进行腐蚀处理、膨胀处理或平滑处理。In some implementations, the server 110 may include, but is not limited to, performing an etching process, an expansion process, or a smoothing process on the processing of the layer.
以车道线图层为例,图8为对车道线图层处理前的原图,白色代表车道线,黑色代表非车道线。Take the lane line layer as an example. Figure 8 is the original image before processing the lane line layer. White represents the lane line and black represents the non-lane line.
在一些实施例中,在提取车道线前,服务器110可以对车道线图层进行形态学处理,主要是使用比较小的矩形核(例如,核(2,2),核(3,3),核(5,5))进行腐蚀。腐蚀操作指当核内的像素点的像素值不全为1时,将核内的像素值为1的像素点由1变为0(即由白色变为黑色),当核内的像素点的像素值全为1时,则不作改变。通过此操作可以绝大部分地过滤掉无效(或者孤立)的噪点。图9为对车道线图层进行腐蚀后的结果图,从图9中可以看出,经过腐蚀,相比原车道线,腐蚀后的车道线变窄。In some embodiments, before the lane line is extracted, the server 110 may perform morphological processing on the lane line layer, mainly using a relatively small rectangular kernel (for example, kernel (2, 2), kernel (3, 3), Core (5,5)). Corrosion operation means that when the pixel value of a pixel in the core is not all 1, the pixel value of the pixel in the core is changed from 1 to 0 (that is, from white to black). When the value is all 1, no change is made. This operation can largely filter out invalid (or isolated) noise. FIG. 9 is a result diagram of the lane line layer after being corroded. As can be seen from FIG. 9, after the erosion, the lane line after the erosion is narrower than the original lane line.
通过腐蚀操作,可以绝大部分地过滤掉无效(或者孤立)的噪点,但是同时对有效的信息将会产生损失,此时可以使用比较大的矩形核(例如,核(20,3)、核(20,5)或核(5,3))进行膨胀操作。膨胀操作指当核内的像素点的像素值不全为0时,将核内的像素值为0的像素点由0变为1(即由黑色变为白色),当核内的像素点的像素值全为0时,则不作改变。图10为对车道线图层进行腐蚀后又经过膨胀处理后的结果图。通过膨胀,一部分可以抵消腐蚀操作中引入的精度损失,一部分可以解决由于部分遮挡导致的车道线断开情况(如图9中的901部分两条车道线是断开的,经过膨胀后得到图10中1001部分两条车道线是连接的)。Through the corrosion operation, most of the invalid (or isolated) noise can be filtered out, but at the same time, valid information will be lost. At this time, a relatively large rectangular kernel (for example, kernel (20,3), kernel (20,5) or core (5,3)). The dilation operation means that when the pixel values of the pixels in the kernel are not all 0, the pixels with the pixel values of 0 in the kernel are changed from 0 to 1 (that is, from black to white). When the value is all 0, no change is made. FIG. 10 is a result diagram of the lane line layer after being corroded and then subjected to expansion processing. Part of the expansion can offset the loss of precision introduced in the corrosion operation, and part of it can solve the situation of lane line disconnection caused by partial occlusion (as shown in Figure 901, the two lane lines are disconnected. After expansion, Figure 10 is obtained. The two lane lines in the middle 1001 section are connected).
在一些实施中,图11示出了响应于图像的图层为车道线图层,服务器110可以对车道线图层进行以下操作获得图像的位置坐标的方法。在一些实施例中,对车道线图层进行以下操作之前,服务器110可以对车道线图层进行腐蚀处理、膨胀处理或平滑处理中的一种或多种处理。In some implementations, FIG. 11 illustrates a method in which the server 110 may perform the following operations on the lane line layer to obtain the position coordinates of the image in response to the image layer being a lane line layer. In some embodiments, before performing the following operations on the lane line layer, the server 110 may perform one or more of an etching process, an expansion process, or a smoothing process on the lane line layer.
步骤1101,基于车道线图层和车道线图层中像素点的像素值,确定至少一条车道线的左边缘点和右边缘点在车道线图层中的位置坐标。Step 1101: Determine the position coordinates of the left edge point and the right edge point of at least one lane line in the lane line layer based on the pixel values of the pixel points in the lane line layer and the lane line layer.
服务器110可以对图层自上而下进行左右车道线边缘提取。例如,服务器110可以每次读取一行数据,确定该行中的第一条车道线的第一左边缘点和第一右边缘点,其中,第一左边缘点为像素值由0变为1(由黑变白)的第一个像素点,第一右边缘点为像素值由1变为0(由白变黑)的第一个像素点。服务器110继续扫描该行数据,得到第二左边缘点和第二右边缘点,如果第二左边缘点与第一左边缘点的距离大于一定的阈值,则说明为存在两条车道线,以此类推,找到第三条或者第四条车道线的左边缘点、右边缘点。服务器110扫描不同行的数据,将该行的车道线的某像素点(例如左边缘点)的位置与前一行的车道线的某像素点(例如左边缘点)的位置对比,如果两点距离小于一定的阈值,则作为同一条车道线,这样将第二行信息与第一行信息结合起来;以此类推,结合多行信息,最后连接得到多条车道线。The server 110 may perform left and right lane line edge extraction on the layer from top to bottom. For example, the server 110 may read one line of data at a time to determine the first left edge point and the first right edge point of the first lane line in the line, where the first left edge point is a pixel value from 0 to 1 The first pixel point (from black to white), and the first right edge point is the first pixel point whose pixel value changes from 1 to 0 (from white to black). The server 110 continues to scan the line of data to obtain the second left edge point and the second right edge point. If the distance between the second left edge point and the first left edge point is greater than a certain threshold, it means that there are two lane lines, and By analogy, find the left edge point and right edge point of the third or fourth lane line. The server 110 scans data of different lines, and compares the position of a pixel point (for example, the left edge point) of the lane line of the line with the position of a pixel point (for example, the left edge point) of the lane line of the previous line. If it is less than a certain threshold, it is regarded as the same lane line, so that the second line of information is combined with the first line of information; and so on, combining multiple lines of information, and finally connecting to obtain multiple lane lines.
步骤1102,基于所述至少一条车道线的左边缘点和右边缘点在车道线图层中的位置坐标,确定所述至少一条车道线的中心线在车道线图层中的位置坐标。Step 1102: Determine position coordinates of a center line of the at least one lane line in the lane line layer based on position coordinates of a left edge point and a right edge point of the at least one lane line in the lane line layer.
在一些实施例中,服务器110可以对车道线图层中每一行车道线的左边缘点和右边缘点的坐标进行一定的运算,将运算结果作为车道线中心线上点的坐标。例如,将车道线图层中每一行车道线的左边缘点和右边缘点的坐标求平均,将结果作为车道线中心线的坐标位置。服务器110通过连接车道线的左边缘点和右边缘点的中点形成该车道线的中心线。图12为对车道线图层确定车道线的中心线后的结果图。In some embodiments, the server 110 may perform a certain operation on the coordinates of the left edge point and the right edge point of each lane line in the lane line layer, and use the operation result as the coordinates of the center line point of the lane line. For example, the coordinates of the left edge point and the right edge point of each lane line in the lane line layer are averaged, and the result is taken as the coordinate position of the lane line center line. The server 110 forms the center line of the lane line by connecting the midpoint of the left edge point and the right edge point of the lane line. FIG. 12 is a result diagram after the center line of the lane line is determined for the lane line layer.
步骤1103,基于车道线的预设宽度以及所述至少一条车道线的中心线的位置坐标,确定所述至少一条车道线在所述图像中的位置坐标。Step 1103: Determine position coordinates of the at least one lane line in the image based on a preset width of the lane line and position coordinates of a center line of the at least one lane line.
服务器110基于车道线的预设宽度(例如,10mm、20mm、30mm、50mm、100mm、200mm等),可以确定车道线图层中的车道线的像素宽度。例如,服务器110可以确定车道线图层中的车道线的像素宽度为30个像素点。在一些实施例中,服务器110可以以车道线的中心线为基准左右拓宽一定像素区域,确定车道线在图像中的位置坐标。例如,服务器110可以以车道线的中心线为基准向左右两侧各拓宽15个像素点,从而确定车道线在图像中的位置坐标。The server 110 may determine the pixel width of the lane line in the lane line layer based on a preset width of the lane line (for example, 10 mm, 20 mm, 30 mm, 50 mm, 100 mm, 200 mm, etc.). For example, the server 110 may determine that the pixel width of the lane lines in the lane line layer is 30 pixels. In some embodiments, the server 110 may widen a certain pixel area from the center line of the lane line to the left and right to determine the position coordinates of the lane line in the image. For example, the server 110 may use the center line of the lane line as a reference to widen 15 pixels to the left and right sides to determine the position coordinates of the lane line in the image.
在一些实施例中,服务器110可以判断某条车道线的中心线的末端和另一条车道线的中心线的首端的在水平方向的距离是否小于第一阈值,并且在竖直方向上的距离是否小于第二阈值。示例性的,第一阈值可以设为15个像素点,第二阈值可以设为100个像素点。当某条车道线的中心线的末端和另一条车道线的中心线的首端的在水平方向的距离小于第一阈值,并且在竖直方向上的距离小于第二阈值时,服务器110可以将所述至少一条车道线的末端和 所述另一条车道线的中心线的首端通过线性插值的方法进行连接。进一步地,服务器110可以通过将拼接后的中心线为基准向左右两侧各拓宽若干个像素点,确定所述至少一条车道线在所述俯视图像中的位置坐标。In some embodiments, the server 110 may determine whether the distance in the horizontal direction between the end of the centerline of one lane line and the head of the centerline of another lane line is less than the first threshold, and whether the distance in the vertical direction is Less than the second threshold. Exemplarily, the first threshold may be set to 15 pixels, and the second threshold may be set to 100 pixels. When the distance in the horizontal direction between the end of the centerline of one lane line and the head of the centerline of another lane line is less than the first threshold and the distance in the vertical direction is less than the second threshold, the server 110 may The ends of the at least one lane line and the head of the center line of the other lane line are connected by a linear interpolation method. Further, the server 110 may determine a position coordinate of the at least one lane line in the bird's-eye view image by widening a number of pixels to the left and right sides by using the spliced center line as a reference.
在一些实施例中,服务器110可以先对车道线图层进行形态学处理(包括腐蚀、膨胀)后进行图11的操作,如图13所示为车道线图层经过腐蚀、膨胀后提取中心车道线中心线的结果图。In some embodiments, the server 110 may perform the morphological processing (including corrosion and expansion) on the lane line layer and then perform the operation in FIG. 11, as shown in FIG. 13, extracting the center lane after the lane line layer is corroded and expanded. Result graph of line centerline.
在一些实施例中,由于行人遮挡或者其他遮挡的原因,检测到的某条车道线可能被分为几段。对于该车道线属于断的情况,服务器110可以对车道线作拼接、滤波和平滑处理。图14示出了一种示例性的处理多段车道线的方法的流程。In some embodiments, a detected lane line may be divided into several segments due to pedestrian occlusion or other occlusion reasons. For the case where the lane line is broken, the server 110 may perform splicing, filtering, and smoothing on the lane line. FIG. 14 shows a flowchart of an exemplary method for processing a multi-segment lane line.
步骤1401,判断车道线图层中至少一条车道线的某一端点和另一条车道线的某一端点在水平方向的距离是否小于第一阈值,并且在竖直方向上的距离是否小于第二阈值。 Step 1401, it is determined whether a distance between a certain end point of at least one lane line and another end point of another lane line in the lane line layer in a horizontal direction is less than a first threshold value, and whether the distance in a vertical direction is less than a second threshold value .
服务器110可以通过步骤1101的方法确定每条车道线某一端部的左边缘点和右边缘点在车道线图层中的位置坐标。基于每条车道线某端部的左边缘点和右边缘点在车道线图层中的位置坐标,服务器110可以确定车道线图层中每条车道线的端点。应当注意的是,端部为车道线在车道线图层的竖直方向上的起始点或终点所在的区域,例如车道线最上方的一行像素点或最下方的一行像素,由于车道线具有一定的宽度,该端部可以包括某一行的多个像素点。而端点为端部区域的某一点,用于表征车道线的端部。服务器110可以确定车道线图层中每条车道线的某一端点和另一条车道线的某一端点在水平方向的距离(也称为水平距离)和竖直方向上的距离(也称为竖直距离)。服务器110可以计算某条车道线末端(可以包括多个像素点)中的一个端点和另一条车道线的首端(可以包括多个像素点)中的一个端点的水平距离和竖直距离。例如,服务器110可以计算某条车道线的中心线的末端和另一条车道线的中心线的首端的水平距离和竖直距离。又例如,服务器110可以计算某条车道线的末端的左边缘点和另一条车道线的首端的左边缘点的水平距离和竖直距离。进一步地,判断车道线图层中至少一条车道线的某一端点和另一条车道线的某一端点在水平方向的距离是否小于第一阈值,并且在竖直方向上的距离是否小于第二阈值。The server 110 may determine the position coordinates of the left edge point and the right edge point of a certain end of each lane line in the lane line layer by the method of step 1101. Based on the position coordinates of the left edge point and the right edge point of a certain end of each lane line in the lane line layer, the server 110 may determine the endpoint of each lane line in the lane line layer. It should be noted that the end is the area where the lane line starts or ends in the vertical direction of the lane line layer, such as the top row of pixels or the bottom row of pixels. Width, the end can include multiple pixels in a row. The end point is a point in the end area, which is used to characterize the end of the lane line. The server 110 may determine a distance in a horizontal direction (also referred to as a horizontal distance) and a distance in a vertical direction (also referred to as a vertical direction) between a certain end point of each lane line and another end point of another lane line in the lane line layer. Straight distance). The server 110 may calculate a horizontal distance and a vertical distance between one end point of a certain lane line end (which may include multiple pixels) and one end point of another lane line head end (which may include multiple pixels). For example, the server 110 may calculate a horizontal distance and a vertical distance between the end of the centerline of a certain lane line and the beginning of the centerline of another lane line. As another example, the server 110 may calculate a horizontal distance and a vertical distance between a left edge point at the end of a certain lane line and a left edge point at the head of another lane line. Further, it is determined whether a distance between a certain end point of at least one lane line and another end point of another lane line in the lane line layer in a horizontal direction is less than a first threshold value, and whether a distance in a vertical direction is less than a second threshold value. .
步骤1402,响应于所述至少一条车道线的某一端点和另一条车道线的某一端点在水平方向的距离小于第一阈值,并且在竖直方向上的距离小于第二阈值时,则将所述至少一条车道线的所述某一端点和所述另一条车道线的所述某一端点进行拼接。In step 1402, in response to a distance between a certain end point of the at least one lane line and a certain end point of the other lane line in the horizontal direction being smaller than the first threshold and the distance in the vertical direction being smaller than the second threshold, The one end point of the at least one lane line and the one end point of the other lane line are spliced.
例如,第一阈值可以设为15个像素点,第二阈值可以设为100个像素点。在一些实施例中,服务器110可以将所述至少一条车道线的末端的左边缘点和右边缘点和另一条车道 线的首端的左边缘点和右边缘点形成的区域范围的像素点的像素值变为1,从而形成两段车道线的拼接段。For example, the first threshold may be set to 15 pixels, and the second threshold may be set to 100 pixels. In some embodiments, the server 110 may combine pixels in a range of pixels formed by the left edge point and the right edge point of the end of the at least one lane line and the left edge point and the right edge point of the head end of the other lane line. The value becomes 1 to form a spliced segment of the two lane lines.
在一些实施例中,服务器110可以将所述至少一条车道线的中心线的末端和所述另一条车道线的中心线的首端通过线性插值的方法进行连接。进一步地,服务器110可以通过将拼接后的中心线为基准向左右两侧各拓宽若干个像素点,实现两段车道线的拼接。In some embodiments, the server 110 may connect the end of the center line of the at least one lane line and the head of the center line of the other lane line by a linear interpolation method. Further, the server 110 can widen a number of pixels to the left and right sides by using the spliced center line as a reference to realize the splicing of two lane lines.
在一些实施例中,服务器110可以将所述至少一条车道线的末端的左边缘点和所述另一条车道线的首端的左边缘点通过线性插值的方法进行连接。进一步地,服务器110可以通过将拼接后的左边缘线为基准向右宽若干个像素点,实现两段车道线的拼接。In some embodiments, the server 110 may connect the left edge point of the end of the at least one lane line and the left edge point of the head of the other lane line by a linear interpolation method. Further, the server 110 can realize the splicing of the two lane lines by using the left edge line after splicing as a reference to be several pixels wide to the right.
在一些实施例中,服务器110可以将所述至少一条车道线的末端的右边边缘点和所述另一条车道线的首端的右边缘点通过线性插值的方法进行连接。进一步地,服务器110可以通过将拼接后的右边缘线为基准向左拓宽若干个像素点,实现两段车道线的拼接。In some embodiments, the server 110 may connect a right edge point at the end of the at least one lane line and a right edge point at the head of the other lane line by a linear interpolation method. Further, the server 110 may widen a number of pixels to the left by using the spliced right edge line as a reference to realize splicing of two lane lines.
步骤1403,判断车道线图层中某条车道线的长度是否小于第三阈值。Step 1403: Determine whether the length of a lane line in the lane line layer is less than a third threshold.
服务器110可以确定某条车道线的首端和末端在车道线图层中的位置。服务器110可以基于某条车道线的首端和末端在车道线图层中的位置,从而确定该条车道线的长度。进一步地,服务器110可以确定某条车道线的长度是否小于第三阈值。该第三阈值可以为服务器110预先设定的值,例如,可以为100个像素点的距离。The server 110 may determine the positions of the head and end of a lane line in the lane line layer. The server 110 may determine the length of the lane line based on the position of the head and end of the lane line in the lane line layer. Further, the server 110 may determine whether the length of a certain lane line is less than a third threshold. The third threshold may be a value set in advance by the server 110, and for example, may be a distance of 100 pixels.
步骤1404,响应于所述某条车道线的长度小于第三阈值,将所述车道线从所述图像中去除。 Step 1404, in response to the length of a certain lane line being less than a third threshold, removing the lane line from the image.
当某条车道线的长度小于第三阈值,服务器110可以将该条车道线在车道线图层中的像素点的像素值由1改为0(即由白色变为黑色)。When the length of a lane line is less than the third threshold, the server 110 may change the pixel value of the pixel point of the lane line in the lane line layer from 1 to 0 (that is, from white to black).
步骤1405,对处理后的车道线图层做窗口平滑处理。Step 1405: Perform window smoothing processing on the processed lane line layer.
在一些实施例中,在提取的车道线中,可能因为车辆的短时颠簸,使得极少数的车道线提取后存在“尖刺”等情况,在不损失精度的情况下,服务器110可以对车道线图层做平滑处理。平滑处理算法包括但不限于均值滤波、中值滤波、高斯滤波、双边滤波等。In some embodiments, in the extracted lane lines, the short-term bumps of the vehicle may cause a very small number of lane lines to have “spikes” after being extracted. Without loss of accuracy, the server 110 may Line layers are smoothed. Smoothing algorithms include, but are not limited to, mean filtering, median filtering, Gaussian filtering, bilateral filtering, and so on.
应当注意的是,方法1400对车道线进行拼接、滤波和平滑的处理顺序只是示例性的,本申请并不限定处理的顺序。例如,服务器110可以先对车道线进行滤波处理后,再进行拼接和平滑处理。此外,方法1400中的某些步骤还可以独立使用。例如步骤1401余1402可以独立使用,进行车道线拼接。步骤1403~1404可以合并使用,去除地面上的干扰交通标线识别的其他痕迹或物体。1405可以独立使用,对交通标线进行平滑处理。诸如此类的变形,均在本申请的保护范围之内。It should be noted that the processing sequence of the method 1400 for splicing, filtering, and smoothing lane lines is only exemplary, and the application does not limit the processing sequence. For example, the server 110 may perform filtering processing on the lane lines before performing stitching and smoothing processing. In addition, certain steps in method 1400 may be used independently. For example, steps 1401 and 1402 can be used independently to perform lane line splicing. Steps 1403 to 1404 can be used in combination to remove other traces or objects on the ground that interfere with traffic markings recognition. 1405 can be used independently to smooth the traffic markings. Such deformations are all within the protection scope of this application.
图15为根据本申请的一些实施例所示的另一种对图层进行处理以获得至少一个交通标线在图像中的位置坐标的方法的流程图。该方法1500可以用于处理图像的图层为地标图层。服务器110可以对地标图层进行以下操作获得地标在图像中的位置坐标。15 is a flowchart of another method for processing a layer to obtain position coordinates of at least one traffic line in an image according to some embodiments of the present application. The method 1500 can be used to process an image as a landmark layer. The server 110 may perform the following operations on the landmark layer to obtain the position coordinates of the landmark in the image.
步骤1501,基于所述至少一个图层和该图层中像素点的像素值,确定该图层中交通标线在水平方向上的最大坐标值和最小坐标值,以及在竖直方向上的最大坐标值和最小坐标值。Step 1501: Based on the at least one layer and the pixel values of the pixels in the layer, determine a maximum coordinate value and a minimum coordinate value of the traffic markings in the layer in the horizontal direction and a maximum value in the vertical direction. Coordinate value and minimum coordinate value.
以代表直行的三角形地标为例,如图16所示,服务器110可以确定出三角形地标在图层水平方向的最大坐标值为x2、最小坐标值为x1,以及在竖直方向上的最大坐标值y2和最小坐标值y1。Taking a triangular landmark representing a straight line as an example, as shown in FIG. 16, the server 110 may determine that the maximum coordinate value of the triangular landmark in the horizontal direction of the layer is x2, the minimum coordinate value is x1, and the maximum coordinate value in the vertical direction y2 and the minimum coordinate value y1.
步骤1502,基于所述交通标线在水平方向上的最大坐标值和最小坐标值以及在竖直方向上的最大坐标值和最小坐标值,确定该地标在所述图像中的位置坐标。Step 1502: Determine the position coordinates of the landmark in the image based on the maximum coordinate value and the minimum coordinate value in the horizontal direction of the traffic marking line and the maximum coordinate value and the minimum coordinate value in the vertical direction.
如图16所示,基于三角形地标在水平方向上的最大坐标值和最小坐标值以及在竖直方向上的最大坐标值和最小坐标值,服务器110可以确定三角形地标的矩形区域。服务器110可以基于坐标A(x1,y1)、坐标B(x2,y2)、坐标C(x1,y2)和坐标D(x2,y2),确定三角形地标的矩形区域ABCD。在绘制地图时,服务器110可以将包含地标的区域以矩形框的方式框出,然后标记是直行、左转或者斑马线等导引线。As shown in FIG. 16, based on the maximum coordinate value and the minimum coordinate value of the triangular landmark in the horizontal direction and the maximum coordinate value and the minimum coordinate value in the vertical direction, the server 110 may determine a rectangular area of the triangular landmark. The server 110 may determine the rectangular area ABCD of the triangular landmark based on the coordinates A (x1, y1), coordinates B (x2, y2), coordinates C (x1, y2), and coordinates D (x2, y2). When drawing a map, the server 110 may frame a region containing a landmark in a rectangular frame manner, and then mark the guide line such as going straight, turning left, or a zebra crossing.
对于其他地标标线而言,服务器110同样可以通过方法1500的方法识别出地标所处的矩形区域。在绘制地图时,服务器110将包含地标的区域以矩形框的方式框出,然后标记是直行、左转或者斑马线等导引线。需要注意的是,该方法1500同样适用于对车道线的处理。For other landmark lines, the server 110 can also identify the rectangular area where the landmark is located by using the method 1500. When drawing a map, the server 110 frames the area containing the landmark in a rectangular frame manner, and then marks the guide line such as going straight, turning left, or a zebra crossing. It should be noted that the method 1500 is also applicable to the processing of lane lines.
图17为根据本申请的一些实施例所示的一种图像数据处理装置的框图。FIG. 17 is a block diagram of an image data processing apparatus according to some embodiments of the present application.
图像数据处理装置1700可以由服务器110实现。为描述方便,图像数据处理装置1700也可以称为图像数据处理系统。The image data processing apparatus 1700 may be implemented by the server 110. For convenience of description, the image data processing apparatus 1700 may also be referred to as an image data processing system.
图像数据处理装置1700图像获取模块1710、图层获取模块1720和图层处理模块1730。The image data processing device 1700 includes an image acquisition module 1710, a layer acquisition module 1720, and a layer processing module 1730.
图像获取模块1710用于获取包含道路的图像。The image acquisition module 1710 is configured to acquire an image including a road.
图层获取模块1720用于基于所述图像和经训练的交通标线识别模型,获得所述图像的至少一个图层;所述图层包括所述图像中的至少一个交通标线,以及至少反映所述至少一个交通标线在所述图像中的区域。A layer acquisition module 1720 is configured to obtain at least one layer of the image based on the image and a trained traffic marking recognition model; the layer includes at least one traffic marking in the image, and reflects at least one A region of the at least one traffic line in the image.
图层处理模块1730,用于对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标。A layer processing module 1730 is configured to process the at least one layer to obtain position coordinates of the at least one traffic line in the image.
在一些实施例中,经训练的交通标线识别模型为经训练的MASK-RCNN模型。In some embodiments, the trained traffic marking recognition model is a trained MASK-RCNN model.
在一些实施例中,图层的图像大小与所述图像的图像大小相同,且所述图层为二值化图像。In some embodiments, the image size of the layer is the same as the image size of the image, and the layer is a binary image.
在一些实施例中,图层还反映所述至少一个交通标线的类别。In some embodiments, the layer also reflects a category of the at least one traffic line.
在一些实施例中,图层处理模块1730还用于对所述至少一个图层进行腐蚀处理、膨胀处理或平滑处理中的一种或多种组合。In some embodiments, the layer processing module 1730 is further configured to perform one or more combinations of an etching process, an expansion process, or a smoothing process on the at least one layer.
在一些实施例中,图像包括道路的俯视图像。In some embodiments, the image includes a top-down image of the road.
在一些实施例中,响应于图像的图层为车道线图层,图层处理模块1730还用于基于所述车道线图层和所述车道线图层中像素点的像素值,确定至少一条车道线的左边缘点和右边缘点在所述车道线图层中的位置坐标;基于所述至少一条车道线的左边缘点和右边缘点在所述车道线图层中的位置坐标,确定所述至少一条车道线的中心线在所述车道线图层中的位置坐标;基于车道线的预设宽度以及所述至少一条车道线的中心线的位置坐标,确定所述至少一条车道线在所述图像中的位置坐标。In some embodiments, in response to the layer of the image being a lane line layer, the layer processing module 1730 is further configured to determine at least one of the lane line layers and the pixel values of the pixels in the lane line layer. Position coordinates of a left edge point and a right edge point of a lane line in the lane line layer; based on position coordinates of the left edge point and the right edge point of the at least one lane line in the lane line layer, determine Position coordinates of a center line of the at least one lane line in the lane line layer; determining a position of the at least one lane line based on a preset width of the lane line and position coordinates of the center line of the at least one lane line Position coordinates in the image.
在一些实施例中,响应于图像的图层为车道线图层,图层处理模块1730还用于判断车道线图层中至少一条车道线的某一端点和另一条车道线的某一端点在水平方向的距离是否小于第一阈值,并且在竖直方向上的距离是否小于第二阈值;响应于所述至少一条车道线的某一端点和另一条车道线的某一端点在水平方向的距离小于第一阈值,并且在竖直方向上的距离小于第二阈值时,则将所述至少一条道路的所述某一端点和所述另一条车道线的所述某一端点进行拼接。In some embodiments, in response to the image layer being a lane line layer, the layer processing module 1730 is further configured to determine whether an end point of at least one lane line and an end point of another lane line in the lane line layer are at Whether the distance in the horizontal direction is less than the first threshold value, and whether the distance in the vertical direction is less than the second threshold value; in response to a distance in the horizontal direction between an end point of the at least one lane line and an end point of the other lane line When the distance is smaller than the first threshold and the distance in the vertical direction is smaller than the second threshold, the certain end point of the at least one road and the certain end point of the other lane line are stitched.
在一些实施例中,响应于图像的图层为车道线图层,图层处理模块1730还用于判断车道线图层中某条车道线的长度是否小于第三阈值;响应于所述某条车道线的长度小于第三阈值,将所述车道线从所述图像中去除。In some embodiments, the layer in response to the image is a lane line layer, and the layer processing module 1730 is further configured to determine whether a length of a lane line in the lane line layer is less than a third threshold; in response to the certain line The length of the lane line is less than the third threshold, and the lane line is removed from the image.
在一些实施例中,图层处理模块1730还用于基于所述至少一个图层和该图层中像素点的像素值,确定该图层中交通标线在水平方向上的最大坐标值和最小坐标值,以及在竖直方向上的最大坐标值和最小坐标值;基于所述交通标线在水平方向上的最大坐标值和最小坐标值以及在竖直方向上的最大坐标值和最小坐标值,确定该交通标线在所述图像中的位置坐标。In some embodiments, the layer processing module 1730 is further configured to determine, based on the at least one layer and the pixel values of the pixels in the layer, the maximum coordinate value and the minimum coordinate value of the traffic markings in the layer in the horizontal direction. Coordinate values, and the maximum and minimum coordinate values in the vertical direction; based on the maximum and minimum coordinate values in the horizontal direction of the traffic markings, and the maximum and minimum coordinate values in the vertical direction To determine the position coordinates of the traffic marking in the image.
图18为根据本申请的一些实施例所示的一种俯视图像获取1710的框图。FIG. 18 is a block diagram of a top-view image acquisition 1710 according to some embodiments of the present application.
俯视图像获取模块1710包括视频获取单元1810、图像获取单元19820、图像数据提取单元1830和图像数据拼接单元1840。The overhead image acquisition module 1710 includes a video acquisition unit 1810, an image acquisition unit 19820, an image data extraction unit 1830, and an image data splicing unit 1840.
视频获取单元1810用于获取车载设备拍摄的道路视频。The video acquiring unit 1810 is configured to acquire a road video captured by a vehicle-mounted device.
图像获取单元1820用于基于所述道路视频,获取多张图像。The image acquisition unit 1820 is configured to acquire multiple images based on the road video.
图像数据提取单元1830用于获取每张图像中同一位置的图像数据。The image data extraction unit 1830 is configured to acquire image data of a same position in each image.
图像数据拼接单元1840用于拼接所述多张图像中同一位置的图像数据,获得所述道路的俯视图像。The image data splicing unit 1840 is configured to splice image data of the same position among the multiple images to obtain a top-view image of the road.
在一些实施例中,每张图像中同一位置的图像数据包括每张图像中相同的至少一行。In some embodiments, the image data at the same position in each image includes at least one line that is the same in each image.
图19为根据本申请的一些实施例所示的一种图层获取模块1720的框图。FIG. 19 is a block diagram of a layer acquisition module 1720 according to some embodiments of the present application.
图层获取模块1720包括子图像获取单元1910、子图层获取单元1920和联合确定单元1930。The layer acquisition module 1720 includes a sub-image acquisition unit 1910, a sub-layer acquisition unit 1920, and a joint determination unit 1930.
子图像获取单元1910用于基于图像,获取多张子图像;所述多张子图像分别包括所述图像的一部分,且至少两张子图像共同包括所述图像中的某一部分;The sub-image obtaining unit 1910 is configured to obtain multiple sub-images based on the image; the multiple sub-images each include a part of the image, and at least two sub-images collectively include a certain part of the image;
子图层获取单元1920用于针对每张子图像,基于所述子图像和所述经训练的交通标线识别模型,获得所述子图像的至少一个图层;所述图层包括所述子图像中的至少一个交通标线的至少一部分,以及至少反映所述至少一个交通标线的至少一部分在所述子图像中的区域;The sub-layer obtaining unit 1920 is configured to obtain, for each sub-image, at least one layer of the sub-image based on the sub-image and the trained traffic marking recognition model; the layer includes the sub-image At least a portion of at least one traffic line in the image, and at least a region reflecting at least a portion of the at least one traffic line in the sub-image;
联合确定单元1930用于基于至少两张子图像的至少一个图层,联合确定出所述图像的至少一个图层。The joint determining unit 1930 is configured to jointly determine at least one layer of the images based on at least one layer of the at least two sub-images.
本申请披露了一种图像数据处理系统。该系统包括至少一个处理器以及至少一个存储器;所述至少一个存储器用于存储计算机指令;所述至少一个处理器用于执行所述计算机指令中的至少部分指令以实现任意一种图像数据处理方法。The present application discloses an image data processing system. The system includes at least one processor and at least one memory; the at least one memory is configured to store computer instructions; the at least one processor is configured to execute at least a part of the computer instructions to implement any image data processing method.
本发披露了一种计算机可读存储介质。该存储介质存储计算机指令,当所述计算机指令被处理器执行时实现任意一种图像数据处理方法。The present disclosure discloses a computer-readable storage medium. The storage medium stores computer instructions, and when the computer instructions are executed by a processor, any image data processing method is implemented.
需要说明的是,上述各个模块可以是通过计算机指令实现的软件模块。以上对于候选项显示、确定系统及其模块的描述,仅为描述方便,并不能把本申请限制在所举实施例范围之内。可以理解,对于本领域的技术人员来说,在了解该系统的原理后,可能在不背离这一原理的情况下,对各个模块进行任意组合,或者构成子系统与其他模块连接。例如,各个模块可以共用一个存储模块,各个模块也可以分别具有各自的存储模块。诸如此类的变形,均在本申请的保护范围之内。It should be noted that the foregoing modules may be software modules implemented by computer instructions. The above description of the candidate selection display and determination system and its modules is for convenience of description only, and cannot limit the application to the scope of the illustrated embodiments. It can be understood that for those skilled in the art, after understanding the principle of the system, it is possible to arbitrarily combine the various modules or form a subsystem to connect with other modules without departing from this principle. For example, each module can share a storage module, and each module can also have its own storage module. Such deformations are all within the protection scope of this application.
本申请实施例可能带来的有益效果包括但不限于:(1)通过交通标线识别模型可以识别出不同类型的交通标线;(2)对提取的车道线进行拼接处理、滤波处理、平滑处理在一定程度上解决了交通标线被部分遮挡而不连续的问题;(3)提取道路视频中的多张图像的同 一位置拼接成道路的俯视图像一定程度上解决了图像畸变的问题;(4)通过Mask-RCNN模型可以识别出不同类型的交通标线在图层中的像素级位置,满足更高精度交通标线识别的要求。需要说明的是,不同实施例可能产生的有益效果不同,在不同的实施例里,可能产生的有益效果可以是以上任意一种或几种的组合,也可以是其他任何可能获得的有益效果。The beneficial effects that the embodiments of the present application may bring include, but are not limited to: (1) different types of traffic markings can be identified through a traffic marking recognition model; (2) splicing processing, filtering processing, and smoothing of the extracted lane lines The processing solves the problem that the traffic markings are partially occluded and discontinuous to a certain extent; (3) extracting the same position of multiple images in the road video to stitch the top view image of the road to a certain extent solves the problem of image distortion; ( 4) The Mask-RCNN model can identify the pixel-level positions of different types of traffic markings in the layer, meeting the requirements for higher-precision traffic marking recognition. It should be noted that different embodiments may have different beneficial effects. In different embodiments, the possible beneficial effects may be any one or a combination of the foregoing, or any other beneficial effects that may be obtained.
上文已对基本概念做了描述,显然,对于本领域技术人员来说,上述详细披露仅仅作为示例,而并不构成对本申请的限定。虽然此处并没有明确说明,本领域技术人员可能会对本申请进行各种修改、改进和修正。该类修改、改进和修正在本申请中被建议,所以该类修改、改进、修正仍属于本申请示范实施例的精神和范围。The basic concepts have been described above. Obviously, for those skilled in the art, the above detailed disclosure is merely an example, and does not constitute a limitation on the present application. Although it is not explicitly stated here, those skilled in the art may make various modifications, improvements and amendments to this application. Such modifications, improvements and amendments are suggested in this application, so such modifications, improvements and amendments still belong to the spirit and scope of the exemplary embodiments of this application.
同时,本申请使用了特定词语来描述本申请的实施例。如“一个实施例”、“一实施例”、和/或“一些实施例”意指与本申请至少一个实施例相关的某一特征、结构或特点。因此,应强调并注意的是,本说明书中在不同位置两次或多次提及的“一实施例”或“一个实施例”或“一个替代性实施例”并不一定是指同一实施例。此外,本申请的一个或多个实施例中的某些特征、结构或特点可以进行适当的组合。Meanwhile, specific words are used in this application to describe the embodiments of this application. For example, "one embodiment", "an embodiment", and / or "some embodiments" means a certain feature, structure, or characteristic related to at least one embodiment of the present application. Therefore, it should be emphasized and noted that the "one embodiment" or "one embodiment" or "an alternative embodiment" mentioned two or more times in different places in this specification does not necessarily refer to the same embodiment . In addition, certain features, structures, or characteristics in one or more embodiments of the present application may be appropriately combined.
此外,本领域技术人员可以理解,本申请的各方面可以通过若干具有可专利性的种类或情况进行说明和描述,包括任何新的和有用的工序、机器、产品或物质的组合,或对他们的任何新的和有用的改进。相应地,本申请的各个方面可以完全由硬件执行、可以完全由软件(包括固件、常驻软件、微码等)执行、也可以由硬件和软件组合执行。以上硬件或软件均可被称为“数据块”、“模块”、“引擎”、“单元”、“组件”或“系统”。此外,本申请的各方面可能表现为位于一个或多个计算机可读介质中的计算机产品,该产品包括计算机可读程序编码。In addition, those skilled in the art can understand that aspects of this application can be illustrated and described through several patentable categories or situations, including any new and useful processes, machines, products, or combinations of materials, or their Any new and useful improvements. Accordingly, various aspects of the present application can be executed entirely by hardware, can be executed entirely by software (including firmware, resident software, microcode, etc.), and can also be executed by a combination of hardware and software. The above hardware or software can be called "data block", "module", "engine", "unit", "component" or "system". In addition, aspects of the present application may manifest as a computer product located in one or more computer-readable media, the product including a computer-readable program code.
计算机存储介质可能包含一个内含有计算机程序编码的传播数据信号,例如在基带上或作为载波的一部分。该传播信号可能有多种表现形式,包括电磁形式、光形式等,或合适的组合形式。计算机存储介质可以是除计算机可读存储介质之外的任何计算机可读介质,该介质可以通过连接至一个指令执行系统、装置或设备以实现通讯、传播或传输供使用的程序。位于计算机存储介质上的程序编码可以通过任何合适的介质进行传播,包括无线电、电缆、光纤电缆、RF、或类似介质,或任何上述介质的组合。Computer storage media may contain a transmitted data signal containing a computer program code, such as on baseband or as part of a carrier wave. The propagation signal may have multiple manifestations, including electromagnetic form, optical form, etc., or a suitable combination form. A computer storage medium may be any computer-readable medium other than a computer-readable storage medium, which may be connected to an instruction execution system, apparatus, or device to enable communication, propagation, or transmission of a program for use. Program code on a computer storage medium may be transmitted through any suitable medium, including radio, cable, fiber optic cable, RF, or similar media, or any combination of the foregoing.
本申请各部分操作所需的计算机程序编码可以用任意一种或多种程序语言编写,包括面向对象编程语言如Java、Scala、Smalltalk、Eiffel、JADE、Emerald、C++、C#、VB.NET、Python等,常规程序化编程语言如C语言、Visual Basic、Fortran 2003、Perl、COBOL 2002、PHP、ABAP,动态编程语言如Python、Ruby和Groovy,或其他编程语言等。该程序编码可 以完全在用户计算机上运行、或作为独立的软件包在用户计算机上运行、或部分在用户计算机上运行部分在远程计算机运行、或完全在远程计算机或服务器上运行。在后种情况下,远程计算机可以通过任何网络形式与用户计算机连接,比如局域网(LAN)或广域网(WAN),或连接至外部计算机(例如通过因特网),或在云计算环境中,或作为服务使用如软件即服务(SaaS)。The computer program code required for the operation of each part of this application can be written in any one or more programming languages, including object-oriented programming languages such as Java, Scala, Smalltalk, Eiffel, JADE, Emerald, C ++, C #, VB.NET, Python Etc., conventional programming languages such as C, Visual Basic, Fortran 2003, Perl, COBOL 2002, PHP, ABAP, dynamic programming languages such as Python, Ruby and Groovy, or other programming languages. The program code may run entirely on the user's computer, or as a stand-alone software package on the user's computer, or partly on the user's computer, partly on a remote computer, or entirely on the remote computer or server. In the latter case, the remote computer can be connected to the user's computer through any network form, such as a local area network (LAN) or wide area network (WAN), or to an external computer (for example, via the Internet), or in a cloud computing environment, or as a service Uses such as software as a service (SaaS).
此外,除非权利要求中明确说明,本申请所述处理元素和序列的顺序、数字字母的使用、或其他名称的使用,并非用于限定本申请流程和方法的顺序。尽管上述披露中通过各种示例讨论了一些目前认为有用的实施例,但应当理解的是,该类细节仅起到说明的目的,附加的权利要求并不仅限于披露的实施例,相反,权利要求旨在覆盖所有符合本申请实施例实质和范围的修正和等价组合。例如,虽然以上所描述的系统组件可以通过硬件设备实现,但是也可以只通过软件的解决方案得以实现,如在现有的服务器或移动设备上安装所描述的系统。In addition, unless explicitly stated in the claims, the order of processing elements and sequences described in this application, the use of alphanumeric characters, or the use of other names is not intended to limit the order of the processes and methods of this application. Although the above disclosure discusses some embodiments that are currently considered useful through various examples, it should be understood that this type of detail is for illustration purposes only, and the appended claims are not limited to the disclosed embodiments. Instead, the claims It is intended to cover all modifications and equivalent combinations that conform to the spirit and scope of the embodiments of the present application. For example, although the system components described above can be implemented by hardware devices, they can also be implemented only by software solutions, such as installing the described system on an existing server or mobile device.
同理,应当注意的是,为了简化本申请披露的表述,从而帮助对一个或多个实施例的理解,前文对本申请实施例的描述中,有时会将多种特征归并至一个实施例、附图或对其的描述中。但是,这种披露方法并不意味着本申请对象所需要的特征比权利要求中提及的特征多。实际上,实施例的特征要少于上述披露的单个实施例的全部特征。In the same way, it should be noted that, in order to simplify the expressions disclosed in this application and thereby help the understanding of one or more embodiments, the foregoing description of the embodiments of the application sometimes incorporates multiple features into one embodiment, attached Figure or description of it. However, this disclosure method does not mean that the subject of the present application requires more features than those mentioned in the claims. Indeed, the features of an embodiment are less than all the features of a single embodiment disclosed above.
一些实施例中使用了描述成分、属性数量的数字,应当理解的是,此类用于实施例描述的数字,在一些示例中使用了修饰词“大约”、“近似”或“大体上”来修饰。除非另外说明,“大约”、“近似”或“大体上”表明所述数字允许有±20%的变化。相应地,在一些实施例中,说明书和权利要求中使用的数值参数均为近似值,该近似值根据个别实施例所需特点可以发生改变。在一些实施例中,数值参数应考虑规定的有效数位并采用一般位数保留的方法。尽管本申请一些实施例中用于确认其范围广度的数值域和参数为近似值,在具体实施例中,此类数值的设定在可行范围内尽可能精确。In some embodiments, numbers describing the number of components and attributes are used. It should be understood that, for such numbers used in the description of the embodiments, the modifiers "about", "approximately" or "substantially" Modification. Unless stated otherwise, "about", "approximately" or "substantially" indicates that the number allows for a variation of ± 20%. Accordingly, in some embodiments, the numerical parameters used in the specification and claims are approximate values, and the approximate values may be changed according to the characteristics required by individual embodiments. In some embodiments, the numerical parameter should take the specified significant digits into account and adopt a general digits retention method. Although the numerical ranges and parameters used to confirm the breadth of the range in some embodiments of this application are approximate values, in specific embodiments, the setting of such values is as accurate as possible within the feasible range.
针对本申请引用的每个专利、专利申请、专利申请公开物和其他材料,如文章、书籍、说明书、出版物、文档等,特此将其全部内容并入本申请作为参考。与本申请内容不一致或产生冲突的申请历史文件除外,对本申请权利要求最广范围有限制的文件(当前或之后附加于本申请中的)也除外。需要说明的是,如果本申请附属材料中的描述、定义、和/或术语的使用与本申请所述内容有不一致或冲突的地方,以本申请的描述、定义和/或术语的使用为准。For each patent, patent application, patent application publication, and other materials cited in this application, such as articles, books, instructions, publications, documents, etc., the entire contents thereof are hereby incorporated by reference. Except for application history files that are inconsistent with or conflicting with the content of this application, except for those files that have the broadest scope of the claims in this application (currently or later attached to this application). It should be noted that if there is any inconsistency or conflict between the use of descriptions, definitions, and / or terms in the accompanying materials of this application, the descriptions, definitions, and / or terms of this application shall prevail .
最后,应当理解的是,本申请中所述实施例仅用以说明本申请实施例的原则。其他的变形也可能属于本申请的范围。因此,作为示例而非限制,本申请实施例的替代配置可视为 与本申请的教导一致。相应地,本申请的实施例不仅限于本申请明确介绍和描述的实施例。Finally, it should be understood that the embodiments described in this application are only used to illustrate the principles of the embodiments of this application. Other variations may also fall within the scope of this application. Therefore, by way of example and not limitation, alternative configurations of embodiments of the present application may be considered consistent with the teachings of the present application. Accordingly, the embodiments of the present application are not limited to the embodiments explicitly introduced and described in the present application.
本领域技术人员应明白,本申请的实施例可提供为方法、系统或计算机程序产品。因此,本申请可采用完全硬件实施例、完全软件实施例或结合软件和硬件方面的实施例的形式。而且,本申请可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。Those skilled in the art should understand that the embodiments of the present application may be provided as a method, a system, or a computer program product. Therefore, this application may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Moreover, this application may take the form of a computer program product implemented on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, etc.) containing computer-usable program code.
以上所述为本申请的基本构思,仅以实施例形式呈现,显而易见地,本领域的技术人员依据本申请作出相应变化、改进或修正。这些变化、改进和修正已被本申请所暗示或间接提出,均包含在本申请实施例的精神或范围之内。The above is the basic idea of the present application, which is only presented in the form of an embodiment. Obviously, those skilled in the art can make corresponding changes, improvements or amendments according to the present application. These changes, improvements, and amendments have been implicitly or indirectly proposed by this application, and are all included within the spirit or scope of the embodiments of this application.

Claims (28)

  1. 一种图像数据处理方法,其特征在于,包括:An image data processing method, comprising:
    获取包含道路的图像;Get images containing roads;
    基于所述图像和经训练的交通标线识别模型,获得所述图像的至少一个图层;所述图层包括所述图像中的至少一个交通标线,以及反映所述至少一个交通标线在所述图像中的区域;Obtaining at least one layer of the image based on the image and the trained traffic marking recognition model; the layer includes at least one traffic marking in the image, and reflecting that the at least one traffic marking is in the A region in the image;
    对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标。And processing the at least one layer to obtain position coordinates of the at least one traffic line in the image.
  2. 如权利要求1所述的方法,其特征在于,所述基于所述图像和所述经训练的交通标线识别模型,获得所述图像的至少一个图层还包括:The method of claim 1, wherein obtaining at least one layer of the image based on the image and the trained traffic marking recognition model further comprises:
    基于所述图像,获取多张子图像;所述多张子图像分别包括所述图像的一部分,且至少两张子图像共同包括所述图像中的某一部分;Obtaining a plurality of sub-images based on the image; each of the plurality of sub-images including a part of the image, and at least two of the sub-images collectively including a part of the image;
    针对每张子图像,For each sub-image,
    基于所述子图像和所述经训练的交通标线识别模型,获得所述子图像的至少一个图层;所述子图像的至少一个图层包括所述子图像中的至少一个交通标线的至少一部分,以及至少反映所述至少一个交通标线的至少一部分在所述子图像中的区域;Obtaining at least one layer of the sub-image based on the sub-image and the trained traffic marking recognition model; at least one layer of the sub-image includes at least one traffic marking in the sub-image At least a portion, and at least a region in the sub-image that reflects at least a portion of the at least one traffic line;
    基于至少两张子图像的至少一个图层,联合确定出所述图像的至少一个图层。Based on at least one layer of at least two sub-images, at least one layer of the images is jointly determined.
  3. 如权利要求1所述的方法,其特征在于,所述经训练的交通标线识别模型为经训练的MASK-RCNN模型。The method of claim 1, wherein the trained traffic marking recognition model is a trained MASK-RCNN model.
  4. 如权利要求1所述的方法,其特征在于,所述图层的图像大小与所述图像的图像大小相同,且所述图层为二值化图像。The method according to claim 1, wherein the image size of the layer is the same as the image size of the image, and the layer is a binary image.
  5. 如权利要求1所述的方法,其特征在于,所述对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标还包括:The method of claim 1, wherein processing the at least one layer to obtain position coordinates of the at least one traffic line in the image further comprises:
    对所述至少一个图层进行腐蚀处理、膨胀处理或平滑处理中的一种或多种的组合。The at least one layer is subjected to one or more combinations of an etching process, an expansion process, or a smoothing process.
  6. 如权利要求1所述的方法,其特征在于,所述图像为包括道路的俯视图像。The method of claim 1, wherein the image is a top-down image including a road.
  7. 如权利要求6所述的方法,其特征在于,所述获取包含道路的图像还包括:The method according to claim 6, wherein the acquiring an image including a road further comprises:
    获取车载设备拍摄的道路视频;Obtain road videos taken by vehicle equipment;
    基于所述道路视频,获取多张图像;Acquiring multiple images based on the road video;
    获取每张图像中同一位置的图像数据;Obtain image data at the same position in each image;
    拼接所述多张图像中同一位置的图像数据,获得所述道路的俯视图像。The image data of the same position in the multiple images are stitched to obtain a top-view image of the road.
  8. 如权利要求7所述的方法,其特征在于,所述每张图像中同一位置的图像数据包括每张图像中相同的至少一行的图像数据。The method according to claim 7, wherein the image data of the same position in each image includes image data of at least one line that is the same in each image.
  9. 如权利要求1所述的方法,其特征在于,所述图层还反映所述至少一个交通标线的类别。The method of claim 1, wherein the layer further reflects a category of the at least one traffic line.
  10. 如权利要求9所述的方法,其特征在于,响应于图像的图层为车道线图层,所述对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标还包括:The method of claim 9, wherein the layer responsive to the image is a lane line layer, and the at least one layer is processed to obtain the at least one traffic line in the image The location coordinates also include:
    基于所述车道线图层和所述车道线图层中像素点的像素值,确定至少一条车道线的左边缘点和右边缘点在所述车道线图层中的位置坐标;Determining position coordinates of the left edge point and the right edge point of at least one lane line in the lane line layer based on pixel values of the lane line layer and pixel points in the lane line layer;
    基于所述至少一条车道线的左边缘点和右边缘点在所述车道线图层中的位置坐标,确定所述至少一条车道线的中心线在所述车道线图层中的位置坐标;Determining position coordinates of a center line of the at least one lane line in the lane line layer based on position coordinates of a left edge point and a right edge point of the at least one lane line in the lane line layer;
    基于车道线的预设宽度以及所述至少一条车道线的中心线的位置坐标,确定所述至少一条车道线在所述图像中的位置坐标。A position coordinate of the at least one lane line in the image is determined based on a preset width of the lane line and a position coordinate of a center line of the at least one lane line.
  11. 如权利要求9所述的方法,其特征在于,响应于图像的图层为车道线图层,所述对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标还包括:The method of claim 9, wherein the layer responsive to the image is a lane line layer, and the at least one layer is processed to obtain the at least one traffic line in the image The location coordinates also include:
    判断车道线图层中至少一条车道线的某一端点和另一条车道线的某一端点 在水平方向的距离是否小于第一阈值,并且在竖直方向上的距离是否小于第二阈值;Judging whether a distance between a certain end point of at least one lane line and another end point of another lane line in a lane line layer in a horizontal direction is less than a first threshold value, and whether a distance in a vertical direction is less than a second threshold value;
    响应于所述至少一条车道线的某一端点和另一条车道线的某一端点在水平方向的距离小于第一阈值,并且在竖直方向上的距离小于第二阈值时,则将所述至少一条道路的所述某一端点和所述另一条车道线的所述某一端点进行拼接。In response to a distance in a horizontal direction between a certain end point of the at least one lane line and a certain end point of another lane line being less than a first threshold value, and when the distance in the vertical direction is less than a second threshold value, The one end point of one road and the one end point of the other lane line are spliced.
  12. 如权利要求9或11所述的方法,其特征在于,响应于图像的图层为车道线图层,所述对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标还包括:The method according to claim 9 or 11, wherein the layer responsive to the image is a lane line layer, and the at least one layer is processed to obtain the at least one traffic marking line at the The position coordinates of the image also include:
    判断车道线图层中某条车道线的长度是否小于第三阈值;Determine whether the length of a lane line in the lane line layer is less than a third threshold;
    响应于所述某条车道线的长度小于第三阈值,将所述车道线从所述图像中去除。In response to the length of a certain lane line being less than a third threshold, the lane line is removed from the image.
  13. 如权利要求1所述的方法,其特征在于,所述对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标还包括:The method of claim 1, wherein processing the at least one layer to obtain position coordinates of the at least one traffic line in the image further comprises:
    基于所述至少一个图层和该图层中像素点的像素值,确定该图层中交通标线在水平方向上的最大坐标值和最小坐标值,以及在竖直方向上的最大坐标值和最小坐标值;Based on the at least one layer and the pixel values of the pixel points in the layer, determine the maximum coordinate value and the minimum coordinate value of the traffic markings in the layer in the horizontal direction, and the maximum coordinate value and Minimum coordinate value
    基于所述交通标线在水平方向上的最大坐标值和最小坐标值以及在竖直方向上的最大坐标值和最小坐标值,确定该交通标线在所述图像中的位置坐标。Based on the maximum coordinate value and the minimum coordinate value in the horizontal direction of the traffic marking line and the maximum coordinate value and the minimum coordinate value in the vertical direction, the position coordinates of the traffic marking line in the image are determined.
  14. 一种图像数据处理系统,其特征在于,所述系统包括:An image data processing system, characterized in that the system includes:
    用于存储计算机指令的至少一个存储器;At least one memory for storing computer instructions;
    与所述存储器通讯的至少一个处理器,其中当所述至少一个处理器执行所述计算机指令时,所述至少一个处理器使所述系统执行:At least one processor in communication with the memory, wherein when the at least one processor executes the computer instructions, the at least one processor causes the system to execute:
    获取包含道路的图像;Get images containing roads;
    基于所述图像和经训练的交通标线识别模型,获得所述图像的至少一个图层;所述图层包括所述图像中的至少一个交通标线,以及反映所述至少一个交通标线 在所述图像中的区域;Obtaining at least one layer of the image based on the image and the trained traffic marking recognition model; the layer includes at least one traffic marking in the image, and reflecting that the at least one traffic marking is in the A region in the image;
    对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标。And processing the at least one layer to obtain position coordinates of the at least one traffic line in the image.
  15. 如权利要求14所述的系统,其特征在于,为基于所述图像和所述经训练的交通标线识别模型,获得所述图像的至少一个图层,所述至少一个处理器使所述系统进一步执行:The system of claim 14, wherein to obtain at least one layer of the image based on the image and the trained traffic marking recognition model, the at least one processor causes the system to Further implementation:
    基于所述图像,获取多张子图像;所述多张子图像分别包括所述图像的一部分,且至少两张子图像共同包括所述俯视图像中的某一部分;Obtaining a plurality of sub-images based on the image; each of the plurality of sub-images including a part of the image, and at least two of the sub-images collectively including a part of the overhead image;
    针对每张子图像,For each sub-image,
    基于所述子图像和所述经训练的交通标线识别模型,获得所述子图像的至少一个图层;所述子图像的至少一个图层包括所述子图像中的至少一个交通标线的至少一部分,以及至少反映所述至少一个交通标线的至少一部分在所述子图像中的区域;Obtaining at least one layer of the sub-image based on the sub-image and the trained traffic marking recognition model; at least one layer of the sub-image includes at least one traffic marking in the sub-image At least a portion, and at least a region in the sub-image that reflects at least a portion of the at least one traffic line;
    基于至少两张子图像的至少一个图层,联合确定出所述图像的至少一个图层。Based on at least one layer of at least two sub-images, at least one layer of the images is jointly determined.
  16. 如权利要求14所述的系统,其特征在于,所述经训练的交通标线识别模型为经训练的MASK-RCNN模型。The system according to claim 14, wherein the trained traffic marking recognition model is a trained MASK-RCNN model.
  17. 如权利要求14所述的系统,其特征在于,所述图层的图像大小与所述图像的图像大小相同,且所述图层为二值化图像。The system according to claim 14, wherein the image size of the layer is the same as the image size of the image, and the layer is a binary image.
  18. 如权利要求14所述的系统,其特征在于,为对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标,所述至少一个处理器使所述系统进一步执行:The system of claim 14, wherein in order to process the at least one layer to obtain position coordinates of the at least one traffic line in the image, the at least one processor causes the at least one processor to The system further executes:
    对所述至少一个图层进行腐蚀处理、膨胀处理或平滑处理中的一种或多种的组合。The at least one layer is subjected to one or more combinations of an etching process, an expansion process, or a smoothing process.
  19. 如权利要求14所述的系统,其特征在于,所述图像包括道路的俯视图像。The system of claim 14, wherein the image comprises a top-view image of a road.
  20. 如权利要求19所述的系统,其特征在于,为获取包含道路的图像,所述至少一个处理器使所述系统进一步执行:The system of claim 19, wherein in order to obtain an image containing a road, the at least one processor causes the system to further execute:
    获取车载设备拍摄的道路视频;Obtain road videos taken by vehicle equipment;
    基于所述道路视频,获取多张图像;Acquiring multiple images based on the road video;
    获取每张图像中同一位置的图像数据;Obtain image data at the same position in each image;
    拼接所述多张图像中同一位置的图像数据,获得所述道路的俯视图像。The image data of the same position in the multiple images are stitched to obtain a top-view image of the road.
  21. 如权利要求20所述的系统,其特征在于,所述每张图像中同一位置的图像数据包括每张图像中相同的至少一行的图像数据。The system according to claim 20, wherein the image data of the same position in each image includes image data of at least one line that is the same in each image.
  22. 如权利要求14所述的系统,其特征在于,所述图层还反映所述至少一个交通标线的类别。The system of claim 14, wherein the layer further reflects a category of the at least one traffic line.
  23. 如权利要求22所述的系统,其特征在于,响应于图像的图层为车道线图层,为对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标,所述至少一个处理器使所述系统进一步执行:The system according to claim 22, wherein the layer in response to the image is a lane line layer, and the at least one layer is processed to obtain the at least one traffic line in the image. Position coordinates, the at least one processor causes the system to further execute:
    基于所述车道线图层和所述车道线图层中像素点的像素值,确定至少一条车道线的左边缘点和右边缘点在所述车道线图层中的位置坐标;Determining position coordinates of the left edge point and the right edge point of at least one lane line in the lane line layer based on the pixel values of the lane line layer and the pixel points in the lane line layer;
    基于所述至少一条车道线的左边缘点和右边缘点在所述车道线图层中的位置坐标,确定所述至少一条车道线的中心线在所述车道线图层中的位置坐标;Determining position coordinates of a center line of the at least one lane line in the lane line layer based on position coordinates of a left edge point and a right edge point of the at least one lane line in the lane line layer;
    基于车道线的预设宽度以及所述至少一条车道线的中心线的位置坐标,确定所述至少一条车道线在所述图像中的位置坐标。A position coordinate of the at least one lane line in the image is determined based on a preset width of the lane line and a position coordinate of a center line of the at least one lane line.
  24. 如权利要求22所述的系统,其特征在于,响应于图像的图层为车道线图层,为对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标,所述至少一个处理器使所述系统进一步执行:The system according to claim 22, wherein the layer in response to the image is a lane line layer, and the at least one layer is processed to obtain the at least one traffic line in the image. Position coordinates, the at least one processor causes the system to further execute:
    判断车道线图层中至少一条车道线的某一端点和另一条车道线的某一端点在水平方向的距离是否小于第一阈值,并且在竖直方向上的距离是否小于第二阈 值;Judging whether a distance between a certain end point of at least one lane line and another end point of another lane line in a lane line layer in a horizontal direction is less than a first threshold, and whether a distance in a vertical direction is less than a second threshold;
    响应于所述至少一条车道线的某一端点和另一条车道线的某一端点在水平方向的距离小于第一阈值,并且在竖直方向上的距离小于第二阈值时,则将所述至少一条道路的所述某一端点和所述另一条车道线的所述某一端点进行拼接。In response to a distance in a horizontal direction between a certain end point of the at least one lane line and a certain end point of another lane line being less than a first threshold value, and when the distance in the vertical direction is less than a second threshold value, The one end point of one road and the one end point of the other lane line are spliced.
  25. 如权利要求22或24所述的系统,其特征在于,响应于图像的图层为车道线图层,为对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标,所述至少一个处理器使所述系统进一步执行:The system according to claim 22 or 24, wherein the layer responsive to the image is a lane line layer, and the at least one layer is processed to obtain the at least one traffic marking line in the Position coordinates of the image, the at least one processor causes the system to further execute:
    判断车道线图层中某条车道线的长度是否小于第三阈值;Determine whether the length of a lane line in the lane line layer is less than a third threshold;
    响应于所述某条车道线的长度小于第三阈值,将所述车道线从所述图像中去除。In response to the length of a certain lane line being less than a third threshold, the lane line is removed from the image.
  26. 如权利要求14所述的系统,其特征在于,为对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标,所述至少一个处理器使所述系统进一步执行:The system of claim 14, wherein in order to process the at least one layer to obtain position coordinates of the at least one traffic line in the image, the at least one processor causes the at least one processor to The system further executes:
    基于所述至少一个图层和该图层中像素点的像素值,确定该图层中交通标线在水平方向上的最大坐标值和最小坐标值,以及在竖直方向上的最大坐标值和最小坐标值;Based on the at least one layer and the pixel values of the pixel points in the layer, determine the maximum coordinate value and the minimum coordinate value of the traffic markings in the layer in the horizontal direction, and the maximum coordinate value and Minimum coordinate value
    基于所述交通标线在水平方向上的最大坐标值和最小坐标值以及在竖直方向上的最大坐标值和最小坐标值,确定该交通标线在所述图像中的位置坐标。Based on the maximum coordinate value and the minimum coordinate value in the horizontal direction of the traffic marking line and the maximum coordinate value and the minimum coordinate value in the vertical direction, the position coordinates of the traffic marking line in the image are determined.
  27. 一种图像数据处理系统,其特征在于,包括:An image data processing system, comprising:
    图像获取模块,用于获取包含道路的图像;An image acquisition module, for acquiring an image containing a road;
    图层获取模块,用于基于所述图像和经训练的交通标线识别模型,获得所述图像的至少一个图层;所述图层包括所述图像中的至少一个交通标线,以及反映所述至少一个交通标线在所述图像中的区域;A layer acquisition module, configured to obtain at least one layer of the image based on the image and the trained traffic line recognition model; the layer includes at least one traffic line in the image, and reflects the The area of the at least one traffic line in the image;
    图层处理模块,用于对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标。A layer processing module is configured to process the at least one layer to obtain position coordinates of the at least one traffic marking in the image.
  28. 一种计算机可读存储介质,其特征在于,所述存储介质存储计算机指令,当计算机读取存储介质中的计算机指令后,计算机执行方法,所述方法包括:A computer-readable storage medium, characterized in that the storage medium stores computer instructions, and after the computer reads the computer instructions in the storage medium, the computer executes the method, and the method includes:
    获取包含道路的图像;Get images containing roads;
    基于所述图像和经训练的交通标线识别模型,获得所述图像的至少一个图层;所述图层包括所述图像中的至少一个交通标线,以及反映所述至少一个交通标线在所述图像中的区域;Obtaining at least one layer of the image based on the image and the trained traffic marking recognition model; the layer includes at least one traffic marking in the image, and reflecting that the at least one traffic marking is in the A region in the image;
    对所述至少一个图层进行处理,以获得所述至少一个交通标线在所述图像的位置坐标。And processing the at least one layer to obtain position coordinates of the at least one traffic line in the image.
PCT/CN2019/104415 2018-09-05 2019-09-04 Image data processing method and system WO2020048487A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201811031784.9A CN110879943B (en) 2018-09-05 2018-09-05 Image data processing method and system
CN201811031784.9 2018-09-05

Publications (1)

Publication Number Publication Date
WO2020048487A1 true WO2020048487A1 (en) 2020-03-12

Family

ID=69721537

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/104415 WO2020048487A1 (en) 2018-09-05 2019-09-04 Image data processing method and system

Country Status (2)

Country Link
CN (1) CN110879943B (en)
WO (1) WO2020048487A1 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111815740A (en) * 2020-07-27 2020-10-23 城云科技(中国)有限公司 Map drawing method, system, terminal and storage medium
CN112163475A (en) * 2020-09-15 2021-01-01 北京三快在线科技有限公司 Method and device for determining lane line direction
US20210319237A1 (en) * 2020-04-10 2021-10-14 Thinkware Corporation Method, apparatus, electronic device, computer program and computer-readable recording medium for detecting lane marking based on vehicle image
CN113780067A (en) * 2021-07-30 2021-12-10 武汉中海庭数据技术有限公司 Lane linear marker detection method and system based on semantic segmentation
CN115797377A (en) * 2023-02-09 2023-03-14 山东省鲁南地质工程勘察院(山东省地质矿产勘查开发局第二地质大队) Data processing system of satellite remote sensing image
CN116071654A (en) * 2023-02-16 2023-05-05 阿里巴巴(中国)有限公司 Map data processing method, device, equipment and medium
GB2616597A (en) * 2022-03-10 2023-09-20 Continental Automotive Tech Gmbh Method and system for identifying rows of objects in images

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111814651B (en) * 2020-07-02 2024-01-12 阿波罗智能技术(北京)有限公司 Lane line generation method, device and equipment
CN114489829B (en) * 2021-12-22 2023-04-18 北京市遥感信息研究所 Remote sensing image sample labeling method based on ArcMap

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105069415A (en) * 2015-07-24 2015-11-18 深圳市佳信捷技术股份有限公司 Lane line detection method and device
CN105260699A (en) * 2015-09-10 2016-01-20 百度在线网络技术(北京)有限公司 Lane line data processing method and lane line data processing device
CN105718870A (en) * 2016-01-15 2016-06-29 武汉光庭科技有限公司 Road marking line extracting method based on forward camera head in automatic driving
CN108470159A (en) * 2018-03-09 2018-08-31 腾讯科技(深圳)有限公司 Lane line data processing method, device, computer equipment and storage medium

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103700081B (en) * 2013-12-17 2016-08-17 河海大学 A kind of shredder crushes the restoration methods of English document
CN103942959B (en) * 2014-04-22 2016-08-24 深圳市宏电技术股份有限公司 A kind of lane detection method and device
CN104281841A (en) * 2014-09-30 2015-01-14 深圳市汇顶科技股份有限公司 Fingerprint identification system and fingerprint processing method and device thereof
CN104766058B (en) * 2015-03-31 2018-04-27 百度在线网络技术(北京)有限公司 A kind of method and apparatus for obtaining lane line
CN105512657B (en) * 2015-08-20 2019-04-30 北京旷视科技有限公司 Character identifying method and equipment
CN105426861B (en) * 2015-12-02 2019-05-21 百度在线网络技术(北京)有限公司 Lane line determines method and device
CN106919915B (en) * 2017-02-22 2020-06-12 武汉极目智能技术有限公司 Map road marking and road quality acquisition device and method based on ADAS system
CN107169415B (en) * 2017-04-13 2019-10-11 西安电子科技大学 Human motion recognition method based on convolutional neural networks feature coding
CN107292229A (en) * 2017-05-08 2017-10-24 北京三快在线科技有限公司 A kind of image-recognizing method and device
CN108009524B (en) * 2017-12-25 2021-07-09 西北工业大学 Lane line detection method based on full convolution network
CN108297867B (en) * 2018-02-11 2019-12-03 江苏金羿智芯科技有限公司 A kind of lane departure warning method and system based on artificial intelligence

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105069415A (en) * 2015-07-24 2015-11-18 深圳市佳信捷技术股份有限公司 Lane line detection method and device
CN105260699A (en) * 2015-09-10 2016-01-20 百度在线网络技术(北京)有限公司 Lane line data processing method and lane line data processing device
CN105718870A (en) * 2016-01-15 2016-06-29 武汉光庭科技有限公司 Road marking line extracting method based on forward camera head in automatic driving
CN108470159A (en) * 2018-03-09 2018-08-31 腾讯科技(深圳)有限公司 Lane line data processing method, device, computer equipment and storage medium

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210319237A1 (en) * 2020-04-10 2021-10-14 Thinkware Corporation Method, apparatus, electronic device, computer program and computer-readable recording medium for detecting lane marking based on vehicle image
US11727693B2 (en) * 2020-04-10 2023-08-15 Thinkware Corporation Method, apparatus, electronic device, computer program and computer-readable recording medium for detecting lane marking based on vehicle image
CN111815740A (en) * 2020-07-27 2020-10-23 城云科技(中国)有限公司 Map drawing method, system, terminal and storage medium
CN111815740B (en) * 2020-07-27 2024-01-30 城云科技(中国)有限公司 Map drawing method, system, terminal and storage medium
CN112163475A (en) * 2020-09-15 2021-01-01 北京三快在线科技有限公司 Method and device for determining lane line direction
CN113780067A (en) * 2021-07-30 2021-12-10 武汉中海庭数据技术有限公司 Lane linear marker detection method and system based on semantic segmentation
GB2616597A (en) * 2022-03-10 2023-09-20 Continental Automotive Tech Gmbh Method and system for identifying rows of objects in images
CN115797377A (en) * 2023-02-09 2023-03-14 山东省鲁南地质工程勘察院(山东省地质矿产勘查开发局第二地质大队) Data processing system of satellite remote sensing image
CN116071654A (en) * 2023-02-16 2023-05-05 阿里巴巴(中国)有限公司 Map data processing method, device, equipment and medium
CN116071654B (en) * 2023-02-16 2023-08-08 阿里巴巴(中国)有限公司 Map data processing method, device, equipment and medium

Also Published As

Publication number Publication date
CN110879943B (en) 2022-08-19
CN110879943A (en) 2020-03-13

Similar Documents

Publication Publication Date Title
WO2020048487A1 (en) Image data processing method and system
EP3462377B1 (en) Method and apparatus for identifying driving lane
US10599930B2 (en) Method and apparatus of detecting object of interest
US11688183B2 (en) System and method of determining a curve
US20200293058A1 (en) Data processing method, apparatus and terminal
JP6855350B2 (en) Head-up display device, navigation device, display method
CN111874006B (en) Route planning processing method and device
KR102675523B1 (en) Method and apparatus of determining lane
US20200064855A1 (en) Method and apparatus for determining road line
CN106064587B (en) Multi-lane vehicle distance recognition method and device based on license plate recognition
JP2012185011A (en) Mobile position measuring apparatus
WO2013018708A1 (en) Vehicle detection device and vehicle detection method
CN111091037A (en) Method and device for determining driving information
JP2009199284A (en) Road object recognition method
CN113868356A (en) Rendering method, rendering apparatus, storage medium, and computer program
KR102564856B1 (en) Method and apparatus of detecting road line
JP2017130155A (en) Object recognition device and object recognition method
US20110069087A1 (en) Method of and computer implemented system for generating a junction view image
CN111931683A (en) Image recognition method, image recognition device and computer-readable storage medium
CN110733416A (en) lane departure early warning method based on inverse perspective transformation
WO2023179030A1 (en) Road boundary detection method and apparatus, and electronic device, storage medium and computer program product
WO2020093351A1 (en) Systems and methods for identifying a road feature
Irshad et al. Real-time lane departure warning system on a lower resource platform
Al Noman et al. A Computer Vision-Based Lane Detection Approach for an Autonomous Vehicle Using the Image Hough Transformation and the Edge Features
JP7405643B2 (en) Sign identification system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19858600

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19858600

Country of ref document: EP

Kind code of ref document: A1