WO2020135730A1 - 车载控制单元、基于fpga的车辆自动驾驶方法及装置 - Google Patents
车载控制单元、基于fpga的车辆自动驾驶方法及装置 Download PDFInfo
- Publication number
- WO2020135730A1 WO2020135730A1 PCT/CN2019/129248 CN2019129248W WO2020135730A1 WO 2020135730 A1 WO2020135730 A1 WO 2020135730A1 CN 2019129248 W CN2019129248 W CN 2019129248W WO 2020135730 A1 WO2020135730 A1 WO 2020135730A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- soc
- fpga
- arm
- mcu
- vehicle
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 76
- 230000008447 perception Effects 0.000 claims abstract description 82
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 50
- 230000008569 process Effects 0.000 claims abstract description 47
- 238000013528 artificial neural network Methods 0.000 claims abstract description 42
- 230000016776 visual perception Effects 0.000 claims description 13
- 230000001360 synchronised effect Effects 0.000 claims description 4
- 230000010354 integration Effects 0.000 claims description 3
- 230000007246 mechanism Effects 0.000 abstract description 11
- 230000006870 function Effects 0.000 description 9
- 238000012545 processing Methods 0.000 description 9
- 238000013461 design Methods 0.000 description 8
- 238000010586 diagram Methods 0.000 description 8
- 230000004927 fusion Effects 0.000 description 8
- 238000013527 convolutional neural network Methods 0.000 description 7
- 239000000306 component Substances 0.000 description 6
- 230000003287 optical effect Effects 0.000 description 4
- 238000004590 computer program Methods 0.000 description 3
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000013135 deep learning Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 238000004891 communication Methods 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 238000005259 measurement Methods 0.000 description 2
- 210000004556 brain Anatomy 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 210000003169 central nervous system Anatomy 0.000 description 1
- 239000008358 core component Substances 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 238000000802 evaporation-induced self-assembly Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/183—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a single remote source
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F30/00—Computer-aided design [CAD]
- G06F30/30—Circuit design
- G06F30/32—Circuit design at the digital level
- G06F30/33—Design verification, e.g. functional simulation or model checking
- G06F30/3308—Design verification, e.g. functional simulation or model checking using simulation
- G06F30/331—Design verification, e.g. functional simulation or model checking using simulation with hardware acceleration, e.g. by using field programmable gate array [FPGA] or emulation
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W60/00—Drive control systems specially adapted for autonomous road vehicles
- B60W60/001—Planning or execution of driving tasks
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/86—Combinations of radar systems with non-radar systems, e.g. sonar, direction finder
- G01S13/867—Combination of radar systems with cameras
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/88—Radar or analogous systems specially adapted for specific applications
- G01S13/93—Radar or analogous systems specially adapted for specific applications for anti-collision purposes
- G01S13/931—Radar or analogous systems specially adapted for specific applications for anti-collision purposes of land vehicles
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S15/00—Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
- G01S15/86—Combinations of sonar systems with lidar systems; Combinations of sonar systems with systems not using wave reflection
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S15/00—Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
- G01S15/88—Sonar systems specially adapted for specific applications
- G01S15/93—Sonar systems specially adapted for specific applications for anti-collision purposes
- G01S15/931—Sonar systems specially adapted for specific applications for anti-collision purposes of land vehicles
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course or altitude of land, water, air, or space vehicles, e.g. automatic pilot
- G05D1/0088—Control of position, course or altitude of land, water, air, or space vehicles, e.g. automatic pilot characterized by the autonomous decision making process, e.g. artificial intelligence, predefined behaviours
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course or altitude of land, water, air, or space vehicles, e.g. automatic pilot
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0231—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
- G05D1/0246—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using a video camera in combination with image processing means
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course or altitude of land, water, air, or space vehicles, e.g. automatic pilot
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0231—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
- G05D1/0246—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using a video camera in combination with image processing means
- G05D1/0251—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using a video camera in combination with image processing means extracting 3D information from a plurality of images taken from different locations, e.g. stereo vision
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course or altitude of land, water, air, or space vehicles, e.g. automatic pilot
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0255—Control of position or course in two dimensions specially adapted to land vehicles using acoustic signals, e.g. ultra-sonic singals
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course or altitude of land, water, air, or space vehicles, e.g. automatic pilot
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0257—Control of position or course in two dimensions specially adapted to land vehicles using a radar
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/06—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
- G06N3/063—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/183—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a single remote source
- H04N7/185—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a single remote source from a mobile camera, e.g. for remote control
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2420/00—Indexing codes relating to the type of sensors based on the principle of their operation
- B60W2420/40—Photo or light sensitive means, e.g. infrared sensors
- B60W2420/403—Image sensing, e.g. optical camera
-
- B60W2420/408—
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/93—Lidar systems specially adapted for specific applications for anti-collision purposes
- G01S17/931—Lidar systems specially adapted for specific applications for anti-collision purposes of land vehicles
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/88—Radar or analogous systems specially adapted for specific applications
- G01S13/93—Radar or analogous systems specially adapted for specific applications for anti-collision purposes
- G01S13/931—Radar or analogous systems specially adapted for specific applications for anti-collision purposes of land vehicles
- G01S2013/9323—Alternative operation using light waves
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/88—Radar or analogous systems specially adapted for specific applications
- G01S13/93—Radar or analogous systems specially adapted for specific applications for anti-collision purposes
- G01S13/931—Radar or analogous systems specially adapted for specific applications for anti-collision purposes of land vehicles
- G01S2013/9324—Alternative operation using ultrasonic waves
Definitions
- the embodiments of the present application relate to the technical field of vehicle control, and in particular, to an on-board control unit and an FPGA-based vehicle automatic driving method and device.
- Self-driving vehicles also known as unmanned vehicles, computer-driven vehicles, or wheeled mobile robots, are intelligent vehicles that realize unmanned driving through computer systems.
- An electronic control unit is installed on the self-driving car.
- the ECU is also called a trip computer and an on-board computer, and is the "brain" of an unmanned vehicle.
- the on-board camera is used to collect the surrounding information of the automatic driving vehicle, and the collected information is sent to the ECU.
- the ECU uses a perception algorithm to visually recognize the received information to generate a decision result, and generates a control command according to the decision result, and then based on The control command completes automatic driving.
- the visual perception of the information collected by the on-board camera using the perception algorithm relies heavily on the ECU, which causes the ECU to increase the burden.
- Embodiments of the present application provide an on-board control unit, an FPGA-based vehicle automatic driving method and device.
- the first SoC is integratedly configured by FPGA and ARM, and the sensor data is processed by FPAG and ARM and sent to the MCU to reduce the burden of the MCU.
- an embodiment of the present application provides an FPGA-based vehicle automatic driving method, which is suitable for a vehicle-mounted control unit.
- the vehicle-mounted control unit includes a first system-on-chip SoC and a micro-control unit MCU.
- the first SoC uses an FPGA It is integrated with the advanced reduced instruction set machine ARM.
- the vehicle-mounted control unit is installed on an autonomous vehicle. The method includes:
- the FPGA of the first SoC receives the video data sent by the vehicle camera
- the FPGA of the first SoC uses a first neural network algorithm to perform visual perception on the video data to obtain first perception information
- the FPGA of the first SoC sends the first perception information to the ARM of the first SoC;
- the ARM of the first SoC processes the first perception information, obtains first decision information, and sends it to the MCU.
- the ARM of the first SoC processes the first perception information to obtain the first decision information and sends it to the MCU, including:
- the ARM of the first SoC receives radar data
- the ARM of the first SoC fuses the first perception information and the radar data
- the ARM of the first SoC processes the merged first perception information and the radar data to obtain the first decision information and sends it to the MCU.
- the radar data includes at least one of ultrasonic radar data, millimeter wave radar data, and lidar data.
- the vehicle-mounted control unit further includes: a second system-on-chip SoC, and the second SoC is integratedly set through an FPGA and an advanced reduced instruction set machine ARM, and the method further includes:
- the FPGA of the second SoC receives the video data sent by the vehicle camera
- the FPGA of the second SoC uses a second neural network algorithm to perform visual perception on the video data to obtain second perception information
- the FPGA of the second SoC sends the second perception information to the ARM of the second SoC;
- the ARM of the second SoC processes the second perception information to obtain second decision information and sends it to the MCU;
- the MCU generates a control instruction according to the first decision information and the second decision information.
- an embodiment of the present application provides a vehicle-mounted control unit, including: a first system-on-chip SoC and a micro control unit MCU, the first SoC is integrated with a field programmable gate array FPGA and an advanced reduced instruction set machine ARM Set, the first SoC and the MCU are connected through an Ethernet switching chip.
- the above-mentioned vehicle-mounted control unit further includes: a second system-on-chip SoC, where an FPGA and an ARM are provided on the second SoC, and the FPGA and the ARM on the second SoC are connected by a bus, and the second The SoC and the MCU are connected through the Ethernet switching chip.
- a second system-on-chip SoC where an FPGA and an ARM are provided on the second SoC, and the FPGA and the ARM on the second SoC are connected by a bus, and the second The SoC and the MCU are connected through the Ethernet switching chip.
- the vehicle-mounted control unit further includes: a first synchronous dynamic random access memory SDRAM and a first flash memory Flash, the first SDRAM is connected to the first SoC, and the first Flash is connected to the first One SoC connection.
- an embodiment of the present application provides an automatic driving device, which is suitable for a vehicle-mounted control unit.
- the automatic driving device includes: a first system-on-chip SoC module and a micro control unit MCU module, and the first SoC module Through the integrated setting of FPGA and advanced reduced instruction set machine ARM, the vehicle-mounted control unit is set on an autonomous vehicle, and the first SoC includes a first FPGA unit and a first ARM unit, wherein,
- the first FPGA unit is configured to receive video data sent by an on-board camera, perform visual perception on the video data using a first neural network algorithm, obtain first perception information, and send the first sensor to the ARM of the first SoC A perceived information;
- the first ARM unit is configured to process the first perception information, obtain first decision information, and send it to the MCU module.
- the first ARM unit is specifically used to receive radar data, fuse the first perception information and the radar data, and perform the fusion on the first perception information and the radar data Processing to obtain the first decision information and send it to the MCU.
- the device further includes: a second SoC module, the second SoC module includes a second FPGA unit and a second ARM unit, wherein,
- the second FPGA unit is configured to receive the video data sent by the on-board camera, and perform visual perception on the video data using a second neural network algorithm to obtain second perception information and send it to the ARM of the second SoC Sending the second perception information;
- the second ARM unit is configured to process the second perception information, obtain second decision information, and send it to the MCU;
- the MCU module is configured to generate a control instruction according to the first decision information and the second decision information.
- the vehicle-mounted control unit and the FPGA-based vehicle automatic driving method and device provided in the embodiments of the present application.
- the vehicle-mounted control unit includes an MCU and a first SoC implemented by FPGA and ARM integrated settings.
- the vehicle-mounted control unit is provided on an automatic driving vehicle.
- the FPGA of the first SoC receives the video data sent by the on-board sensors, uses the first neural network algorithm to visually perceive the video data to obtain the first perception information, and sends the first perception information to the ARM of the first SoC,
- the ARM of the first SoC processes the first perception information to obtain the first decision information and sends it to the MCU.
- the MCU generates a control instruction according to the first decision information and sends it to the corresponding execution mechanism, so that the control mechanism automatically performs the control according to the control execution. drive.
- the first SoC is integrated with FPGA and ARM, and the sensor data is processed by FPAG and ARM and sent to the MCU to reduce the burden on the MCU.
- Figure 1 is an ECU based on FPGA implemented in the prior art
- FIG. 2 is a schematic structural diagram of a vehicle-mounted control unit provided by an embodiment of the present application.
- FIG. 3 is a flowchart of an automatic driving method based on a field programmable gate array FPGA provided by an embodiment of the present application;
- FIG. 4 is a schematic structural diagram of another vehicle-mounted control unit provided by an embodiment of the present application.
- FIG. 5 is a schematic structural diagram of an automatic driving device provided by an embodiment of the present application.
- FIG. 6 is a schematic structural diagram of another automatic driving device provided by an embodiment of the present application.
- On-board control unit also called electronic control unit (ECU, ECU)
- ECU electronice control unit
- ECU electronice control unit
- Multiple ECUs are installed on a car to be responsible for different functions. Information exchange is possible, and multiple ECUs form the control system of the car.
- an ECU includes a micro controller unit (MCU), input and output interfaces, etc.
- MCU micro controller unit
- the ECU and other electronic components together form the central nervous system of the car, process the information sent by various sensors according to a preset program, generate decision results and then generate control instructions, and then send the control instructions to the actuator, which executes various Scheduled control functions.
- Fig. 1 is an ECU implemented based on FPGA in the prior art. Please refer to FIG. 1, the FPGA-based ECU is to set the FPGA outside the ECU. The FPGA is connected to the external ECU, and is suitable for ECUs that do not have a large number of computing needs.
- technologies such as assisted driving and automatic driving, more and more sensors such as on-board cameras and radars connected to the ECU are required, and the ECU is required to have strong computing power to process data from various sensors.
- the data collected by the car camera is the most huge and difficult to process, and the current mainstream processing method for the video data collected by the car camera is the neural network algorithm for deep learning.
- the neural network algorithm for deep learning needs to consume huge ECU CPU (central processing) unit resources, that is, the ECU needs to have strong computing power.
- the existing FPGA-based ECU uses FPGA to process the video data sent by the on-board camera and send the processed data to an external ECU, the external ECU further processes the received data, generates a control command according to the processing result, and then completes the automatic driving according to the control command.
- the data calculated by the FPGA needs to be further processed by an external ECU, resulting in an increased burden on the ECU.
- the embodiments of the present application provide an FPGA based on a field programmable gate array, and the sensor data is processed by the FPGA integrated with ARM and sent to the MCU of the ECU to reduce the burden of the MCU.
- the FPGA integrated with ARM is processed by the FPGA integrated with ARM and sent to the MCU of the ECU to reduce the burden of the MCU.
- FIG. 2 is a schematic structural diagram of a vehicle-mounted control unit provided by an embodiment of the present application.
- the vehicle-mounted control unit provided by the embodiment of the present application includes: a first system on chip (system on chip, SoC) and a micro control unit (micro controller unit (MCU), the first SoC uses a field programmable gate array
- SoC system on chip
- MCU micro controller unit
- the FPGA and the advanced reduced instruction set machine (advanced RISC Machine, ARM) are integrated, and the first SoC and the MCU are connected through an Ethernet switching chip.
- the first SoC integrates FPGA and ARM, and encapsulates the two into a single chip, which improves the integration and reduces the data connection between FPGA and ARM, so that the first SoC can do more operations. Improve the reliability while reducing the complexity of the ECU.
- the first SoC is connected to the MCU through an Ethernet switching chip, thereby implementing data exchange between the first SoC and the MCU.
- the Ethernet switching chip can also interact with external on-board Ethernet.
- the MCU is responsible for the functional safety of the printed circuit board (PCB) of the ECU and the internal and external interfaces.
- the MCU internally interacts with the first SoC through the Ethernet switch chip, and the global positioning system (global positioning system) system (GPS), inertial measurement unit (inertial measurement unit (IMU), temperature sensor, etc.) all interact with the MCU.
- GPS global positioning system
- IMU inertial measurement unit
- the MCU leads the debugging interface and the vehicle control CAN interface.
- the debugging interface is mainly responsible for the debugging of the board, and the CAN interface realizes the control function of the vehicle.
- the vehicle-mounted control unit provided in the embodiment of the present application includes an MCU and a first SoC implemented by integrating FPGA and ARM settings.
- the vehicle-mounted control unit is set on an autonomous vehicle.
- the FPGA of the first SoC receives the sensor sent by the vehicle-mounted sensor.
- For video data use the first neural network algorithm to visually perceive the video data to obtain first perception information, and send the first perception information to the ARM of the first SoC, and the ARM of the first SoC processes the first perception information to obtain the first
- the decision information is sent to the MCU, and finally the MCU generates a control instruction according to the first decision information and sends it to the corresponding execution mechanism, so that the control mechanism performs automatic driving according to the control execution.
- the first SoC is integrated with FPGA and ARM, and the sensor data is processed by FPAG and ARM and sent to the MCU to reduce the burden on the MCU.
- FIG. 3 is a flowchart of an automatic driving method based on a field programmable gate array FPGA provided by an embodiment of the present application. This embodiment includes:
- the FPGA of the first SoC receives video data sent by a vehicle sensor.
- the vehicle-mounted control unit is provided on the self-driving car, and the vehicle-mounted control unit includes a first SoC and an MCU, and the first SoC is integratedly set through FPGA and ARM.
- the on-board camera on the autonomous vehicle collects video data and sends it to the FPGA on the first SoC.
- the FPGA on the first SoC receives the video data.
- the FPGA of the first SoC performs visual perception on the video data using a first neural network algorithm to obtain first perception information.
- the FPGA on the first SoC uses the first neural network algorithm to perform visual perception on the received video data to obtain first perception information.
- the first neural network algorithm is a preset neural network algorithm based on deep learning, such as a convolutional neural network (convolutional neural networks, CNN) algorithm, etc.
- a convolutional neural network convolutional neural networks, CNN
- TDNN time delay network
- SIANN shift-invariant artificial neural network
- LeNet-5 neural network algorithm e.g., a neural network algorithm for example, a convolutional neural network (convolutional neural networks, SIANN), LeNet-5 neural network algorithm, VGGNet neural network algorithm, GoogLeNet neural network algorithm or ResNet Neural network algorithms, etc.
- the FPGA of the first SoC sends the first perception information to the ARM of the first SoC.
- the FPGA of the first SoC sends the first sensing information to the ARM of the first SoC through the FPGA internal bus.
- the ARM of the first SoC processes the first perception information, obtains first decision information, and sends it to the MCU.
- the ARM of the first SoC models the environment based on the first perception information, etc., and uses the decision algorithm to process the first perception information, and then sends the first decision information to the MCU.
- the MCU After receiving the first decision information, the MCU generates a control instruction according to the first decision information and sends it to the corresponding execution mechanism, so that the control mechanism performs automatic driving according to the control execution.
- the field-codeable gate array FPGA-based automatic driving method provided by an embodiment of the present application is applicable to an on-vehicle control unit.
- the on-vehicle control unit includes an MCU and a first SoC implemented by FPGA and ARM integrated settings.
- the on-vehicle control unit is set to automatic In the driving car, during the automatic driving process, the FPGA of the first SoC receives the video data sent by the vehicle sensor, uses the first neural network algorithm to visually perceive the video data, obtains the first perception information, and sends the first SoC's ARM The first perception information, the ARM of the first SoC processes the first perception information to obtain the first decision information and sends it to the MCU, and finally the MCU generates a control instruction according to the first decision information and sends it to the corresponding executing agency, so that the control agency This control is executed for automatic driving.
- the first SoC is integrated with FPGA and ARM, and the sensor data is processed by FPAG and ARM and sent to the MCU to reduce the burden on the MCU.
- the ARM of the first SoC when the ARM of the first SoC processes the first perception information to obtain the first decision information and sends it to the MCU, the ARM of the first SoC also receives radar data and fuse the first Perceive information and radar data, and then process the merged first perception information and radar data to obtain first decision information and send it to the MCU.
- the ARM of the first SoC after receiving the first perception information sent by the FPGA of the first SoC, the ARM of the first SoC aggregates the first perception information and the data sent by other sensors to obtain fusion data, and uses the fusion data to build an environment After the first decision information is obtained, it is sent to the MCU.
- the MCU After receiving the first decision information, the MCU generates a control instruction according to the first decision information and sends it to the corresponding execution mechanism, so that the control mechanism performs automatic driving according to the control execution.
- the tactile sensor includes ultrasonic radar, millimeter wave radar, lidar, etc.
- the data sent by the sensor includes at least one of ultrasonic radar data, millimeter wave radar data, and lidar data.
- the vehicle-mounted control unit provided by the embodiment of the present application on the basis of the above-mentioned FIG. 2, further includes: a second SoC, which is integratedly set through an FPGA and an advanced reduced instruction set machine ARM.
- the onboard camera in addition to sending the collected video data to the FPGA of the first SoC, the onboard camera also sends the video data to the FPGA of the second SoC.
- the FPGA of the second SoC uses the second neural network algorithm to visually sense the video data to obtain the second perception information and sends the second perception information to the ARM of the second SoC through the FPGA internal bus.
- the ARM of the second SoC After receiving the second perception information, the ARM of the second SoC processes the second perception information to obtain second decision information and sends it to the MCU. In this way, the MCU will receive the first decision information sent by the ARM of the first SoC and the second decision information sent by the second SoC. The MCU compares and analyzes the two decision-making information, selects appropriate strategy information from them, generates a control instruction according to the selected strategy information and sends it to the corresponding executing agency, so that the control agency performs automatic driving according to the control execution.
- the FPGA of the first SoC uses the first neural network algorithm to visually perceive the video data to obtain first perception information; the FPGA of the second SoC uses the second neural network algorithm to visually perceive the video data To get the second perception information.
- the first neural network algorithm and the second neural network algorithm are different neural network algorithms from each other.
- the first neural network algorithm is a convolutional neural network algorithm
- the second neural network algorithm is a Bp algorithm.
- the first neural network algorithm is a time delay network algorithm
- the second neural network algorithm is a ResNet neural network algorithm.
- the first SoC, the second SoC, and the MCU obtain a highly integrated vehicle control unit architecture with huge processing capabilities, and use the first SoC FPGA and the second SoC FPGA to perform neural network algorithm on the video data.
- Calculation, and the perception information obtained by calculating the video data by each FPGA is sent to the ARM corresponding to the SoC, and the ARM integrates the perception information and the radar data. Therefore, the vehicle-mounted control unit provided by the embodiment of the present application has stronger processing capabilities.
- the first SoC and the second SoC are used to implement a dual SoC, thereby achieving a heterogeneous redundant structure, that is, the FPGA of the first SoC and the FPGA of the second SoC use different
- the neural network algorithm processes the video data, which not only ensures the functional safety of the vehicle-mounted control unit but also can be used as the main and standby functions.
- the vehicle-mounted control unit provided by the embodiment of the present application may further include a first synchronous dynamic random access memory (SDRAM) and a first flash memory, the first SDRAM is connected to the first SoC, The first Flash is connected to the first SoC.
- the vehicle-mounted control unit provided in the embodiment of the present application may further include: a second synchronous dynamic random access memory (SDRAM) and a second flash memory, the second SDRAM is connected to the second SoC, and the second Flash Connect with the second SoC.
- SDRAM synchronous dynamic random access memory
- the second SoC loads the program from the second Flash. After the program is loaded, it waits for data input from the vehicle sensor and other sensors.
- the FPGA of the first SoC uses the first neural network algorithm to visually perceive the video data to obtain first perception information; the FPGA of the second SoC uses the second neural network algorithm to visually perceive the video data, Get the second perception information.
- the first perception information is sent by the FPGA of the first SoC to the ARM of the first SoC, and the ARM of the first SoC fuse the first perception information with the detection data of various radars to obtain the first fusion data, and use the first Fusion data is used to model the environment, and then the first decision information is generated according to the decision algorithm and sent to the MCU through the Ethernet switching chip;
- the second perception information is sent to the second SoC's ARM by the second SoC's FPGA, and the second SoC's ARM fuses the second perception information and the detection data of various radars to obtain the second fusion data, and uses the second fusion data to model the environment, and then generates the second decision information according to the decision algorithm and exchanges the chip through the Ethernet Send to MCU.
- the MCU After receiving the first decision information and the second decision information, the MCU analyzes and compares the two decision information, selects the appropriate decision information, generates a control command according to the selected decision information, and sends the control command to the CAN interface to The actuator, the actuator executes automatic driving according to the control instructions.
- FIG. 5 is a schematic structural diagram of an automatic driving device provided by an embodiment of the present application.
- the automatic driving device may be a vehicle-mounted control unit, or may be a chip applied inside the vehicle-mounted control unit.
- the automatic driving device can be used to perform the functions of the vehicle-mounted control unit in the above-described embodiment.
- the automatic driving device 100 may include: a first system-on-chip SoC module 11 and a micro-control unit MCU module 12, the first SoC module 11 is integrated with an advanced reduced instruction set machine ARM through FPGA,
- the vehicle-mounted control unit is provided on an autonomous driving vehicle, and the first SoC module 11 includes a first FPGA unit 111 and a first ARM unit 112, wherein,
- the first FPGA unit 111 is configured to receive video data sent by an on-board camera, perform visual perception on the video data using a first neural network algorithm, obtain first perception information, and send the first ARM unit 112 the First perception information;
- the first ARM unit 112 is configured to process the first perception information, obtain first decision information, and send it to the MCU module 12.
- the first ARM unit 112 is specifically used to receive radar data, fuse the first perception information and the radar data, and combine the first perception information and the radar data after fusion Perform processing to obtain the first decision information and send it to the MCU.
- the automatic driving device provided in the embodiment of the present application is suitable for an on-board control unit.
- the automatic driving device includes an MCU module and a first SoC module.
- the first SoC module includes a first FPGA unit and a first ARM unit.
- the first FPGA unit receives the video data sent by the vehicle sensor, uses the first neural network algorithm to visually perceive the video data to obtain the first perception information, and sends the first perception information to the first ARM unit, which is processed by the first ARM unit
- the first perception information obtains the first decision information and sends it to the MCU module.
- the MCU module generates a control instruction according to the first decision information and sends it to the corresponding execution mechanism, so that the control mechanism performs automatic driving according to the control execution.
- the first SoC is integrated with FPGA and ARM, and the sensor data is processed by FPAG and ARM and sent to the MCU to reduce the burden on the MCU.
- a second SoC module 13 the second SoC module 13 includes a second FPGA unit 131 and a second FARM unit 132, wherein,
- the second FPGA unit 131 is configured to receive the video data sent by the vehicle-mounted camera, and perform visual perception on the video data using a second neural network algorithm to obtain second perception information and send the second perception information to the second ARM unit 132 sending the second perception information;
- the second ARM unit 132 is configured to process the second perception information, obtain second decision information, and send it to the MCU;
- the MCU module 12 is configured to generate a control instruction according to the first decision information and the second decision information.
- An embodiment of the present application further provides a storage medium that stores computer-executed instructions, which are used by the processor to implement the above-mentioned field programmable gate array FPGA-based automatic driving method when executed by the processor .
- the described device and method may be implemented in other ways.
- the device embodiments described above are only schematic.
- the division of the modules is only a division of logical functions.
- there may be other divisions for example, multiple modules may be combined or integrated To another system, or some features can be ignored, or not implemented.
- the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or modules, and may be in electrical, mechanical or other forms.
- modules described as separate components may or may not be physically separated, and the components displayed as modules may or may not be physical units, that is, they may be located in one place or may be distributed on multiple network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of this embodiment.
- the functional modules in the embodiments of the present application may be integrated into one processing unit, or each module may exist alone physically, or two or more modules may be integrated into one unit.
- the unit formed by the above modules can be implemented in the form of hardware, or in the form of hardware plus software functional units.
- the above-mentioned integrated modules implemented in the form of software function modules may be stored in a computer-readable storage medium.
- the above software function modules are stored in a storage medium, and include several instructions to enable an electronic device (which may be a personal computer, server, or network device, etc.) or a processor (English: processor) to perform the embodiments described in this application Part of the method.
- processor may be a central processing unit (central processing unit, CPU), or other general-purpose processors, digital signal processors (DSPs), application specific integrated circuits (ASICs) Wait.
- the general-purpose processor may be a microprocessor or the processor may be any conventional processor or the like. The steps of the method disclosed in conjunction with the invention may be directly embodied and completed by a hardware processor, or may be performed and completed by a combination of hardware and software modules in the processor.
- the memory may include a high-speed RAM memory, and may also include a non-volatile storage NVM, such as at least one magnetic disk storage, and may also be a U disk, a mobile hard disk, a read-only memory, a magnetic disk, or an optical disk.
- NVM non-volatile storage
- the bus may be an industry standard architecture (ISA) bus, an external device interconnection (peripheral component, PCI) bus, or an extended industry standard architecture (extended Industry standard architecture, EISA) bus, etc.
- ISA industry standard architecture
- PCI peripheral component
- EISA extended Industry standard architecture
- the bus can be divided into address bus, data bus, control bus and so on.
- the bus in the drawings of this application does not limit to only one bus or one type of bus.
- the above storage medium may be implemented by any type of volatile or non-volatile storage device or a combination thereof, such as static random access memory (SRAM), electrically erasable programmable read only memory (EEPROM), erasable In addition to programmable read only memory (EPROM), programmable read only memory (PROM), read only memory (ROM), magnetic memory, flash memory, magnetic disk or optical disk.
- SRAM static random access memory
- EEPROM electrically erasable programmable read only memory
- EPROM erasable In addition to programmable read only memory
- PROM programmable read only memory
- ROM read only memory
- magnetic memory flash memory
- flash memory magnetic disk or optical disk.
- optical disk any available medium that can be accessed by a general-purpose or special-purpose computer.
- An exemplary storage medium is coupled to the processor so that the processor can read information from the storage medium and can write information to the storage medium.
- the storage medium may also be a component of the processor.
- the processor and the storage medium may be located in application-specific integrated circuits (application specific integrated circuits, ASIC).
- ASIC application specific integrated circuits
- the processor and the storage medium may also exist as discrete components in the terminal or server.
- the computer program product includes one or more computer instructions.
- the computer may be a general-purpose computer, a dedicated computer, a computer network, or other programmable devices.
- the computer instructions can be stored in a computer-readable storage medium or transmitted from one computer-readable storage medium to another computer-readable storage medium.
- the computer instructions can be transmitted from a website site, computer, server, or data center via wire (e.g.
- the computer-readable storage medium may be any available medium that can be accessed by a computer or a data storage device including one or more available media integrated servers, data centers, and the like.
- the available media may be magnetic media (eg, floppy disk, hard disk, magnetic tape), optical media (eg, DVD), or semiconductor media (eg, Solid State Disk, SSD), and so on.
- plural herein refers to two or more.
- the term “and/or” in this article is just an association relationship that describes an associated object, indicating that there can be three relationships, for example, A and/or B, which can mean: A exists alone, A and B exist at the same time, exist alone B these three cases.
- the character "/" in this article generally indicates that the related object is a "or” relationship; in the formula, the character "/" indicates that the related object is a "divide” relationship.
Abstract
Description
Claims (10)
- 一种基于现场可编程门阵列FPGA的自动驾驶方法,其特征在于,适用于车载控制单元,所述车载控制单元包括第一片上系统SoC和微控制单元MCU,所述第一SoC通过FPGA和进阶精简指令集机器ARM集成设置,所述车载控制单元设置在自动驾驶车辆上,所述方法包括:所述第一SoC的FPGA接收车载摄像头发送的视频数据;所述第一SoC的FPGA利用第一神经网络算法对所述视频数据进行视觉感知,得到第一感知信息;所述第一SoC的FPGA向所述第一SoC的ARM发送所述第一感知信息;所述第一SoC的ARM处理所述第一感知信息,得到第一决策信息并发送给所述MCU。
- 根据权利要求1所述的方法,其特征在于,所述第一SoC的ARM处理所述第一感知信息,得到第一决策信息并发送给所述MCU,包括:所述第一SoC的ARM接收雷达数据;所述第一SoC的ARM融合所述第一感知信息与所述雷达数据;所述第一SoC的ARM对融合后的所述第一感知信息与所述雷达数据进行处理,得到所述第一决策信息并发送给所述MCU。
- 根据权利要求2所述的方法,其特征在于,所述雷达数据包括超声波雷达数据、毫米波雷达数据、激光雷达数据中的至少一个。
- 根据权利要求1~3任一项所述的方法,其特征在于,所述车载控制单元还包括:第二片上系统SoC,所述第二SoC通过FPGA和进阶精简指令集机器ARM集成设置,所述方法还包括:所述第二SoC的FPGA接收所述车载摄像头发送的所述视频数据;所述第二SoC的FPGA利用第二神经网络算法对所述视频数据进行视觉感知,得到第二感知信息;所述第二SoC的FPGA向所述第二SoC的ARM发送所述第二感知信息;所述第二SoC的ARM处理所述第二感知信息,得到第二决策信息并发送给所述MCU;所述MCU根据所述第一决策信息和所述第二决策信息,生成控制指令。
- 一种车载控制单元,其特征在于,包括:第一片上系统SoC和微控制单元MCU,所述第一SoC通过现场可编程门阵列FPGA和进阶精简指令集机器ARM集成设置,所述第一SoC与所述MCU通过以太网交换芯片连接。
- 根据权利要求5所述的车载控制单元,其特征在于,还包括:第二片上系统SoC,所述第二SoC上设置FPGA和ARM,所述第二SoC上的FPGA与ARM通过总线连接,所述第二SoC与所述MCU通过所述以太网交换芯片连接。
- 根据权利要求5或6所述的车载控制单元,其特征在于,还包括:第一同步动态随机存储器SDRAM和第一闪存Flash,所述第一SDRAM与所述第一SoC连接,所述第一Flash与所述第一SoC连接。
- 一种自动驾驶装置,其特征在于,该装置适用于车载控制单元,所述自动驾驶装置包括:第一片上系统SoC模块和微控制单元MCU模块,所述第一SoC模块通过FPGA和进阶精简指令集机器ARM集成设置,所述车载控制单元设置在自动驾驶车辆上,所述第一SoC包括第一FPGA单元和第一ARM单元,其中,所述第一FPGA单元,用于接收车载摄像头发送的视频数据,利用第一神经网络算法对所述视频数据进行视觉感知,得到第一感知信息,向所述第一SoC的ARM发送所述第一感知信息;所述第一ARM单元,用于处理所述第一感知信息,得到第一决策信息并发送给所述MCU模块。
- 根据权利要求8所述的装置,其特征在于,所述第一ARM单元,具体用于接收雷达数据,融合所述第一感知信息与所述雷达数据,对融合后的所述第一感知信息与所述雷达数据进行处理,得到所述第一决策信息并发送给所述MCU。
- 根据权利要求8或9所述的装置,其特征在于,所述装置还包括:第二SoC模块,所述第二SoC模块包括第二FPGA单元和第二ARM单元,其中,所述第二FPGA单元,用于接收所述车载摄像头发送的所述视频数据,利用第二神经网络算法对所述视频数据进行视觉感知,得到第二感知信息,向所述第二SoC的ARM发送所述第二感知信息;所述第二ARM单元,用于处理所述第二感知信息,得到第二决策信息并发送给所述MCU;所述MCU模块,用于根据所述第一决策信息和所述第二决策信息,生成控制指令。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2020572668A JP7416731B2 (ja) | 2018-12-28 | 2019-12-27 | 車載制御ユニット、fpgaベースの車両自動運転方法及び装置 |
KR1020207024742A KR102471010B1 (ko) | 2018-12-28 | 2019-12-27 | 차량 탑재 제어 유닛, fpga 기반 차량 자동 주행 방법 및 장치 |
EP19903938.9A EP3839686B1 (en) | 2018-12-28 | 2019-12-27 | Vehicle-mounted control unit, and fpga-based vehicle automatic driving method and device |
US17/024,137 US20210001880A1 (en) | 2018-12-28 | 2020-09-17 | Vehicle-mounted control unit, and method and apparatus for fpga based automatic driving of vehicle |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811625816.8 | 2018-12-28 | ||
CN201811625816 | 2018-12-28 | ||
CN201910013122.7 | 2019-01-07 | ||
CN201910013122.7A CN109814552A (zh) | 2018-12-28 | 2019-01-07 | 车载控制单元、基于fpga的车辆自动驾驶方法及装置 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/024,137 Continuation US20210001880A1 (en) | 2018-12-28 | 2020-09-17 | Vehicle-mounted control unit, and method and apparatus for fpga based automatic driving of vehicle |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020135730A1 true WO2020135730A1 (zh) | 2020-07-02 |
Family
ID=66604077
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/129248 WO2020135730A1 (zh) | 2018-12-28 | 2019-12-27 | 车载控制单元、基于fpga的车辆自动驾驶方法及装置 |
Country Status (6)
Country | Link |
---|---|
US (1) | US20210001880A1 (zh) |
EP (1) | EP3839686B1 (zh) |
JP (1) | JP7416731B2 (zh) |
KR (1) | KR102471010B1 (zh) |
CN (1) | CN109814552A (zh) |
WO (1) | WO2020135730A1 (zh) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112793481A (zh) * | 2021-01-11 | 2021-05-14 | 宁波均联智行科技股份有限公司 | 一种座舱主动安全方法及系统 |
Families Citing this family (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109814552A (zh) * | 2018-12-28 | 2019-05-28 | 百度在线网络技术(北京)有限公司 | 车载控制单元、基于fpga的车辆自动驾驶方法及装置 |
CN111025959B (zh) * | 2019-11-20 | 2021-10-01 | 华为技术有限公司 | 一种数据管理的方法、装置、设备及智能汽车 |
CN110780608B (zh) * | 2019-11-26 | 2023-03-10 | 北京百度网讯科技有限公司 | 仿真测试方法及装置 |
CN110955232A (zh) * | 2019-12-13 | 2020-04-03 | 深圳市英博超算科技有限公司 | 一种自动驾驶系统架构 |
CN110928693B (zh) * | 2020-01-23 | 2021-01-15 | 飞诺门阵(北京)科技有限公司 | 一种计算设备及资源分配方法 |
US11706076B2 (en) | 2020-01-23 | 2023-07-18 | Novnet Computing System Tech Co., Ltd. | Computer system with computing devices, communication device, task processing device |
JP7425615B2 (ja) * | 2020-02-05 | 2024-01-31 | マツダ株式会社 | 車両用制御システム |
CN111290310B (zh) * | 2020-02-13 | 2021-03-16 | 天津鹍骐科技有限公司 | 一种车载计算系统 |
US20220048525A1 (en) * | 2020-08-14 | 2022-02-17 | Nvidia Corporation | Hardware fault detection for feedback control systems in autonomous machine applications |
CN112527003B (zh) * | 2021-02-18 | 2021-07-16 | 北京图森智途科技有限公司 | 数据传输装置和系统 |
CN113110407B (zh) * | 2021-06-16 | 2021-09-10 | 奥特酷智能科技(南京)有限公司 | 基于区块网关电控单元的汽车控制器 |
CN113200003A (zh) * | 2021-06-18 | 2021-08-03 | 北京理工华创电动车技术有限公司 | 一种商用车整车多域集成控制系统及方法 |
CN113721503A (zh) * | 2021-08-16 | 2021-11-30 | 北京超星未来科技有限公司 | 一种车载计算平台、无人驾驶系统及车辆 |
CN113442938A (zh) * | 2021-08-31 | 2021-09-28 | 国汽智控(北京)科技有限公司 | 车载计算系统、电子设备和车辆 |
CN114348018A (zh) * | 2021-12-17 | 2022-04-15 | 际络科技(上海)有限公司 | 商用车辆的自动驾驶系统及方法 |
CN114407872A (zh) * | 2022-02-17 | 2022-04-29 | 驭势科技(浙江)有限公司 | 自动驾驶系统及车辆 |
CN114625189A (zh) * | 2022-02-28 | 2022-06-14 | 北京梧桐车联科技有限责任公司 | 车辆管理方法、装置及设备 |
CN114604260A (zh) * | 2022-05-11 | 2022-06-10 | 青岛慧拓智能机器有限公司 | 用于无人车的域控制器及域控制装置 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180093538A1 (en) * | 2017-11-07 | 2018-04-05 | Michael Benn Rothschild | Digital hardware method, system and device for preventing the initial onset of trailer oscillations and control thereof |
CN108037756A (zh) * | 2017-11-29 | 2018-05-15 | 江苏若博机器人科技有限公司 | 一种多传感器融合中速无人车探测避障系统 |
CN108162977A (zh) * | 2017-11-25 | 2018-06-15 | 深圳市元征科技股份有限公司 | 驾驶辅助系统及控制方法 |
CN108205279A (zh) * | 2017-12-08 | 2018-06-26 | 联创汽车电子有限公司 | 异构多芯片智能驾驶控制器 |
CN108508882A (zh) * | 2017-02-28 | 2018-09-07 | 郑州宇通客车股份有限公司 | 一种车辆辅助驾驶控制平台 |
CN109814552A (zh) * | 2018-12-28 | 2019-05-28 | 百度在线网络技术(北京)有限公司 | 车载控制单元、基于fpga的车辆自动驾驶方法及装置 |
Family Cites Families (33)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2008021034A (ja) * | 2006-07-11 | 2008-01-31 | Fujitsu Ten Ltd | 画像認識装置、画像認識方法、歩行者認識装置および車両制御装置 |
US20100063652A1 (en) * | 2008-09-11 | 2010-03-11 | Noel Wayne Anderson | Garment for Use Near Autonomous Machines |
US8478493B2 (en) * | 2008-09-11 | 2013-07-02 | Deere & Company | High integrity perception program |
US9188980B2 (en) * | 2008-09-11 | 2015-11-17 | Deere & Company | Vehicle with high integrity perception system |
US20130151088A1 (en) * | 2011-11-16 | 2013-06-13 | Flextronics Ap, Llc | Method and system for vehicle data collection regarding traffic |
NL2007104C2 (nl) * | 2011-03-18 | 2012-09-19 | Innovius B V | Voertuigvolgsysteem voor toepassing in een voertuig. |
US10380022B2 (en) * | 2011-07-28 | 2019-08-13 | Netlist, Inc. | Hybrid memory module and system and method of operating the same |
EP2973487B1 (en) * | 2013-03-13 | 2019-07-31 | Bosch Automotive Service Solutions Inc. | Vehicle measurement apparatus having a system-on-a-chip device, a sensor and a wireless adapter |
US20140267727A1 (en) * | 2013-03-14 | 2014-09-18 | Honda Motor Co., Ltd. | Systems and methods for determining the field of view of a processed image based on vehicle information |
US9132746B2 (en) * | 2013-08-15 | 2015-09-15 | Honda Motor Co., Ltd. | Method and system for reducing range anxiety |
US9720797B2 (en) * | 2015-06-30 | 2017-08-01 | Nxp Usa, Inc. | Flash memory controller, data processing system with flash memory controller and method of operating a flash memory controller |
EP3358369A4 (en) * | 2015-09-30 | 2019-05-08 | Sony Corporation | INFORMATION PROCESSING DEVICE, INFORMATION PROCESSING METHOD, AND PROGRAM |
CN105335327B (zh) * | 2015-10-13 | 2017-11-21 | 电子科技大学 | 基于Soc的可重构/双冗余VPX3U信号处理载板 |
US10003554B1 (en) * | 2015-12-22 | 2018-06-19 | Amazon Technologies, Inc. | Assisted sideband traffic management |
JP6858002B2 (ja) * | 2016-03-24 | 2021-04-14 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 物体検出装置、物体検出方法及び物体検出プログラム |
US11378954B2 (en) * | 2016-06-28 | 2022-07-05 | Faraday&Future Inc. | Multi-processor SoC system |
US10509407B2 (en) * | 2016-07-01 | 2019-12-17 | Samsung Electronics Co., Ltd. | Apparatus and method for a vehicle platform |
CN111897751A (zh) * | 2017-01-26 | 2020-11-06 | 华为技术有限公司 | 一种数据传输的方法、装置、设备和系统 |
US9919702B1 (en) * | 2017-01-27 | 2018-03-20 | GM Global Technology Operations LLC | Systems and methods for managing battery state of charge |
CN108462726A (zh) * | 2017-02-14 | 2018-08-28 | 广州市联奥信息科技有限公司 | 面向未知状况的车辆辅助驾驶决策系统和装置 |
KR20180097247A (ko) * | 2017-02-23 | 2018-08-31 | 주식회사 만도 | 사용자 친화적 차량 제어 시스템 및 방법 |
JP6799590B2 (ja) * | 2017-03-10 | 2020-12-16 | バイドゥ ドットコム タイムス テクノロジー (ベイジン) カンパニー リミテッド | 自律走行車のアンダーステアリングを解決するための自動的ステアリング制御の基準適用 |
CN106951627A (zh) * | 2017-03-15 | 2017-07-14 | 北京百度网讯科技有限公司 | 车辆自动驾驶的仿真测试方法、装置、设备及计算机可读存储介质 |
US10816973B2 (en) * | 2017-06-02 | 2020-10-27 | Baidu Usa Llc | Utilizing rule-based and model-based decision systems for autonomous driving control |
WO2018230492A1 (ja) * | 2017-06-16 | 2018-12-20 | 本田技研工業株式会社 | 情報処理装置、情報処理方法、及びプログラム |
CN107404529A (zh) * | 2017-07-26 | 2017-11-28 | 成都网烁信息科技有限公司 | 一种基于网络云的大数据分析系统及方法 |
CN107885214A (zh) * | 2017-11-22 | 2018-04-06 | 济南浪潮高新科技投资发展有限公司 | 一种基于fpga的加速自动驾驶视觉感知的方法及装置 |
US10620266B2 (en) * | 2017-11-29 | 2020-04-14 | Intel Corporation | System, apparatus and method for in-field self testing in a diagnostic sleep state |
US11029744B2 (en) * | 2017-12-29 | 2021-06-08 | Intel Corporation | System, apparatus and method for controlling a processor based on effective stress information |
CN108196547B (zh) * | 2018-01-08 | 2020-06-23 | 北京图森未来科技有限公司 | 一种自动驾驶系统 |
CN108445885A (zh) * | 2018-04-20 | 2018-08-24 | 鹤山东风新能源科技有限公司 | 一种基于纯电动物流车的自动驾驶系统及其控制方法 |
CN108614790B (zh) * | 2018-05-03 | 2021-02-05 | 浪潮集团有限公司 | 一种车载计算单元、车辆及车辆自动驾驶方法 |
CN108776472A (zh) * | 2018-05-17 | 2018-11-09 | 驭势(上海)汽车科技有限公司 | 智能驾驶控制方法及系统、车载控制设备和智能驾驶车辆 |
-
2019
- 2019-01-07 CN CN201910013122.7A patent/CN109814552A/zh active Pending
- 2019-12-27 EP EP19903938.9A patent/EP3839686B1/en active Active
- 2019-12-27 KR KR1020207024742A patent/KR102471010B1/ko active IP Right Grant
- 2019-12-27 WO PCT/CN2019/129248 patent/WO2020135730A1/zh unknown
- 2019-12-27 JP JP2020572668A patent/JP7416731B2/ja active Active
-
2020
- 2020-09-17 US US17/024,137 patent/US20210001880A1/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108508882A (zh) * | 2017-02-28 | 2018-09-07 | 郑州宇通客车股份有限公司 | 一种车辆辅助驾驶控制平台 |
US20180093538A1 (en) * | 2017-11-07 | 2018-04-05 | Michael Benn Rothschild | Digital hardware method, system and device for preventing the initial onset of trailer oscillations and control thereof |
CN108162977A (zh) * | 2017-11-25 | 2018-06-15 | 深圳市元征科技股份有限公司 | 驾驶辅助系统及控制方法 |
CN108037756A (zh) * | 2017-11-29 | 2018-05-15 | 江苏若博机器人科技有限公司 | 一种多传感器融合中速无人车探测避障系统 |
CN108205279A (zh) * | 2017-12-08 | 2018-06-26 | 联创汽车电子有限公司 | 异构多芯片智能驾驶控制器 |
CN109814552A (zh) * | 2018-12-28 | 2019-05-28 | 百度在线网络技术(北京)有限公司 | 车载控制单元、基于fpga的车辆自动驾驶方法及装置 |
Non-Patent Citations (1)
Title |
---|
See also references of EP3839686A4 |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112793481A (zh) * | 2021-01-11 | 2021-05-14 | 宁波均联智行科技股份有限公司 | 一种座舱主动安全方法及系统 |
CN112793481B (zh) * | 2021-01-11 | 2021-10-15 | 宁波均联智行科技股份有限公司 | 一种座舱主动安全方法及系统 |
Also Published As
Publication number | Publication date |
---|---|
US20210001880A1 (en) | 2021-01-07 |
KR20200115594A (ko) | 2020-10-07 |
EP3839686A1 (en) | 2021-06-23 |
CN109814552A (zh) | 2019-05-28 |
EP3839686A4 (en) | 2022-05-04 |
EP3839686B1 (en) | 2024-05-01 |
KR102471010B1 (ko) | 2022-11-25 |
JP2021529388A (ja) | 2021-10-28 |
JP7416731B2 (ja) | 2024-01-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020135730A1 (zh) | 车载控制单元、基于fpga的车辆自动驾驶方法及装置 | |
CN109917765B (zh) | 一种基于自动驾驶系统的网络架构的集散式域控制器系统 | |
US11308739B2 (en) | Automatic driving system, vehicle control method and device | |
US11378954B2 (en) | Multi-processor SoC system | |
CN209842367U (zh) | 一种基于自动驾驶系统的网络架构的集散式域控制器系统 | |
CN110979314A (zh) | 一种自主代客泊车方法、车载设备及存储介质 | |
US20190132555A1 (en) | Methods and systems to broadcast sensor outputs in an automotive environment | |
CN113655788B (zh) | 车辆遥控泊车方法、系统、终端设备及可读存储介质 | |
JP2023534048A (ja) | データ送信方法および装置 | |
KR101703500B1 (ko) | 차량 유닛 | |
US20220212692A1 (en) | Automated driving control system and automated driving control apparatus | |
US11390243B2 (en) | Method and apparatus for processing vehicle collision, vehicle, device and storage medium | |
WO2023124354A1 (zh) | 一种车辆控制器、车辆和车辆控制方法 | |
US20200114905A1 (en) | Electronic Control Apparatus, Electronic Control System, and Electronic Control Method | |
EP4266184A1 (en) | Method and device for processing sensor data | |
US20230085381A1 (en) | Vehicle sensor data processing method and system | |
JP2019188958A (ja) | 駐車支援装置 | |
CN108243244B (zh) | 车载通信系统以及其控制方法 | |
CN111766866A (zh) | 信息处理装置和包括信息处理装置的自动行驶控制系统 | |
CN116533757A (zh) | 电压故障判断方法、装置、存储介质、控制器及车辆 | |
CN115503626A (zh) | 车辆的控制系统、方法、电子设备及存储介质 | |
CN113968227B (zh) | 用于非自动驾驶车的控制方法和装置 | |
EP3896915A1 (en) | Data processing method and apparatus | |
JP6979630B2 (ja) | 監視装置、監視方法及びプログラム | |
EP4144593A1 (en) | Method and system for processing data of sensor of vehicle |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19903938 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 20207024742 Country of ref document: KR Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2020572668 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2019903938 Country of ref document: EP Effective date: 20210317 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |