WO2019066477A1 - AUTONOMOUS VEHICLE AND ITS CONTROL METHOD - Google Patents

AUTONOMOUS VEHICLE AND ITS CONTROL METHOD Download PDF

Info

Publication number
WO2019066477A1
WO2019066477A1 PCT/KR2018/011403 KR2018011403W WO2019066477A1 WO 2019066477 A1 WO2019066477 A1 WO 2019066477A1 KR 2018011403 W KR2018011403 W KR 2018011403W WO 2019066477 A1 WO2019066477 A1 WO 2019066477A1
Authority
WO
WIPO (PCT)
Prior art keywords
vehicle
route
location
learned
parking
Prior art date
Application number
PCT/KR2018/011403
Other languages
English (en)
French (fr)
Inventor
Ho Yun
Junghee Park
Boyoung JEONG
Soonhong Jung
Original Assignee
Lg Electronics Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Lg Electronics Inc. filed Critical Lg Electronics Inc.
Priority to CN201880063165.XA priority Critical patent/CN111148674A/zh
Publication of WO2019066477A1 publication Critical patent/WO2019066477A1/en

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B62LAND VEHICLES FOR TRAVELLING OTHERWISE THAN ON RAILS
    • B62DMOTOR VEHICLES; TRAILERS
    • B62D15/00Steering not otherwise provided for
    • B62D15/02Steering position indicators ; Steering position determination; Steering aids
    • B62D15/027Parking aids, e.g. instruction means
    • B62D15/0285Parking performed automatically
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W30/00Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units, or advanced driver assistance systems for ensuring comfort, stability and safety or drive control systems for propelling or retarding the vehicle
    • B60W30/14Adaptive cruise control
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W30/00Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units, or advanced driver assistance systems for ensuring comfort, stability and safety or drive control systems for propelling or retarding the vehicle
    • B60W30/06Automatic manoeuvring for parking
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W30/00Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units, or advanced driver assistance systems for ensuring comfort, stability and safety or drive control systems for propelling or retarding the vehicle
    • B60W30/18Propelling the vehicle
    • B60W30/18009Propelling the vehicle related to particular drive situations
    • B60W30/18163Lane change; Overtaking manoeuvres
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/34Route searching; Route guidance
    • G01C21/3453Special cost functions, i.e. other than distance or default speed limit of road segments
    • G01C21/3484Personalized, e.g. from learned user behaviour or user-defined profiles
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/34Route searching; Route guidance
    • G01C21/36Input/output arrangements for on-board computers
    • G01C21/3679Retrieval, searching and output of POI information, e.g. hotels, restaurants, shops, filling stations, parking facilities
    • G01C21/3685Retrieval, searching and output of POI information, e.g. hotels, restaurants, shops, filling stations, parking facilities the POI's being parking facilities
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/38Electronic maps specially adapted for navigation; Updating thereof
    • G01C21/3804Creation or updating of map data
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/01Detecting movement of traffic to be counted or controlled
    • G08G1/0104Measuring and analyzing of parameters relative to traffic conditions
    • G08G1/0108Measuring and analyzing of parameters relative to traffic conditions based on the source of data
    • G08G1/0112Measuring and analyzing of parameters relative to traffic conditions based on the source of data from the vehicle, e.g. floating car data [FCD]
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/01Detecting movement of traffic to be counted or controlled
    • G08G1/0104Measuring and analyzing of parameters relative to traffic conditions
    • G08G1/0125Traffic data processing
    • G08G1/0133Traffic data processing for classifying traffic situation
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/09Arrangements for giving variable traffic instructions
    • G08G1/0962Arrangements for giving variable traffic instructions having an indicator mounted inside the vehicle, e.g. giving voice messages
    • G08G1/09623Systems involving the acquisition of information from passive traffic signs by means mounted on the vehicle
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/14Traffic control systems for road vehicles indicating individual free spaces in parking areas
    • G08G1/145Traffic control systems for road vehicles indicating individual free spaces in parking areas where the indication depends on the parking areas
    • G08G1/146Traffic control systems for road vehicles indicating individual free spaces in parking areas where the indication depends on the parking areas where the parking area is a limited parking space, e.g. parking garage, restricted space
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/14Traffic control systems for road vehicles indicating individual free spaces in parking areas
    • G08G1/145Traffic control systems for road vehicles indicating individual free spaces in parking areas where the indication depends on the parking areas
    • G08G1/147Traffic control systems for road vehicles indicating individual free spaces in parking areas where the indication depends on the parking areas where the parking area is within an open public zone, e.g. city centre
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/16Anti-collision systems
    • G08G1/166Anti-collision systems for active traffic, e.g. moving vehicles, pedestrians, bikes
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/16Anti-collision systems
    • G08G1/168Driving aids for parking, e.g. acoustic or visual feedback on parking space
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2554/00Input parameters relating to objects
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60YINDEXING SCHEME RELATING TO ASPECTS CROSS-CUTTING VEHICLE TECHNOLOGY
    • B60Y2300/00Purposes or special features of road vehicle drive control systems
    • B60Y2300/14Cruise control
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/09Arrangements for giving variable traffic instructions
    • G08G1/0962Arrangements for giving variable traffic instructions having an indicator mounted inside the vehicle, e.g. giving voice messages
    • G08G1/0967Systems involving transmission of highway information, e.g. weather, speed limits

Definitions

  • the present disclosure relates to an autonomous vehicle and a method of controlling the same.
  • a vehicle is an apparatus that can transport a user in a desired direction.
  • the vehicle may include an automobile.
  • Vehicles may be equipped with sensors and electronic devices to provide user convenience. For example, research has been actively conducted on advanced driver assistance systems (ADAS) to provide user convenience for various operations in driving. Further, autonomous vehicles are under active development.
  • ADAS advanced driver assistance systems
  • the autonomous vehicle may be unable to perform autonomous parking. In some cases, the autonomous vehicle may be unable to actively learn a plurality of parking slots.
  • One aspect of the present disclosure is to provide an autonomous vehicle configured to determine and store a property of an object detected in a driving route during learning the driving route from a starting location to an ending location.
  • Another aspect of the present disclosure is to provide an autonomous vehicle for starting to drive in a learned route, detecting an empty parking slot in the route, and parking in the detected parking slot.
  • a method of controlling a vehicle that is configured to be autonomously driven includes determining a learned route based on a driving route that the vehicle has driven in a manual mode from a starting location to an ending location, driving the vehicle along the learned route in an autonomous mode, detecting a parking space based on driving the vehicle along the learned route in the autonomous mode, and based on a detection of the parking space in the learned route, parking the vehicle in the detected parking space.
  • determining the learned route based on the driving route may include detecting an object located in the driving route from the starting location to the ending location, determining a property of the object detected in the driving route, and storing information about the property of the object.
  • storing the information about the property of the object comprises storing first information related to an object that is located at a static location in the driving route or second information related to an object that is movable in the driving route.
  • Determining the learned route may further include, based on the property of the object, determining a second route between a first location and a second location in the driving route, the second route being different from the driving route in which the vehicle has driven from the first location to the second location, and storing the second route as the learned route.
  • determining the learned route may further include determining a partial route including at least one of the starting location or the ending location, and based on the partial route, generating the learned route from the starting location to the ending location. In some implementations, determining the learned route may further include detecting at least one branch point in the driving route in which the vehicle has driven, and storing a route that extends from the at least one branch point as the learned route.
  • Driving the vehicle along the learned route in the autonomous mode may include determining, by an object detection device, that the vehicle has arrived at a location within a distance from the starting location, receiving, through a user interface device, a user input for starting to drive along the learned route, and based on the user input, driving the vehicle along the learned route from the location within the distance from the starting location.
  • the method may further include receiving information about a parking facility, where driving the vehicle along the learned route in the autonomous mode may include based on the information about the parking facility, modifying the learned route, and driving the vehicle along the modified learned route in the autonomous mode.
  • driving the vehicle along the learned route in the autonomous mode may include receiving a user input for returning the vehicle to one of the starting location or the ending location, and in response to the user input, controlling the vehicle to return to the one of the starting location or the ending location along the learned route.
  • controlling the vehicle to the one of the starting location or the ending location along the learned route may further include determining a current location of the vehicle in the learned route, and based on the determined current location of the vehicle, determining whether to maintain a driving direction of the vehicle.
  • detecting the parking space may include receiving, from an object detection unit, a detection result including detection of the object, and based on the received detection result and the stored information about the property of the object, detecting the parking space.
  • Parking the vehicle in the detected parking space may include, based on a detection of a plurality of parking spaces, selecting a parking space according to a predetermined prioritization.
  • the parking space may include a reserved parking space that is located at the ending location.
  • parking the vehicle in the detected parking space may include detecting an object in the reserved parking space, the object not being included in the learned route, based on an area occupied by the object in the reserved parking space, determining an available area for parking of the vehicle, and parking the vehicle in the available area.
  • the method may further include, based on parking the vehicle in the detected parking space, storing information about a location of the detected parking space in a memory, in response to a user input, obtaining the information about the location of the parking space and information about the learned route from the memory, and controlling the vehicle to move out from the parking space to the ending location along at least a part of the learned route.
  • the method may further include detecting, by an object detection device, an object located outside of the vehicle, and generating a connection route to the learned route based on the detected object.
  • controlling the vehicle to move out from the parking space may include determine a current location of the vehicle, determining whether the current location of the vehicle has changed since the vehicle parked, based on a determination that the current location of the vehicle has changed since the vehicle parked, determine whether the current location of the vehicle is in the learned route, based on a determination that the current location of the vehicle is in the learned route, controlling the vehicle to move out from the parking space along the learned route, and based on a determination that the current location of the vehicle is outside of the learned route, generating a connection route to the learned route and controlling the vehicle to move out from the parking space along the connection route to the learned route.
  • the method may further include identifying a user of the vehicle, determining a location of the user, based on the location of the user, determining whether the user is located in the learned route, and based on a determination that the user is located in the learned route, controlling the vehicle to stop at the location of the identified user.
  • the method may further include based on a determination that the identified user is located outside of the learned route, generating a connection route to the location of the user, and driving the vehicle to the location of the user along the connection route.
  • a vehicle includes a wheel, a power source configured to drive the wheel, and at least one processor.
  • the at least one processor is configured to determine a learned route based on a driving route that the vehicle has driven in a manual mode from a starting location to an ending location, drive the vehicle along the learned route in an autonomous mode, detect a parking space based on the vehicle being driven along the learned route in the autonomous mode, and based on a detection of the parking space in the learned route, park the vehicle in the detected parking space.
  • the implementation of the present disclosure achieves the technical effect that an optimum driving route may be provided for the vehicle in the autonomous mode by modifying a route learned according to the type of an obstacle as well as an actually driven route.
  • the implementation of the present disclosure achieves the technical effect that instead of unconditionally driving in a learned route from a starting location to an ending location, the vehicle generates an optimum driving route based on information received from an external device.
  • the implementation of the present disclosure achieves the technical effect that driving and parking of other vehicles are not obstructed by perceiving the ambient space of a vehicle, in obstacle-avoiding parking.
  • the implementation of the present disclosure achieves the technical effect that a pull-out route changed adaptively according to an ambient environment condition of a vehicle is generated during pull-out.
  • the implementation of the present disclosure achieves the technical effect that if a more favorable pull-out route than a learned pull-out route may be generated, an adaptively changed pull-out route is generated.
  • FIG. 1 is a view illustrating an example external appearance of an example vehicle.
  • FIG. 2 illustrates different angled views of example external appearances of an example vehicle.
  • FIGS. 3 and 4 are views illustrating example interior configurations of an example vehicle.
  • FIGS. 5 and 6 are views illustrating example objects around an example vehicle.
  • FIG. 7 is a block diagram illustrating example components of an example vehicle.
  • FIG. 8 is a flowchart illustrating an example learning-based vehicle parking method.
  • FIG. 9 is a view illustrating an example of route learning in an example manual mode or an example learning mode.
  • FIG. 10 is a view illustrating examples of route learning in the manual mode or the learning mode.
  • FIG. 11 is a view illustrating an example of route learning an example vehicle in the manual mode or the learning mode.
  • FIG. 12 is a view illustrating examples of route learning of an example vehicle in the manual mode or the learning mode.
  • FIG. 13 is a view illustrating an example of route learning of an example vehicle in the manual mode or the learning mode.
  • FIG. 14 is a view illustrating an example of route learning of an example vehicle in the manual mode or the learning mode.
  • FIG. 15 is a view illustrating examples of route learning of an example vehicle in the manual mode or the learning mode.
  • FIG. 16 is a view illustrating an example vehicle driving in an example learned route.
  • FIG. 17 is a view illustrating an example vehicle driving in an example learned route.
  • FIG. 18 is a view illustrating example vehicles driving in example learned routes.
  • FIG. 19 is a view illustrating example vehicles driving in example learned routes.
  • FIG. 20 is a view illustrating an example change of a driving route based on input by a user.
  • FIG. 21 is a view illustrating example vehicles driving in example learned routes.
  • FIG. 22 is a view illustrating an example vehicle parking in an empty parking slot in an example learned route.
  • FIG. 23 is a view illustrating an example vehicle parking in an empty parking slot in an example learned route.
  • FIG. 24 is a view illustrating an example vehicle parking in an empty parking slot in an example learned route.
  • FIG. 25 is a view illustrating another example of a learning-based vehicle pulling-out method.
  • FIG. 26 is a view illustrating another example of a learning-based vehicle pulling-out method.
  • FIG. 27 is a view illustrating another example of a learning-based vehicle pulling-out method.
  • FIG. 28 is a view illustrating examples of a learning-based vehicle pulling-out method.
  • FIG. 29 is a view illustrating examples of a learning-based vehicle pulling-out method.
  • FIG. 30 is a view illustrating another example of a learning-based vehicle pulling-out method.
  • FIG. 31 is a view illustrating another example of a learning-based vehicle pulling-out method.
  • a vehicle described in this specification may include, but is not limited to, an automobile and a motorcycle.
  • a description will be given based on an automobile.
  • a vehicle described in this specification may include, but is not limited to, various types of internal combustion engine vehicles including an engine as a power source, a hybrid vehicle including both an engine and an electric motor as a power source, and an electric vehicle including an electric motor as a power source.
  • the left side of the vehicle refers to the left side in the forward driving direction of the vehicle
  • the right side of the vehicle refers to the right side in the forward driving direction of the vehicle
  • FIG. 1 is a view of the external appearance of a vehicle according to an implementation of the present disclosure
  • FIG. 2 is different angled views of a vehicle according to an implementation of the present disclosure
  • FIGS. 3 and 4 are views of the internal configuration of a vehicle according to an implementation of the present disclosure
  • FIGS. 5 and 6 are views for explanation of objects according to an implementation of the present disclosure
  • FIG. 7 is a block diagram illustrating a vehicle according to an implementation of the present disclosure.
  • a vehicle 100 may include a plurality of wheels, which are rotated by a power source, and a steering input device 510 for controlling a driving direction of the vehicle 100.
  • the vehicle 100 may be an autonomous vehicle.
  • the vehicle 100 may be switched to an autonomous mode or a manual mode in response to a user input.
  • the vehicle 100 may be switched from a manual mode to an autonomous mode, or vice versa.
  • the vehicle 100 may be switched to the autonomous mode or to the manual mode based on driving environment information.
  • the driving environment information may include at least one of the following: information on an object outside a vehicle, navigation information, and vehicle state information.
  • the vehicle 100 may be switched from the manual mode to the autonomous mode, or vice versa, based on driving environment information generated by the object detection device 300.
  • the vehicle 100 may be switched from the manual mode to the autonomous mode, or vice versa, based on driving environment information received through a communication device 400.
  • the vehicle 100 may be switched from the manual mode to the autonomous mode, or vice versa, based on information, data, and a signal provided from an external device.
  • the autonomous vehicle 100 may operate based on an operation system 700.
  • the autonomous vehicle 100 may operate based on information, data, or signals generated by a driving system 710, a vehicle pulling-out system 740, and a vehicle parking system 750.
  • the autonomous vehicle 100 may receive a user input for driving of the vehicle 100 through a maneuvering device 500. In response to the user input received through the maneuvering device 500, the vehicle 100 may operate.
  • overall length is the length from the front end to the rear end of the vehicle 100
  • the term “overall width” is the width of the vehicle 100
  • the term “overall height” is the height from the bottom of the wheel to the roof.
  • the term “overall length direction L” may mean the reference direction for the measurement of the overall length of the vehicle 100
  • the term “overall width direction W” may mean the reference direction for the measurement of the overall width of the vehicle 100
  • the term “overall height direction H” may mean the reference direction for the measurement of the overall height of the vehicle 100.
  • the vehicle 100 may include the user interface device 200, the object detection device 300, the communication device 400, the maneuvering device 500, a vehicle drive device 600, the operation system 700, a navigation system 770, a sensing unit 120, an interface 130, a memory 140, a controller 170, and a power supply unit 190.
  • the vehicle 100 may further include other components in addition to the aforementioned components, or may not include some of the aforementioned components.
  • the sensing unit 120 may sense the state of the vehicle.
  • the sensing unit 120 may include an attitude sensor (for example, a yaw sensor, a roll sensor, or a pitch sensor), a collision sensor, a wheel sensor, a speed sensor, a gradient sensor, a weight sensor, a heading sensor, a gyro sensor, a position module, a vehicle forward/reverse movement sensor, a battery sensor, a fuel sensor, a tire sensor, a steering sensor based on the rotation of the steering wheel, an in-vehicle temperature sensor, an in-vehicle humidity sensor, an ultrasonic sensor, an illumination sensor, an accelerator pedal position sensor, and a brake pedal position sensor.
  • an attitude sensor for example, a yaw sensor, a roll sensor, or a pitch sensor
  • a collision sensor for example, a yaw sensor, a roll sensor, or a pitch sensor
  • a wheel sensor for example, a speed sensor, a gradient sensor, a weight sensor, a
  • the sensing unit 120 may acquire sensing signals with regard to, for example, vehicle attitude information, vehicle collision information, vehicle driving direction information, vehicle location information (GPS information), vehicle angle information, vehicle speed information, vehicle acceleration information, vehicle tilt information, vehicle forward/reverse movement information, battery information, fuel information, tire information, vehicle lamp information, in-vehicle temperature information, in-vehicle humidity information, steering-wheel rotation angle information, outside illumination information, information about the pressure applied to an accelerator pedal, and information about the pressure applied to a brake pedal.
  • GPS information vehicle location information
  • vehicle angle information vehicle speed information
  • vehicle acceleration information vehicle acceleration information
  • vehicle tilt information vehicle forward/reverse movement information
  • battery information fuel information
  • tire information tire information
  • vehicle lamp information in-vehicle temperature information
  • in-vehicle humidity information in-vehicle humidity information
  • steering-wheel rotation angle information outside illumination information
  • the sensing unit 120 may further include, for example, an accelerator pedal sensor, a pressure sensor, an engine speed sensor, an Air Flow-rate Sensor (AFS), an Air Temperature Sensor (ATS), a Water Temperature Sensor (WTS), a Throttle Position Sensor (TPS), a Top Dead Center (TDC) sensor, and a Crank Angle Sensor (CAS).
  • AFS Air Flow-rate Sensor
  • ATS Air Temperature Sensor
  • WTS Water Temperature Sensor
  • TPS Throttle Position Sensor
  • TDC Top Dead Center
  • CAS Crank Angle Sensor
  • the sensing unit 120 may generate vehicle state information based on sensing data.
  • the vehicle condition information may be information that is generated based on data sensed by a variety of sensors inside a vehicle.
  • the vehicle state information may include vehicle position information, vehicle speed information, vehicle tilt information, vehicle weight information, vehicle direction information, vehicle battery information, vehicle fuel information, vehicle tire pressure information, vehicle steering information, in-vehicle temperature information, in-vehicle humidity information, pedal position information, vehicle engine temperature information, etc.
  • the interface 130 may serve as a passage for various kinds of external devices that are connected to the vehicle 100.
  • the interface 130 may have a port that is connectable to a mobile terminal and may be connected to the mobile terminal via the port. In this case, the interface 130 may exchange data with the mobile terminal.
  • the interface 130 may serve as a passage for the supply of electrical energy to a mobile terminal connected thereto.
  • the interface 130 may provide electrical energy, supplied from the power supply unit 190, to the mobile terminal under control of the controller 170.
  • the memory 140 is electrically connected to the controller 170.
  • the memory 140 may store basic data for each unit, control data for the operational control of each unit, and input/output data.
  • the memory 140 may be any of various hardware storage devices, such as a ROM, a RAM, an EPROM, a flash drive, and a hard drive.
  • the memory 140 may store various data for the overall operation of the vehicle 100, such as programs for the processing or control of the controller 170.
  • the memory 140 may be integrally formed with the controller 170, or may be provided as an element of the controller 170.
  • the controller 170 may control the overall operation of each unit inside the vehicle 100.
  • the controller 170 may be referred to as an Electronic Controller (ECU).
  • ECU Electronic Controller
  • the power supply unit 190 may supply power required to operate each component under control of the controller 170.
  • the power supply unit 190 may receive power from, for example, a battery inside the vehicle 100.
  • At least one processor and the controller 170 included in the vehicle 100 may be implemented using at least one selected from among Application Specific Integrated Circuits (ASICs), Digital Signal Processors (DSPs), Digital Signal Processing Devices (DSPDs), Programmable Logic Devices (PLDs), Field Programmable Gate Arrays (FPGAs), processors, controllers, micro-controllers, microprocessors, and electric units for the implementation of other functions.
  • ASICs Application Specific Integrated Circuits
  • DSPs Digital Signal Processors
  • DSPDs Digital Signal Processing Devices
  • PLDs Programmable Logic Devices
  • FPGAs Field Programmable Gate Arrays
  • processors controllers, micro-controllers, microprocessors, and electric units for the implementation of other functions.
  • each of the sensing unit 120, the interface unit 130, the memory 140, the power supply unit 190, the user interface device 200, the object detection device 300, the communication device 400, the maneuvering device 500, the vehicle drive device 600, the operation system 700, and the navigation system 770 may have an individual processor or may be incorporated in the controller 170.
  • the user interface device 200 is provided to support communication between the vehicle 100 and a user.
  • the user interface device 200 may receive a user input, and provide information generated in the vehicle 100 to the user.
  • the vehicle 100 may enable User Interfaces (UI) or User Experience (UX) through the user interface device 200.
  • UI User Interfaces
  • UX User Experience
  • the user interface device 200 may include an input unit 210, an internal camera 220, a biometric sensing unit 230, an output unit 250, and a processor 270. Each component of the user interface device 200 may be separated from or integrated with the afore-described interface 130, structurally or operatively.
  • the user interface device 200 may further include other components in addition to the aforementioned components, or may not include some of the aforementioned components.
  • the input unit 210 is configured to receive information from a user, and data collected in the input unit 210 may be analyzed by the processor 270 and then processed into a control command of the user.
  • the input unit 210 may be disposed inside the vehicle 100.
  • the input unit 210 may be disposed in a region of a steering wheel, a region of an instrument panel, a region of a seat, a region of each pillar, a region of a door, a region of a center console, a region of a head lining, a region of a sun visor, a region of a windshield, or a region of a window.
  • the input unit 210 may include a voice input unit 211, a gesture input unit 212, a touch input unit 213, and a mechanical input unit 214.
  • the voice input unit 211 may convert a voice input of a user into an electrical signal.
  • the converted electrical signal may be provided to the processor 270 or the controller 170.
  • the voice input unit 211 may include one or more microphones.
  • the gesture input unit 212 may convert a gesture input of a user into an electrical signal.
  • the converted electrical signal may be provided to the processor 270 or the controller 170.
  • the gesture input unit 212 may include at least one selected from among an infrared sensor and an image sensor for sensing a gesture input of a user.
  • the gesture input unit 212 may sense a three-dimensional (3D) gesture input of a user.
  • the gesture input unit 212 may include a plurality of light emitting units for outputting infrared light, or a plurality of image sensors.
  • the gesture input unit 212 may sense the 3D gesture input by employing a time of flight (TOF) scheme, a structured light scheme, or a disparity scheme.
  • TOF time of flight
  • the touch input unit 213 may convert a user's touch input into an electrical signal.
  • the converted electrical signal may be provided to the processor 270 or the controller 170.
  • the touch input unit 213 may include a touch sensor for sensing a touch input of a user.
  • the touch input unit 210 may be formed integral with a display unit 251 to implement a touch screen.
  • the touch screen may provide an input interface and an output interface between the vehicle 100 and the user.
  • the mechanical input unit 214 may include at least one selected from among a button, a dome switch, a jog wheel, and a jog switch. An electrical signal generated by the mechanical input unit 214 may be provided to the processor 270 or the controller 170.
  • the mechanical input unit 214 may be located on a steering wheel, a center fascia, a center console, a cockpit module, a door, etc.
  • the processor 270 may start a learning mode of the vehicle 100 in response to a user input to at least one of the afore-described voice input unit 211, gesture input unit 212, touch input unit 213, or mechanical input unit 214.
  • the vehicle 100 may learn a driving route and ambient environment of the vehicle 100. The learning mode will be described later in detail in relation to the object detection device 300 and the operation system 700.
  • the internal camera 220 may acquire images of the inside of the vehicle 100.
  • the processor 270 may sense a user's condition based on the images of the inside of the vehicle 100.
  • the processor 270 may acquire information on an eye gaze of the user.
  • the processor 270 may sense a gesture of the user from the images of the inside of the vehicle 100.
  • the biometric sensing unit 230 may acquire biometric information of the user.
  • the biometric sensing unit 230 may include a sensor for acquire biometric information of the user, and may utilize the sensor to acquire finger print information, heart rate information, etc. of the user.
  • the biometric information may be used for user authentication.
  • the output unit 250 is configured to generate a visual, audio, or tactile output.
  • the output unit 250 may include at least one selected from among a display unit 251, a sound output unit 252, and a haptic output unit 253.
  • the display unit 251 may display graphic objects corresponding to various types of information.
  • the display unit 251 may include at least one selected from among a Liquid Crystal Display (LCD), a Thin Film Transistor-Liquid Crystal Display (TFT LCD), an Organic Light-Emitting Diode (OLED), a flexible display, a 3D display, and an e-ink display.
  • LCD Liquid Crystal Display
  • TFT LCD Thin Film Transistor-Liquid Crystal Display
  • OLED Organic Light-Emitting Diode
  • the display unit 251 may form an inter-layer structure together with the touch input unit 213, or may be integrally formed with the touch input unit 213 to implement a touch screen.
  • the display unit 251 may be implemented as a head up display (HUD).
  • the display unit 251 may include a projector module in order to output information through an image projected on a windshield or a window.
  • the display unit 251 may include a transparent display.
  • the transparent display may be attached on the windshield or the window.
  • the transparent display may display a predetermined screen with a predetermined transparency.
  • the transparent display may include at least one selected from among a transparent Thin Film Electroluminescent (TFEL) display, an Organic Light Emitting Diode (OLED) display, a transparent Liquid Crystal Display (LCD), a transmissive transparent display, and a transparent Light Emitting Diode (LED) display.
  • TFEL Thin Film Electroluminescent
  • OLED Organic Light Emitting Diode
  • LCD transparent Liquid Crystal Display
  • LED transparent Light Emitting Diode
  • the transparency of the transparent display may be adjustable.
  • the user interface device 200 may include a plurality of display units 251a to 251g.
  • the display unit 251 may be disposed in a region of a steering wheel, a region 251a, 251b or 251e of an instrument panel, a region 251d of a seat, a region 251f of each pillar, a region 251g of a door, a region of a center console, a region of a head lining, a region of a sun visor, a region 251c of a windshield, or a region 251h of a window.
  • the sound output unit 252 converts an electrical signal from the processor 270 or the controller 170 into an audio signal, and outputs the audio signal. To this end, the sound output unit 252 may include one or more speakers.
  • the haptic output unit 253 generates a tactile output.
  • the haptic output unit 253 may operate to vibrate a steering wheel, a safety belt, and seats 110FL, 110FR, 110RL, and 110RR so as to allow a user to recognize the output.
  • the processor 270 may control the overall operation of each unit of the user interface device 200.
  • the user interface device 200 may include a plurality of processors 270 or may not include the processor 270.
  • the user interface device 200 may operate under control of the controller 170 or a processor of a different device inside the vehicle 100.
  • the user interface device 200 may be referred to as a display device for a vehicle.
  • the user interface device 200 may operate under control of the controller 170.
  • the object detection device 300 is used to detect an object outside the vehicle 100.
  • the object detection device 300 may generate object information based on sensing data.
  • the object information may include information about the presence of an object, location information of the object, information on distance between the vehicle and the object, and the speed of the object relative to the vehicle 100.
  • the object may include various objects related to travelling of the vehicle 100.
  • an object o may include a lane OB10, a nearby vehicle OB11, a pedestrian OB12, a two-wheeled vehicle OB13, a traffic signal OB14 and OB15, a light, a road, a structure, a bump, a geographical feature, an animal, etc.
  • the lane OB10 may be a lane in which the vehicle 100 is traveling (hereinafter, referred to as the current driving lane), a lane next to the current driving lane, and a lane in which a vehicle travelling in the opposite direction is travelling.
  • the lane OB10 may include left and right lines that define the lane.
  • the nearby vehicle OB11 may be a vehicle that is travelling in the vicinity of the vehicle 100.
  • the nearby vehicle OB11 may be a vehicle within a predetermined distance from the vehicle 100.
  • the nearby vehicle OB11 may be a vehicle that is preceding or following the vehicle 100.
  • the pedestrian OB12 may be a person in the vicinity of the vehicle 100.
  • the pedestrian OB12 may be a person within a predetermined distance from the vehicle 100.
  • the pedestrian OB12 may be a person on a sidewalk or on the roadway.
  • the two-wheeled vehicle OB13 is a vehicle that is located in the vicinity of the vehicle 100 and moves with two wheels.
  • the two-wheeled vehicle OB13 may be a vehicle that has two wheels within a predetermined distance from the vehicle 100.
  • the two-wheeled vehicle OB13 may be a motorcycle or a bike on a sidewalk or the roadway.
  • the traffic signal may include a traffic light OB15, a traffic sign plate OB14, and a pattern or text painted on a road surface.
  • the light may be light generated by a lamp provided in the nearby vehicle.
  • the light may be light generated by a street light.
  • the light may be solar light.
  • the road may include a road surface, a curve, and slopes, such as an upward slope and a downward slope.
  • the structure may be a body located around the road in the state of being fixed onto the ground.
  • the structure may include a streetlight, a roadside tree, a building, a traffic light, and a bridge.
  • the geographical feature may include a mountain and a hill.
  • the object may be classified as a movable object or a stationary object.
  • the movable object may include a nearby vehicle and a pedestrian.
  • the stationary object may include a traffic signal, a road, and a structure.
  • the object detection device 300 may include a camera 310, a radar 320, a LIDAR 330, an ultrasonic sensor 340, an infrared sensor 350, and a processor 370. Each component of the object detection device may be separated from or integrated with the sensing unit, structurally or operatively.
  • the object detection device 300 may further include other components in addition to the aforementioned components, or may not include some of the aforementioned components.
  • the camera 310 may be located at an appropriate position outside the vehicle 100 in order to acquire images of the outside of the vehicle 100.
  • the camera 310 may be a mono camera, a stereo camera 310a, an around view monitoring (AVM) camera 310b, or a 360-degree camera.
  • AVM around view monitoring
  • the camera 310 may acquire location information of an object, information on distance to the object, and information on speed relative to the object.
  • the camera 310 may acquire information on distance to the object and information on speed relative to the object.
  • the camera 310 may acquire the information on distance to the object and the information on speed relative to the object by utilizing a pin hole model or by profiling a road surface.
  • the camera 310 may acquire the information on distance to the object and the information on the speed relative to the object, based on information on disparity of stereo images acquired by a stereo camera 310a.
  • the camera 310 may be disposed near a front windshield in the vehicle 100 in order to acquire images of the front of the vehicle 100.
  • the camera 310 may be disposed around a front bumper or a radiator grill.
  • the camera 310 may be disposed near a rear glass in the vehicle 100 in order to acquire images of the rear of the vehicle 100.
  • the camera 310 may be disposed around a rear bumper, a trunk, or a tailgate.
  • the camera 310 may be disposed near at least one of the side windows in the vehicle 100 in order to acquire images of the side of the vehicle 100.
  • the camera 310 may be disposed around a side mirror, a fender, or a door.
  • the camera 310 may provide an acquired image to the processor 370.
  • the radar 320 may include an electromagnetic wave transmission unit and an electromagnetic wave reception unit.
  • the radar 320 may be realized as a pulse radar or a continuous wave radar depending on the principle of emission of an electronic wave.
  • the radar 320 may be realized as a Frequency Modulated Continuous Wave (FMCW) type radar or a Frequency Shift Keying (FSK) type radar depending on the waveform of a signal.
  • FMCW Frequency Modulated Continuous Wave
  • FSK Frequency Shift Keying
  • the radar 320 may detect an object through the medium of an electromagnetic wave by employing a time of flight (TOF) scheme or a phase-shift scheme, and may detect a location of the detected object, the distance to the detected object, and the speed relative to the detected object.
  • TOF time of flight
  • the radar 320 may be located at an appropriate position outside the vehicle 100 in order to sense an object located in front of the vehicle 100, an object located to the rear of the vehicle 100, or an object located to the side of the vehicle 100.
  • the LIDAR 330 may include a laser transmission unit and a laser reception unit.
  • the LIDAR 330 may be implemented by the TOF scheme or the phase-shift scheme.
  • the LIDAR 330 may be implemented as a drive type LIDAR or a non-drive type LIDAR.
  • the LIDAR 330 may rotate by a motor and detect an object in the vicinity of the vehicle 100.
  • the LIDAR 330 may utilize a light steering technique to detect an object located within a predetermined distance from the vehicle 100.
  • the LIDAR 330 may detect an object through the medium of laser light by employing the TOF scheme or the phase-shift scheme, and may detect a location of the detected object, the distance to the detected object, and the speed relative to the detected object.
  • the LIDAR 330 may be located at an appropriate position outside the vehicle 100 in order to sense an object located in front of the vehicle 100, an object located to the rear of the vehicle 100, or an object located to the side of the vehicle 100.
  • the ultrasonic sensor 340 may include an ultrasonic wave transmission unit and an ultrasonic wave reception unit.
  • the ultrasonic sensor 340 may detect an object based on an ultrasonic wave, and may detect a location of the detected object, the distance to the detected object, and the speed relative to the detected object.
  • the ultrasonic sensor 340 may be located at an appropriate position outside the vehicle 100 in order to detect an object located in front of the vehicle 100, an object located to the rear of the vehicle 100, and an object located to the side of the vehicle 100.
  • the infrared sensor 350 may include an infrared light transmission unit and an infrared light reception unit.
  • the infrared sensor 350 may detect an object based on infrared light, and may detect a location of the detected object, the distance to the detected object, and the speed relative to the detected object.
  • the infrared sensor 350 may be located at an appropriate position outside the vehicle 100 in order to sense an object located in front of the vehicle 100, an object located to the rear of the vehicle 100, or an object located to the side of the vehicle 100.
  • the processor 370 may control the overall operation of each unit of the object detection device 300.
  • the processor 370 may detect or classify an object by comparing data sensed by the camera 310, the radar 320, the LIDAR 330, the ultrasonic sensor 340, and the infrared sensor 350 with pre-stored data.
  • the processor 370 may detect and track an object based on acquired images.
  • the processor 370 may, for example, calculate the distance to the object and the speed relative to the object.
  • the processor 370 may acquire information on the distance to the object and information on the speed relative to the object based on a variation in size over time of the object in acquired images.
  • the processor 370 may acquire information on the distance to the object or information on the speed relative to the object by employing a pin hole model or by profiling a road surface.
  • the processor 370 may acquire information on the distance to the object and information on the speed relative to the object based on information on disparity of stereo images acquired from the stereo camera 310a.
  • the processor 370 may detect and track an object based on a reflection electromagnetic wave which is formed as a result of reflection a transmission electromagnetic wave by the object. Based on the electromagnetic wave, the processor 370 may, for example, calculate the distance to the object and the speed relative to the object.
  • the processor 370 may detect and track an object based on a reflection laser light which is formed as a result of reflection of transmission laser by the object. Based on the laser light, the processor 370 may, for example, calculate the distance to the object and the speed relative to the object.
  • the processor 370 may detect and track an object based on a reflection ultrasonic wave which is formed as a result of reflection of a transmission ultrasonic wave by the object. Based on the ultrasonic wave, the processor 370 may, for example, calculate the distance to the object and the speed relative to the object.
  • the processor 370 may detect and track an object based on reflection infrared light which is formed as a result of reflection of transmission infrared light by the object. Based on the infrared light, the processor 370 may, for example, calculate the distance to the object and the speed relative to the object.
  • the processor 370 may store data sensed by the camera 310, the radar 320, the LIDAR 330, the ultrasonic sensor 340, and the infrared sensor 350 in the memory 140.
  • the object detection device 300 may include a plurality of processors 370 or no processor 370.
  • the camera 310, the radar 320, the LIDAR 330, the ultrasonic sensor 340, and the infrared sensor 350 may include individual processors.
  • the object detection device 300 may operate under control of the controller 170 or a processor inside the vehicle 100.
  • the object detection device 300 may operate under control of the controller 170.
  • the communication device 400 is configured to perform communication with an external device.
  • the external device may be a nearby vehicle, a mobile terminal, or a server.
  • the communication device 400 may include at least one selected from among a transmission antenna, a reception antenna, a Radio Frequency (RF) circuit capable of implementing various communication protocols, and an RF device.
  • RF Radio Frequency
  • the communication device 400 may include a short-range communication unit 410, a location information unit 420, a V2X communication unit 430, an optical communication unit 440, a broadcast transmission and reception unit 450, an Intelligent Transport Systems (ITS) communication unit 460, and a processor 470.
  • a short-range communication unit 410 may include a short-range communication unit 410, a location information unit 420, a V2X communication unit 430, an optical communication unit 440, a broadcast transmission and reception unit 450, an Intelligent Transport Systems (ITS) communication unit 460, and a processor 470.
  • ITS Intelligent Transport Systems
  • the communication device 400 may further include other components in addition to the aforementioned components, or may not include some of the aforementioned components.
  • the short-range communication unit 410 is configured to perform short-range communication.
  • the short-range communication unit 410 may support short-range communication using at least one selected from among BluetoothTM, Radio Frequency IDdentification (RFID), Infrared Data Association (IrDA), Ultra-WideBand (UWB), ZigBee, Near Field Communication (NFC), Wireless-Fidelity (Wi-Fi), Wi-Fi Direct, and Wireless USB (Wireless Universal Serial Bus).
  • RFID Radio Frequency IDdentification
  • IrDA Infrared Data Association
  • UWB Ultra-WideBand
  • ZigBee Near Field Communication
  • NFC Near Field Communication
  • Wi-Fi Wireless-Fidelity
  • Wi-Fi Direct Wireless USB (Wireless Universal Serial Bus).
  • the short-range communication unit 410 may form wireless area networks to perform short-range communication between the vehicle 100 and at least one external device.
  • the location information unit 420 is configured to acquire location information of the vehicle 100.
  • the location information unit 420 may include a Global Positioning System (GPS) module or a Differential Global Positioning System (DGPS) module.
  • GPS Global Positioning System
  • DGPS Differential Global Positioning System
  • the V2X communication unit 430 is configured to perform wireless communication between a vehicle and a server (that is, vehicle to infra (V2I) communication), wireless communication between a vehicle and a nearby vehicle (that is, vehicle to vehicle (V2V) communication), or wireless communication between a vehicle and a pedestrian (that is, vehicle to pedestrian (V2P) communication).
  • V2I vehicle to infra
  • V2V vehicle to vehicle
  • V2P vehicle to pedestrian
  • the optical communication unit 440 is configured to perform communication with an external device through the medium of light.
  • the optical communication unit 440 may include a light emitting unit, which converts an electrical signal into an optical signal and transmits the optical signal to the outside, and a light receiving unit which converts a received optical signal into an electrical signal.
  • the light emitting unit may be integrally formed with a lamp provided included in the vehicle 100.
  • the broadcast transmission and reception unit 450 is configured to receive a broadcast signal from an external broadcasting management server or transmit a broadcast signal to the broadcasting management server through a broadcasting channel.
  • the broadcasting channel may include a satellite channel, and a terrestrial channel.
  • the broadcast signal may include a TV broadcast signal, a radio broadcast signal, and a data broadcast signal.
  • the ITS communication unit 460 may exchange information, data, or signals with a traffic system.
  • the ITS communication unit 460 may provide acquired information or data to the traffic system.
  • the ITS communication unit 460 may receive information, data, or signals from the traffic system.
  • the ITS communication unit 460 may receive traffic information from the traffic system and provide the traffic information to the controller 170.
  • the ITS communication unit 460 may receive a control signal from the traffic system, and provide the control signal to the controller 170 or a processor provided in the vehicle 100.
  • the processor 470 may control the overall operation of each unit of the communication device 400.
  • the communication device 400 may include a plurality of processors 470, or may not include the processor 470.
  • the communication device 400 may operate under control of the controller 170 or a processor of a device inside of the vehicle 100.
  • the communication device 400 may implement a vehicle display device, together with the user interface device 200.
  • the vehicle display device may be referred to as a telematics device or an audio video navigation (AVN) device.
  • APN audio video navigation
  • the communication device 400 may operate under control of the controller 170.
  • the maneuvering device 500 is configured to receive a user input for driving the vehicle 100.
  • the vehicle 100 may operate based on a signal provided by the maneuvering device 500.
  • the maneuvering device 500 may include a steering input device 510, an acceleration input device 530, and a brake input device 570.
  • the steering input device 510 may receive a user input with regard to the direction of travel of the vehicle 100.
  • the steering input device 510 may take the form of a wheel to enable a steering input through the rotation thereof.
  • the steering input device may be provided as a touchscreen, a touch pad, or a button.
  • the acceleration input device 530 may receive a user input for acceleration of the vehicle 100.
  • the brake input device 570 may receive a user input for deceleration of the vehicle 100.
  • Each of the acceleration input device 530 and the brake input device 570 may take the form of a pedal.
  • the acceleration input device or the break input device may be configured as a touch screen, a touch pad, or a button.
  • the maneuvering device 500 may operate under control of the controller 170.
  • the vehicle drive device 600 is configured to electrically control the operation of various devices of the vehicle 100.
  • the vehicle drive device 600 may include a power train drive unit 610, a chassis drive unit 620, a door/window drive unit 630, a safety apparatus drive unit 640, a lamp drive unit 650, and an air conditioner drive unit 660.
  • the vehicle drive device 600 may further include other components in addition to the aforementioned components, or may not include some of the aforementioned components.
  • the vehicle drive device 600 may include a processor. Each unit of the vehicle drive device 600 may include its own processor.
  • the power train drive unit 610 may control the operation of a power train.
  • the power train drive unit 610 may include a power source drive unit 611 and a transmission drive unit 612.
  • the power source drive unit 611 may control a power source of the vehicle 100.
  • the power source drive unit 611 may perform electronic control of the engine. As such the power source drive unit 611 may control, for example, the output torque of the engine. The power source drive unit 611 may adjust the output toque of the engine under control of the controller 170.
  • the power source drive unit 611 may control the motor.
  • the power train drive unit 610 may control, for example, the RPM and toque of the motor under control of the controller 170.
  • the transmission drive unit 612 may control a transmission.
  • the transmission drive unit 612 may adjust the state of the transmission.
  • the transmission drive unit 612 may adjust a state of the transmission to a drive (D), reverse (R), neutral (N), or park (P) state.
  • the transmission drive unit 612 may adjust a gear-engaged state to the drive position D.
  • the chassis drive unit 620 may control the operation of a chassis.
  • the chassis drive unit 620 may include a steering drive unit 621, a brake drive unit 622, and a suspension drive unit 623.
  • the steering drive unit 621 may perform electronic control of a steering apparatus provided inside the vehicle 100.
  • the steering drive unit 621 may change the direction of travel of the vehicle 100.
  • the brake drive unit 622 may perform electronic control of a brake apparatus provided inside the vehicle 100. For example, the brake drive unit 622 may reduce the speed of the vehicle 100 by controlling the operation of a brake located at a wheel.
  • the brake drive unit 622 may control a plurality of brakes individually.
  • the brake drive unit 622 may apply a different degree-braking force to each wheel.
  • the suspension drive unit 623 may perform electronic control of a suspension apparatus inside the vehicle 100. For example, when the road surface is uneven, the suspension drive unit 623 may control the suspension apparatus so as to reduce the vibration of the vehicle 100.
  • the suspension drive unit 623 may control a plurality of suspensions individually.
  • the door/window drive unit 630 may perform electronic control of a door apparatus or a window apparatus inside the vehicle 100.
  • the door/window drive unit 630 may include a door drive unit 631 and a window drive unit 632.
  • the door drive unit 631 may control the door apparatus.
  • the door drive unit 631 may control opening or closing of a plurality of doors included in the vehicle 100.
  • the door drive unit 631 may control opening or closing of a trunk or a tail gate.
  • the door drive unit 631 may control opening or closing of a sunroof.
  • the window drive unit 632 may perform electronic control of the window apparatus.
  • the window drive unit 632 may control opening or closing of a plurality of windows included in the vehicle 100.
  • the safety apparatus drive unit 640 may perform electronic control of various safety apparatuses provided inside the vehicle 100.
  • the safety apparatus drive unit 640 may include an airbag drive unit 641, a safety belt drive unit 642, and a pedestrian protection equipment drive unit 643.
  • the airbag drive unit 641 may perform electronic control of an airbag apparatus inside the vehicle 100. For example, upon detection of a dangerous situation, the airbag drive unit 641 may control an airbag to be deployed.
  • the safety belt drive unit 642 may perform electronic control of a seatbelt apparatus inside the vehicle 100. For example, upon detection of a dangerous situation, the safety belt drive unit 642 may control passengers to be fixed onto seats 110FL, 110FR, 110RL, and 110RR with safety belts.
  • the pedestrian protection equipment drive unit 643 may perform electronic control of a hood lift and a pedestrian airbag. For example, upon detection of a collision with a pedestrian, the pedestrian protection equipment drive unit 643 may control a hood lift and a pedestrian airbag to be deployed.
  • the lamp drive unit 650 may perform electronic control of various lamp apparatuses provided inside the vehicle 100.
  • the air conditioner drive unit 660 may perform electronic control of an air conditioner inside the vehicle 100. For example, when the inner temperature of the vehicle 100 is high, an air conditioner drive unit 660 may operate the air conditioner so as to supply cool air to the inside of the vehicle 100.
  • the vehicle drive device 600 may include a processor. Each unit of the vehicle drive device 600 may include its own processor.
  • the vehicle drive device 600 may operate under control of the controller 170.
  • the operation system 700 is a system for controlling the overall driving operation of the vehicle 100.
  • the operation system 700 may operate in the autonomous driving mode.
  • the operation system 700 may include the driving system 710, the vehicle pulling-out system 740, and the vehicle parking system 750.
  • the operation system 700 may further include other components in addition to the aforementioned components, or may not include some of the aforementioned component.
  • the operation system 700 may include a processor. Each unit of the operation system 700 may include its own processor.
  • the operation system 700 may control driving in the autonomous mode based on learning.
  • the learning mode and an operating mode based on the premise of completion of learning may be performed.
  • a description will be given below of a method of executing the learning mode and the operating mode by the processor of the operation system 700.
  • the learning mode may be performed in the afore-described manual mode.
  • the processor of the operation system 700 may learn a driving route and ambient environment of the vehicle 100.
  • the learning of the driving route may include generating map data for a route in which the vehicle 100 drives.
  • the processor of the operation system 700 may generate map data based on information detected through the object detection device 300 during driving from a departure to a destination.
  • the learning of the ambient environment may include storing and analyzing information about an ambient environment of the vehicle 100 during driving and parking.
  • the processor of the operation system 700 may store and analyze the information about the ambient environment of the vehicle based on information detected through the object detection device 300 during parking of the vehicle 100, for example, information about a location, size, and a fixed (or mobile) obstacle of a parking space.
  • the operating mode may be performed in the afore-described autonomous mode.
  • the operating mode will be described based on the premise that the driving route or the ambient environment has been learned in the learning mode.
  • the operating mode may be performed in response to a user input through the input unit 210, or when the vehicle 100 reaches the learned driving route and parking space, the operating mode may be performed automatically.
  • the operating mode may include a semi-autonomous operating mode requiring some user's manipulations of the maneuvering device 500, and a full autonomous operating mode requiring no user's manipulation of the maneuvering device 500.
  • the processor of the operation system 700 may drive the vehicle 100 along the learned driving route by controlling the driving system 710 in the operating mode.
  • the processor of the operation system 700 may pull out the vehicle 100 from the learned parking space by controlling the vehicle pulling-out system 740 in the operating mode.
  • the processor of the operation system 700 may park the vehicle 100 in the learned parking space by controlling the vehicle parking system 750 in the operating mode.
  • the operation system 700 may be a subordinate concept of the controller 170.
  • the operation system 700 may be a concept including at least one selected from among the user interface device 200, the object detection device 300, the communication device 400, the vehicle drive device 600, and the controller 170.
  • the driving system 710 may perform driving of the vehicle 100.
  • the driving system 710 may perform driving of the vehicle 100 by providing a control signal to the vehicle drive device 600 in response to reception of navigation information from the navigation system 770.
  • the driving system 710 may perform driving of the vehicle 100 by providing a control signal to the vehicle drive device 600 in response to reception of object information from the object detection device 300.
  • the driving system 710 may perform driving of the vehicle 100 by providing a control signal to the vehicle drive device 600 in response to reception of a signal from an external device through the communication device 400.
  • the driving system 710 may be a system that drives the vehicle 100, including at least one of the user interface device 200, the object detection device 300, the communication device 400, the maneuvering device 500, the vehicle drive device 600, the navigation system 770, the sensing unit 120, or the controller 170.
  • the driving system 710 may be referred to as a vehicle driving control device.
  • the vehicle pulling-out system 740 may perform an operation of pulling the vehicle 100 out of a parking space.
  • the vehicle pulling-out system 740 may perform an operation of pulling the vehicle 100 out of a parking space, by providing a control signal to the vehicle drive device 600 in response to reception of navigation information from the navigation system 770.
  • the vehicle pulling-out system 740 may perform an operation of pulling the vehicle 100 out of a parking space, by providing a control signal to the vehicle drive device 600 in response to reception of object information from the object detection device 300.
  • the vehicle pulling-out system 740 may perform an operation of pulling the vehicle 100 out of a parking space, by providing a control signal to the vehicle drive device 600 in response to reception of a signal from an external device.
  • the vehicle pulling-out system 740 may be a system that performs pulling-out of the vehicle 100, including at least one of the user interface device 200, the object detection device 300, the communication device 400, the maneuvering device 500, the vehicle drive device 600, the navigation system 770, the sensing unit 120, or the controller 170.
  • the vehicle pulling-out system 740 may be referred to as a vehicle pulling-out control device.
  • the vehicle parking system 750 may perform an operation of parking the vehicle 100 in a parking space.
  • the vehicle parking system 750 may perform an operation of parking the vehicle 100 in a parking space, by providing a control signal to the vehicle drive device 600 in response to reception of navigation information from the navigation system 770.
  • the vehicle parking system 750 may perform an operation of parking the vehicle 100 in a parking space, by providing a control signal to the vehicle drive device 600 in response to reception of object information from the object detection device 300.
  • the vehicle parking system 750 may perform an operation of parking the vehicle 100 in a parking space, by providing a control signal to the vehicle drive device 600 in response to reception of a signal from an external device.
  • the vehicle parking system 750 may be a system that performs parking of the vehicle 100, including at least one of the user interface device 200, the object detection device 300, the communication device 400, the maneuvering device 500, the vehicle drive device 600, the navigation system 770, the sensing unit 120, or the controller 170.
  • the vehicle parking system 750 may be referred to as a vehicle parking control device.
  • the navigation system 770 may provide navigation information.
  • the navigation information may include at least one selected from among map information, information on a set destination, information on a route to the set destination, information on various objects along the route, lane information, and information on a current location of the vehicle.
  • the navigation system 770 may include a memory and a processor.
  • the memory may store navigation information.
  • the processor may control the operation of the navigation system 770.
  • the navigation system 770 may update pre-stored information by receiving information from an external device through the communication device 400.
  • the navigation system 770 may be classified as an element of the user interface device 200.
  • FIG. 8 is a flowchart illustrating a method of learning-based vehicle parking method according to an implementation of the present disclosure.
  • a processor of the vehicle 100 as described below is a component corresponding to the controller 170 illustrated in FIG. 7.
  • step 810 the processor of the vehicle 100 learns a route in which the vehicle 100 has driven from a starting location to an ending location in the manual mode or the learning mode.
  • Step 810 may correspond to execution of the learning mode in the operation system 700.
  • Various implementations related to step 810 will be described in detail with reference to FIGS. 9 to 15.
  • step 820 the processor of the vehicle 100 starts driving of the vehicle 100 along the learned route in the operating mode as in step 820.
  • various implementations of step 820 will be described in detail with reference to FIGS. 16 to 21B.
  • step 830 the processor of the vehicle 100 parks the vehicle 100 in the detected parking slot as in step 830.
  • step 830 Various implementations of step 830 will be described in detail with reference to FIGS. 22, 23A, 23B, and 24A to 24D.
  • FIG. 9 is a view illustrating route learning of a processor of a vehicle in a manual mode or a learning mode according to an implementation of the present disclosure.
  • the vehicle 100 is shown as having entered a parking space 900 such as a public parking lot or a garage through an entrance 910.
  • the processor of the vehicle 100 initiates the learning mode at a starting location 920 in response to a user input received through the user interface device 200.
  • the learning mode may be performed in the manual mode in which the vehicle 100 is manually maneuvered by a user. That is, while the user is driving the vehicle 100 by manual maneuvering, a route 940 is learned in the learning mode.
  • Learning the route 940 includes generating map data for the route 940 in which the vehicle 100 drives.
  • the processor of the operation system 700 may generate map data based on information detected through the object detection device 300 during driving of the vehicle 100 from the starting location 920 to an ending location 930.
  • Learning the route 940 may be distinguished conceptually from learning a road in a space in which the vehicle 100 may drive, that is, the parking space 900. Learning the route 940 amounts to learning a driving trace of the vehicle 100 from the starting location 920 to the ending location 930.
  • learning of the space in which the vehicle 100 may drive amounts to learning a drivable space based on the overall width and length of the vehicle 100, the width of a road, and parking lines drawn on the floor surface of the parking space 900.
  • learning the space in which the vehicle 100 may drive may be included in the step of determining the property of an object.
  • learning the route 940 includes learning the driving trace of the vehicle 100 in the parking space 900 in a narrow sense
  • learning the route 940 may be understood as including even the concept of learning an object, a parking slot, and so on detected through the object detection device 300 in the parking space 900 by the vehicle in a broad sense.
  • the processor of the vehicle 100 may learn information detected through the object detection device 300, that is, information about the properties of objects by classifying the information into fixed information and movable information in the route learning step.
  • the fixed information refers to information which is not variable over time, such as a pillar in a parking lot, a tree, or the foregoing space in which the vehicle 100 may drive.
  • the fixed information is used when the vehicle 100 autonomously drives in the learned route.
  • the movable information refers to non-permanent information such as information about the location of any other parked vehicle.
  • the processor of the vehicle 100 may learn a location at which a parking line is not recognized as a parking-available location, based on information about another vehicle parked at the location.
  • the processor of the vehicle 100 may detect an empty parking slot using information related to the properties of objects stored in the route learning step. For example, the processor of the vehicle 100 may use not only the fixed information such as the area of the parking slot where the parking line is recognized and the space occupied by other parked vehicle, but also the movable information such as other parked vehicle itself.
  • the maximum speed of the vehicle 100 may be limited. That is, if the speed of the vehicle 100 becomes equal to or larger than a predetermined maximum speed, the processor of the vehicle 100 may output an alarm message to the user through the output unit 250, to thereby control cancellation of the learning.
  • a driving pattern of the user may also be learned in the route learning step.
  • the processor of the vehicle 100 may generate a speed profile of changes in the speed of the vehicle 100 in a driving route and store the speed profile in the memory 140, during learning.
  • the processor of the vehicle 100 may store, in the memory 140, information about manipulations of the steering wheel of the vehicle 100 in the driving route through a steering sensor, during learning.
  • a learned driving pattern including the speed profile or the information about manipulations of the steering wheel may be used in later autonomous driving of the vehicle 100 in the autonomous mode.
  • the learning also ends in response to a user input through the user interface device 200. That is, upon receipt of a user input for ending the learning when the vehicle 100 reaches the ending location 930, the processor of the vehicle 100 stores the route 940 learned from the starting location 920 to the ending location 930.
  • the initiation or ending of the learning may be performed through the recognition and determination of the situation by the processor of the vehicle 100 as well as the direct input through the physical button or touch display.
  • the processor of the vehicle 100 may determine a specific location within sensing area as the starting location or the ending location by recognizing a situation that all of the passengers except for the driver are unloaded and that the entrance is located within a predetermined distance from the vehicle 100 and that the vehicle 100 stopped at the parking slot and that the gear of the vehicle 100 is changed to the park (P) state and that the ignition switch is turned off.
  • the ending location 930 of the route 940 may be substantially the same as the starting location 920. In this case, when the processor of the vehicle 100 recognizes that the vehicle 100 arrives at the starting location 920, the processor may end the route learning without user input.
  • the route 940 learned in the parking space 900 is used for parking and pulling-out of the vehicle 100, which will be described in detail below.
  • FIG. 10 illustrates route learning of the processor of the vehicle 100 in the manual mode or the learning mode according to another implementation of the present disclosure.
  • each floor is designed to have the same structure in a multi-story parking facility such as a tower parking system. Accordingly, it may not be necessary to learn a route on every floor.
  • the processor of the vehicle 100 learns a route 1040 on one floor, as illustrated in FIG. 10(a). Subsequently, the processor of the vehicle 100 may store a route 1050 on another floor as the same as the route 1040, as illustrated in FIG. 10(b).
  • FIG. 11 is a view illustrating route learning of the processor of the vehicle 100 in the manual mode or the learning mode according to another implementation of the present disclosure.
  • FIG. 11 further illustrates correction of the learned route and input of additional information by a user input through the user interface device 200 in addition to the route learning.
  • FIG. 11 illustrates an outside image 1100 of the vehicle 100, captured by the camera 310 of the object detection device 300.
  • the processor of the vehicle 100 may correct a learned route and receive additional information. The following may be taken as examples of the rout correction and the input of additional information.
  • the specific location may be set as a parking-available location.
  • information indicating whether the obstacle 1120 is a temporary or permanent obstacle may be added.
  • the processor of the vehicle 100 may learn the route 1130 as an available route.
  • various pieces of information such as the speed of the vehicle 100 and the distance between vehicles may be added in response to a user input received through the user interface device 200.
  • FIG. 12 illustrates route learning of the processor of the vehicle 100 in the manual mode or the learning mode according to another implementation of the present disclosure.
  • the route learning step may further include a step of storing, as a learned route, a route other than a route in which the vehicle 100 has driven between a first location 1230 and a second location 1240 based on determined properties of objects between the first location 1230 and the second location 1240 in the driven route.
  • a route 1210 and the route 1250 illustrated in FIG. 12(a) are routes in which the vehicle 100 has driven. Further, FIG. 12(a) illustrates driving of the vehicle 100 in the alternative route 1250 from the first location 1230 to the second location 1240 due to an obstacle 1220.
  • the processor of the vehicle 100 stores the alternative route 1250 in which the vehicle 100 has actually driven, and another route 1260 between the first location 1230 and the second location 1240, as learned routes.
  • the other route 1260 may be, for example, a shortest route, a minimum-time route, or a safer route, which connects the first location 1230 to the second location 1240.
  • the processor of the vehicle 100 may perceive a route in which the vehicle 100 has driven as a straight line or set the route as a shortest route.
  • the implementation of the present disclosure illustrated in FIG. 12 achieves the technical effect that an optimum driving route may be provided for the vehicle 100 in the autonomous mode by modifying a route learned according to the type of an obstacle as well as an actually driven route.
  • FIG. 13 is a view illustrating route learning of the processor of the vehicle 100 in the manual mode or the learning mode according to another implementation of the present disclosure.
  • FIG. 13 illustrates arrival of the vehicle 100 at a branch point 1320 in the route learning step by the processor of the vehicle 100.
  • the branch point (or intersection) is a point at which two or more paths meet or intersect.
  • the object detection device 300 of the vehicle 100 detects an available space to the left of the vehicle 100 at the branch point 1320.
  • the camera 310, the radar 320, the LIDAR 330, the ultrasonic sensor 340, and the IR sensor 350 may be used.
  • the processor of the vehicle 100 may recognize the branch point 1320 when the processor detects the movable object entering a path that the vehicle 100 does not currently travel or the other vehicles parked along the path that the vehicle 100 does not currently travel.
  • the processor of the vehicle 100 stores information about the location of the branch point 1320, and information about a direction to the available space with respect to a driving route.
  • the processor of the vehicle 100 stores a new route connecting the two branch points 1320 and 1330 as a learned route in response to a user input received through the user interface device 200. Or, the processor of the vehicle 100 may autonomously store the new route connecting the two branch points 1320 and 1330 as a learned route.
  • FIG. 14 is a view illustrating route learning of the processor of the vehicle 100 in the manual mode or the learning mode according to another implementation of the present disclosure.
  • a parking space 1400 may be limited to a one-way driving direction.
  • traffic is limited to one-way traffic in the clockwise direction in the parking space 1400 of FIG. 14.
  • the processor of the vehicle 100 notifies the user that the vehicle 100 is driving in the wrong direction by controlling the output unit 250 of the user interface device 200.
  • FIG. 15 illustrates route learning of the processor of the vehicle 100 in the manual mode or the learning mode according to another implementation of the present disclosure.
  • one driving route is learned in one parking space.
  • a plurality of partial routes including a starting location and an ending location may be learned even in one parking space.
  • the processor of the vehicle 100 may learn two partial routes 1510 and 1520 having a starting location and an ending location. As illustrated in FIG. 15(b), the processor of the vehicle 100 may generate a full route by incorporating the partial route 1510 with the partial route 1520 in response to a user input received through the user interface device 200, and store the generated full route in the memory 140. Or the processor of the vehicle 100 may autonomously generate the full route and store the generated full route in the memory 140, without a user input.
  • the processor of the vehicle 100 may receive information indicating whether to learn the partial route 1520 successively to non-completed learning of the partial route 1510 from the user through the user interface device 200. Upon receipt of the user input, the processor of the vehicle 100 may generate a full route by incorporating the partial route 1510 with the partial route 1520, and store the generated full route in the memory 140.
  • the processor of the vehicle 100 may merge the first learned route and the second learned route.
  • the processor of the vehicle 100 can use both the first learned route and the second learned route regardless of whether the starting location of the autonomous parking is the first starting location or the second starting location.
  • FIG. 16 is a view illustrating driving of the vehicle 100 in a learned route by the processor of the vehicle 100 according to an implementation of the present disclosure.
  • the following conditions may be required to enable the vehicle 100 to start autonomous driving along a learned route 1640.
  • the vehicle 100 may reach within a predetermined distance (e.g., 2m) from a starting location 1620.
  • the current location of the vehicle 100 may be detected through the camera 310 and the location information unit 420.
  • the processor of the vehicle 100 may pre-store semantic information such as a landmark in an image captured through the camera 310, and extracts the semantic information from an image captured through the camera 310 at the current location of the vehicle 100.
  • the processor of the vehicle 100 compares the detected current location of the vehicle 100 with the starting location 1620 in the above operation.
  • a user input through the user interface device 200 may be required, such as a button input for initiating autonomous driving. This is done to grant user input-based reliability to initiation of autonomous driving.
  • both the conditions may be required, or only one of the conditions may be selectively required.
  • the processor of the vehicle 100 detects an empty parking slot during autonomous driving in the learned route 1640 from the starting location 1620 to the ending location 1630.
  • the maximum speed of the vehicle 100 may be limited.
  • FIG. 17 is a view illustrating driving of the vehicle 100 in a learned route by the processor of the vehicle 100 according to an implementation of the present disclosure.
  • the processor of the vehicle 100 may fail to detect an empty parking slot while the vehicle 100 is driving along a learned route 1740 from a starting location 1720 to an ending location 1730.
  • the processor of the vehicle 100 may pull over the vehicle 100 to the ending location 1730, and notify the user of failure in detecting a parking slot through the output unit 250.
  • FIG. 18 illustrates driving of the vehicle 100 in a learned route by the processor of the vehicle 100 according to an implementation of the present disclosure.
  • FIG. 18 illustrates generation of a new route in the autonomous mode, because of an obstacle 1810 which has been perceived as a fixed obstacle in the learning mode and disappeared with passage of time.
  • FIG. 18(a) illustrates route learning of the vehicle 100 by bypassing a fixed obstacle 1810 in the learning mode.
  • FIG. 18(b) illustrates driving of the vehicle 100 in a new route in the autonomous mode, due to disappearance of the fixed obstacle 1810.
  • the vehicle 100 stops at a location where the fixed obstacle 1810 used to be.
  • the processor of the vehicle 100 notifies the user of disappearance of the fixed obstacle 1810 through the output unit 250.
  • the processor of the vehicle 100 requests the user to confirm whether to generate a new route 1820 through the output unit 250.
  • the processor of the vehicle 100 Upon confirmation of generation of the new route 1820 by a user input through the user interface device 200, the processor of the vehicle 100 generates the new route 1820, and drives the vehicle 100 along the generated new route 1820.
  • the processor of the vehicle 100 may calculate an expected time to collide with the obstacle based on the current speed of the vehicle 100 and the distance to the obstacle. If the expected time to collide is larger than a predetermined value, that is, a collision risk is not very high, the processor of the vehicle 100 may control deceleration of the vehicle 100 without stopping the vehicle 100 immediately.
  • FIG. 19 illustrates driving of the vehicle 100 in a learned route by the processor of the vehicle 100 according to an implementation of the present disclosure.
  • FIG. 19 describes the vehicle 100 driving in a learned route in a multi-story parking facility such as a tower parking system.
  • FIG. 19(a) illustrates driving of the vehicle 100 in a learned route 1910 in order from the first basement to the third basement.
  • the communication device 400 of the vehicle 100 may receive additional information about use of the parking facility from an external device such as a parking lot server.
  • the communication device 400 of the vehicle 100 receives additional information indicating that parking is not allowed in the first and second basements from the parking lot server, by way of example.
  • the processor of the vehicle 100 may modify the learned route by reflecting the received additional information. That is, as illustrated in FIG. 19(b), the processor of the vehicle 100 may generate a new route 1920 which directly runs to the third basement, without driving in the learned route 1910 in the first and second basements.
  • the processor of the vehicle 100 controls driving of the vehicle 100 along the generated route 1920.
  • the implementation of the present disclosure illustrated in FIG. 19 achieves the technical effect that instead of unconditionally driving in a learned route from a starting location to an ending location, the vehicle 100 generates an optimum driving route based on information received from an external device.
  • FIG. 20 is a view illustrating driving of the vehicle 100 in a learned route by the processor of the vehicle 100 according to an implementation of the present disclosure.
  • FIG. 20 illustrates adaptive change of a driving route of the vehicle 100 in the autonomous mode based on information input directly by a user.
  • the user may set an exclusion list 2000 illustrated in FIGS. 10A and 10B and store the exclusion list 2000 in the memory 140.
  • the exclusion list 2000 includes information indicating parking not allowed on the first basement on March 1 to March 3, and information indicating parking not allowed on the second basement on April 1 to April 8.
  • the processor of the vehicle 100 may generate a first parking profile (parking profile 1) 2010 to a second parking profile (parking profile 2) 2020 on a date basis by reflecting the set exclusion list 2000.
  • the processor of the vehicle 100 may add information indicating parking not allowed on the second basement by reflecting information of the exclusion list 2000 in parking profile 1 2010 for April 3. Further, the processor of the vehicle 100 may add information indicating parking allowed on all of the first, second, and third basements by reflecting the information of the exclusion list 2000 in parking profile 2 2020 for April 10.
  • the processor of the vehicle 100 may generate per-date parking profiles by reflecting information received from the user through the user interface device 200, and adaptively change a driving route of the vehicle 100 in the autonomous mode by applying the per-date parking profiles.
  • FIG. 21 illustrates driving of the vehicle 100 in a learned route by the processor of the vehicle 100 according to an implementation of the present disclosure.
  • FIG. 21 illustrates the vehicle 100 that drives in a route modified by reflecting various pieces of information, rather than the vehicle 100 unconditionally drives from a starting location to an ending location along a learned route.
  • the processor of the vehicle 100 may additionally learn an available parking space by time or day and reflect the learned parking space in a parking route.
  • the processor of the vehicle 100 may generate a route 2120 which directly runs to section C expected to have lots of current empty parking slots as illustrated in FIG. 21(b), instead of driving along a learned route 2110 as illustrated in FIG. 21(a).
  • a user input received through the user interface device 200 may be used in this operation.
  • the processor of the vehicle 100 may learn user parking patterns by time or by day and reflect the learned user parking patterns in a parking route.
  • the processor of the vehicle 100 may learn a related parking pattern and drive the vehicle 100 along the route 2110, as illustrated in FIG. 21(a).
  • the processor of the vehicle 100 may learn a related parking pattern and drive the vehicle 100 along the route 2120, as illustrated in FIG. 21(b).
  • the processor of the vehicle 100 may learn the parking pattern of the vehicle 100 in the autonomous mode in the learned route.
  • FIG. 22 is a view illustrating parking of the vehicle 100 in an empty parking slot in a learned route by the processor of the vehicle 100 according to an implementation of the present disclosure.
  • the processor of the vehicle 100 searches for an empty parking slot by controlling the object detection device 300.
  • the processor of the vehicle 100 detects the parking slot 2240 by controlling the camera 310, the radar 320, the LIDAR 330, the ultrasonic sensor 340, and the IR sensor 350 of the object detection device 300.
  • the processor of the vehicle 100 may detect the empty parking slot 2240, using information about the properties of objects stored in the route learning step. For example, the processor of the vehicle 100 may detect the empty parking slot 2240 by detecting a parking line or a location at which another vehicle was parked in the afore-described learning step. For example, the vehicle 100 may search empty parking slots while traveling along the space stored in the map data as a parking slot and along the parking slot in which other vehicles was parked.
  • the processor of the vehicle 100 may receive information related to a parking slot from an infrastructure through the V2X communication unit 430. Based on the received information, vehicle 100 may search the empty parking slots.
  • the processor of the vehicle 100 may detect the empty parking slot 2240 more accurately using information about an obstacle in the route, stored in the learning step as well as the result of detection of the object detection device 300.
  • the processor of the vehicle 100 automatically parks the vehicle 100 in the detected parking slot 2240 by controlling the afore-described parking system 750.
  • a step of selecting a specific parking slot may be added.
  • the processor of the vehicle 100 may preset the priorities of parking slots. For example, the parking slots may be prioritized in an order of the distances from an exit to the parking slots. In this case, upon detection of a plurality of parking slots, the processor of the vehicle 100 may control parking of the vehicle 100 in a parking slot nearer to the exit.
  • various factors such as whether there is a parking line in the parking slot, the size of the parking slot, frequently parked parking slot and parking slot for disabled may be considered as priorities for the parking slot.
  • the processor of the vehicle 100 may store information about the location of the empty parking slot 2240. When the vehicle 100 is to be pulled out, the processor of the vehicle 100 may determine whether to return to the starting location 2210 along the route 2230 in which the vehicle 100 has driven for parking or go to the ending location 2220 along the other route, based on the stored location information.
  • FIG. 23 illustrates parking of the vehicle 100 in an empty parking slot in a learned route by the processor of the vehicle 100 according to an implementation of the present disclosure.
  • FIG. 23 depicts a method for modifying a learned parking route in the presence of an obstacle absent in the learning step during parking, for a designated or reserved parking slot of a personal parking space such as a garage.
  • an ending location 2320 corresponds to the location of the reserved parking slot.
  • the processor of the vehicle 100 determines an available area for parking the vehicle 100 in consideration of an area occupied by the object in the reserved parking slot. Subsequently, the processor of the vehicle 100 parks the vehicle 100 in the available area.
  • FIG. 23(a) illustrates driving of the vehicle 100 along a learned driving route 2330 from a starting location 2310 to the ending location 2320. More specifically, unlike the afore-described implementation of FIG. 22, FIG. 23(a) illustrates that an obstacle 2340 absent in the learning step obstructs automatic parking of the vehicle 100 at the ending location 2320.
  • the processor of the vehicle 100 detects the obstacle 2340 through the object detection device 300 as illustrated in FIG. 23(b).
  • the processor of the vehicle 100 determines whether the vehicle 100 may be parked in an area except for an area occupied by the obstacle 2340 in the parking slot.
  • the processor of the vehicle 100 controls parking of the vehicle 100 by generating a new route 2350 which allows parking with the obstacle 2340 avoided.
  • FIG. 24 illustrates parking of the vehicle 100 in an empty parking slot in a learned route by the processor of the vehicle 100 according to an implementation of the present disclosure.
  • FIG. 24 illustrates other implementations of the present disclosure related to FIG. 23. For example, in the presence of an obstacle which has not existed in a learned parking slot in the learning step, a new route for parking through obstacle avoidance is generated in FIG. 24.
  • parking slot illustrated in FIG. 23 is in a personal parking space such as a garage, there is no need for considering other vehicles.
  • parking slots illustrated in FIG. 24 is in a public parking lot. Thus, it is important not to obstruct parking and pulling-out of other vehicles.
  • FIG. 24 illustrates the presence of an obstacle 2410 which has not existed in the learning step, in a parking slot.
  • the processor of the vehicle 100 controls parking of the vehicle 100 in a parking slot next to the parking slot having the obstacle 2410 by recognizing parking lines as illustrated in FIGS. 24(c) and 24(d), rather than the processor of the vehicle 100 simply generates a parking trace for avoiding the obstacle as illustrated in FIGS. 24(a) and 24(b).
  • FIGS. 24(a) to 24(d) achieves the technical effect that driving and parking of other vehicles are not obstructed by perceiving the ambient space of a vehicle, in obstacle-avoiding parking.
  • FIG. 25 is a view illustrating a learning-based vehicle pulling-out method according to an implementation of the present disclosure.
  • FIG. 25 illustrates a pulling-out process after the vehicle 100 was parked in a detected parking slot according to the implementations of FIGS. 22 to 24.
  • the processor of the vehicle 100 Before the pulling-out process, when parking the vehicle 100 in a detected parking slot 2520, the processor of the vehicle 100 stores information about the location of the detected parking slot 2520 in the memory 140.
  • the processor of the vehicle 100 receives a user input for pulling out the vehicle 100 through the user interface device 200.
  • the processor of the vehicle 100 reads out the information about the location of the parking slot 2520 in which the vehicle 100 has been parked, and information about a learned route 2530 from the memory 140 in response to the user input.
  • the processor of the vehicle 100 generates the pull-out route 2530 running from the parking slot to an ending location, including at least a part of the learned route, and controls pulling-out of the vehicle 100 along the generated pull-out route 2530.
  • the processor of the vehicle 100 may store information about a parking trace in the parking step, and control pulling-out of the vehicle 100, using the stored information about the parking trace during pull-out.
  • the vehicle 100 may occur that the vehicle 100 returns to a starting location before the vehicle 100 is parked in a parking slot, unlike FIG. 25. For example, after the user gets off the vehicle 100 at the starting location and the vehicle 100 starts to drive in a learned route in the autonomous mode, the user may want the vehicle 100 to return to the starting location again.
  • the processor of the vehicle 100 receives a user input requesting the vehicle 100 to return to the starting location or the ending location through the user interface device 200.
  • the request is referred to as a roll-back request.
  • the processor of the vehicle 100 Upon receipt of the roll-back request, the processor of the vehicle 100 determines the location of the vehicle 100 in learned routes 2510 and 2530. The processor of the vehicle 100 determines whether to continue driving in a current driving direction or drive in a direction opposite to the current driving direction, based on the determined location of the vehicle 100.
  • the processor of the vehicle 100 may continue to travel keeping the driving direction.
  • the processor of the vehicle 100 may control the vehicle 100 to travel to the starting location again by changing the driving direction.
  • the distance from the starting location to the current location of the vehicle 100 may be the distance of the learned route 2510 traveled by the vehicle 100. In this case, the distance from the current location of the vehicle 100 to the ending location may be the distance of the learned route 2530.
  • the distance from the starting location to the current location of the vehicle 100 may corresponds to an elapsed time for travelling from the starting location to the current location of the vehicle 100.
  • the distance from the current location of the vehicle 100 to the ending location may correspond to an expected time for reaching the ending location from the current location of the vehicle 100.
  • the processor of the vehicle 100 may control the vehicle 100 to reach the ending location when it is determined that it is impossible to change the driving direction due to the rear obstacle or one-way road.
  • the processor of the vehicle 100 responds to the roll-back request so that the vehicle 100 reaches the starting location or the location designated by the user regardless of the current location of the vehicle 100 .
  • the processor of the vehicle 100 may control the vehicle 100 to reach the ending location even when receiving the roll-back request from the user.
  • various factors for determining whether to maintain the driving direction can be considered. For example, whether a roll-back request is received within a predetermined time (for example, one minute), the remaining time until parking is completed, the progress indicating the parking process, the distance that the vehicle 100 is apart from the user and the like can be considered.
  • a predetermined time for example, one minute
  • the processor of the vehicle 100 stops the vehicle 100 and detects a space in which the driving direction is to be changed, through the object detection device 300. Finally, the processor of the vehicle 100 controls change of the driving direction of the vehicle 100 in the detected space, and drives the vehicle 100 to return to the starting location or the ending location.
  • the processor of the vehicle 100 may control the transmission drive unit 612 to adjust the transmission state from the forward direction D to the reverse direction R and control the vehicle 100 to return to the starting location.
  • the roll-back process in the parking process of the vehicle 100 has been described above.
  • the roll-back process in the pulling-out process will be described.
  • the user may want the vehicle 100 to return to the parking slot that the vehicle 100 was parked and to be parked again.
  • the processor of the vehicle 100 may control the vehicle 100 to return to the parking slot that the vehicle 100 was parked and to be parked again.
  • the processor of the vehicle 100 may detect another available parking slot and control the vehicle 100 to be parked in the another parking slot.
  • the processor of the vehicle 100 may set the above mentioned 'predetermined distance' differently based on whether the starting location and the ending location are the same(including substantially same) or not.
  • the processor of the vehicle 100 may control the vehicle 100 to search for an available parking slot while traveling toward the ending location.
  • FIG. 26 is a view illustrating a learning-based vehicle pulling-out method according to another implementation of the present disclosure.
  • FIG. 26 illustrates generation of a new pull-out route, in the case where the vehicle 100 may not pull out of a parking slot using information about a parking trace stored in the parking step due to one other vehicle 2610 in front of the parking slot.
  • the object detection device 300 of the vehicle 100 may detect an object outside the vehicle 100.
  • the processor of the vehicle 100 may acquire information indicating the presence of the other vehicle 2610 in front of the vehicle 100 and information indicating the absence of an obstacle behind the vehicle 100 through the object detection device 300.
  • the processor of the vehicle 100 generates a new driving route 2620 in which the vehicle 100 may move backward and then return to a learned driving route 2630. Then, the processor of the vehicle 100 controls driving of the vehicle 100 along the generated driving route 2620.
  • the processor of the vehicle 100 determines the current location of the vehicle 100 through the object detection device 300, the location information unit 420, and so on.
  • the processor of the vehicle 100 determines whether meaningful information such as a corner or a landmark like a geographical feature stored in the learning step has been detected in an image captured through the camera 310 at the current location of the vehicle 100. Upon detection of the meaningful information, the processor of the vehicle 100 may determine that the processor 100 has reached the learned route 2630.
  • the processor of the vehicle 100 controls pulling-out of the vehicle 100 along the existing learned driving route 2630.
  • the present disclosure according to the implementation of FIG. 26 achieves the technical effect that a pull-out route changed adaptively according to an ambient environment condition of a vehicle is generated during pull-out.
  • FIG. 27 is a view illustrating a learning-based vehicle pulling-out method according to another implementation of the present disclosure.
  • FIG. 27 illustrates a case in which the location of the vehicle 100 has been changed after the vehicle 100 was parked, like forward or backward movement of the vehicle 100 by external force after double-parking of the vehicle 100.
  • the processor of the vehicle 100 may determine the changed location of the vehicle 100 through the object detection device 300.
  • the processor of the vehicle 100 controls pulling-out of the vehicle 100 along an existing learned route 2710.
  • the processor of the vehicle 100 determines whether the changed location of the vehicle 100 is outside the existing learned route, that is, if the vehicle 100 moves backward to position (b) in FIG. 27, the processor of the vehicle 100 generates a route in which to reach the existing learned route 2710, and controls pulling-out of the vehicle 100 along the generated route and the existing learned route 2710, in the method described with reference to FIG. 26.
  • FIG. 28 illustrates a learning-based vehicle pulling-out method according to another implementation of the present disclosure.
  • the example shown in FIG. 28 is related to the example shown in FIG. 12.
  • the processor of the vehicle 100 stores, as a learned route, a shortest route instead of the route in which the vehicle 100 has driven, avoiding the obstacle.
  • the processor of the vehicle 100 learns a driving route that avoids an obstacle 2810 in the learning step in FIG. 28(a). Then, if the processor of the vehicle 100 determines that the object 2810 has disappeared, through the object detection device 300 in the pull-out step, the processor of the vehicle 100 generates a shortest route 2830, not the existing learned route 2820, and controls pulling-out of the vehicle 100 along the generated shortest route 2830, as illustrated in FIG. 28(b).
  • FIG. 12 depicts a method of storing a learned route as a shortest route
  • FIG. 28 depicts a method of generating a shortest route during pull-out, while storing a learned route as an avoidance driving route.
  • FIG. 29 illustrates a learning-based vehicle pulling-out method according to another implementation of the present disclosure.
  • the processor of the vehicle 100 may automatically change a pull-out route by reflecting changed ambient environment information of the vehicle 100. That is, the processor of the vehicle 100 may detect an object outside the vehicle 100 through the object detection device 300, and generate a favorable route to reach a learned route based on the detected object.
  • FIGS. 29(a) and 29(b) illustrate vertical parking of the vehicle 100
  • FIGS. 29(c) and 29(d) illustrate parallel parking of the vehicle 100.
  • FIG. 29(a) illustrates that a pull-out route 2920 is learned due to one other vehicle 2910 parked to the right of the vehicle 100.
  • the processor of the vehicle 100 controls pulling-out of the vehicle 100 by generating an easier or more favorable pull-out route 2930.
  • FIG. 29(c) illustrates that a pull-out route 2950 is learned due to one other vehicle 2940 parked in front of the vehicle 100.
  • the processor of the vehicle 100 controls pulling-out of the vehicle 100 by generating an easier or more favorable pull-out route 2960.
  • the present disclosure illustrated in FIG. 29 achieves the technical effect that if a more favorable pull-out route than a learned pull-out route may be generated, an adaptively changed pull-out route is generated.
  • FIGS. 30 and 31 are views illustrating learning-based vehicle pulling-out methods according to other implementations of the present disclosure.
  • FIG. 30 illustrates that when the vehicle 100 identifies a user 3020 in the learned driving route 3010 by the camera 310 or the communication device 400 during driving in a learned driving route 3010, the vehicle 100 stops in front of the identified user 3020 without driving to an ending location 3030.
  • a GPS module of the location information unit 420 or a method such as facial recognition or voice recognition may be used.
  • FIG. 31 illustrates a case in which a user 3120 is located outside a learned driving route.
  • the vehicle 100 generates a new route 3140 to stop in front of the identified user 3120, without driving to an ending location 3130.
  • the processor of the vehicle 100 first determines whether a new route 3140 can be generated by using a branch point or the like. When it is determined that the new route 3140 can be generated, the processor of the vehicle 100 may control the vehicle 100 to stop in front of the user using the generated new route 3140.
  • the processor of the vehicle 100 may designate a boarding location (e.g. a location closest to the user 3120) on the learned route to where the user will board, or may receive a boarding location from the user. the processor of the vehicle 100 may control the vehicle 100 to stop at the designated or the received boarding location.
  • a boarding location e.g. a location closest to the user 3120
  • an autonomous vehicle may actively learn at least one parking slot in a route learning step.
  • the autonomous vehicle may detect an empty parking slot, and even in the absence of the history of previous parking in the parking slot, the autonomous vehicle may be parked in the parking slot.
  • the vehicle may be pulled out of a parking slot in a route other than a route used for parking in a learned route.
  • the above-described present disclosure may be implemented as code that can be written on a computer-readable recording medium and thus read by a computer system.
  • the computer-readable recording medium may be any type of recording device in which data is stored in a computer-readable manner. Examples of the computer-readable recording medium include a hard disk drive (HDD), a solid state disk (SSD), a silicon disk drive (SDD), a read only memory (ROM), a random access memory (RAM), a compact disk ROM (CD-ROM), a magnetic tape, a floppy disc, an optical data storage, and a carrier wave (e.g., data transmission over the Internet).
  • the computer may include a processor or a controller.
PCT/KR2018/011403 2017-05-12 2018-09-27 AUTONOMOUS VEHICLE AND ITS CONTROL METHOD WO2019066477A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201880063165.XA CN111148674A (zh) 2017-05-12 2018-09-27 自动驾驶车辆及其控制方法

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US201762505129P 2017-05-12 2017-05-12
KR10-2017-0125311 2017-09-27
KR1020170125311A KR102120108B1 (ko) 2017-05-12 2017-09-27 자율주행 차량 및 그 제어 방법

Publications (1)

Publication Number Publication Date
WO2019066477A1 true WO2019066477A1 (en) 2019-04-04

Family

ID=64602197

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2018/011403 WO2019066477A1 (en) 2017-05-12 2018-09-27 AUTONOMOUS VEHICLE AND ITS CONTROL METHOD

Country Status (3)

Country Link
KR (1) KR102120108B1 (ko)
CN (1) CN111148674A (ko)
WO (1) WO2019066477A1 (ko)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112109700A (zh) * 2020-06-18 2020-12-22 上汽通用五菱汽车股份有限公司 记忆泊车系统及其控制方法和计算机存储介质
EP3862996A1 (en) * 2020-02-07 2021-08-11 Volvo Car Corporation Automatic parking assistance system, in-vehicle device and method

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102025809B1 (ko) * 2017-09-29 2019-09-26 엘지전자 주식회사 차량 주행 제어 장치 및 차량 주행 방법
CN111959497A (zh) * 2020-06-29 2020-11-20 北京百度网讯科技有限公司 自动泊车方法和装置、电子设备、存储介质
CN112365637B (zh) * 2020-11-09 2021-08-17 成都中科大旗软件股份有限公司 基于门票销售平台的开车入园监测系统及监测方法
CN114394098B (zh) * 2021-04-16 2023-06-23 阿波罗智联(北京)科技有限公司 车辆停靠的方法、装置、电子设备和可读存储介质
KR102330985B1 (ko) 2021-06-02 2021-12-01 주식회사 엠데이터싱크 주차장에서의 주차 안내 방법
CN113799770B (zh) * 2021-10-28 2023-03-31 广州小鹏自动驾驶科技有限公司 一种基于自动驾驶的数据处理方法和装置
KR20230098100A (ko) 2021-12-24 2023-07-03 (주)스마트모빌리티랩 자율주행 하위 제어 시스템

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2016033500A (ja) * 2014-07-31 2016-03-10 トヨタ自動車株式会社 出発地・目的地抽出装置、出発地・目的地抽出方法
KR20160066115A (ko) * 2014-12-01 2016-06-10 호서대학교 산학협력단 주차위치 자동 안내시스템
KR101637842B1 (ko) * 2015-07-08 2016-07-07 현대자동차주식회사 주차장 내 자율주행 시스템 및 방법
US9557741B1 (en) * 2015-08-24 2017-01-31 Ford Global Technologies, Llc System and method for autonomous valet parking using plenoptic cameras

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0769741B2 (ja) * 1992-09-07 1995-07-31 株式会社テレシステムズ ゴルフカート制御システム
DE102010023162A1 (de) * 2010-06-09 2011-12-15 Valeo Schalter Und Sensoren Gmbh Verfahren zum Unterstützen eines Fahrers eines Kraftfahrzeugs beim Einparken in eine Parklücke, Fahrerassistenzeinrichtung und Kraftfahrzeug
DE102011121722A1 (de) * 2011-12-20 2013-06-20 Gm Global Technology Operations, Llc Vorrichtung zur Einparksteuerung
DE102012015968A1 (de) * 2012-08-11 2014-03-06 Audi Ag Verfahren zum fahrerlosen Bewegen eines Fahrzeugs auf einer Parkfläche
DE102014011697A1 (de) * 2014-08-07 2016-02-11 Daimler Ag Verfahren und Vorrichtung zur Durchführung eines Parkvorganges eines Fahrzeuges in einem Parkhaus
DE102015214798B4 (de) * 2015-08-04 2023-03-30 Robert Bosch Gmbh Verfahren zum Betreiben eines Fahrzeugs
DE102016211179A1 (de) * 2015-09-08 2017-03-09 Volkswagen Aktiengesellschaft Verfahren und Vorrichtung zum Durchführen einer automatisierten Fahrt eines Fahrzeugs entlang einer bereitgestellten Trajektorie

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2016033500A (ja) * 2014-07-31 2016-03-10 トヨタ自動車株式会社 出発地・目的地抽出装置、出発地・目的地抽出方法
KR20160066115A (ko) * 2014-12-01 2016-06-10 호서대학교 산학협력단 주차위치 자동 안내시스템
KR101637842B1 (ko) * 2015-07-08 2016-07-07 현대자동차주식회사 주차장 내 자율주행 시스템 및 방법
US9557741B1 (en) * 2015-08-24 2017-01-31 Ford Global Technologies, Llc System and method for autonomous valet parking using plenoptic cameras

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
HYUNDAI KIA, FROM CAR PARKING TO CAR DEPARTURE MAGICAL PARKING TECHNOLOGY FIRST RELEASED!, 25 August 2016 (2016-08-25), XP055587284, Retrieved from the Internet <URL:https://blog.hmgjournal.com/Tech/Item/KIA-AVP-Story.blg> *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3862996A1 (en) * 2020-02-07 2021-08-11 Volvo Car Corporation Automatic parking assistance system, in-vehicle device and method
US11787396B2 (en) 2020-02-07 2023-10-17 Volvo Car Corporation Automatic parking assistance system, in-vehicle device and method
CN112109700A (zh) * 2020-06-18 2020-12-22 上汽通用五菱汽车股份有限公司 记忆泊车系统及其控制方法和计算机存储介质

Also Published As

Publication number Publication date
KR20180124685A (ko) 2018-11-21
CN111148674A (zh) 2020-05-12
KR102120108B1 (ko) 2020-06-09

Similar Documents

Publication Publication Date Title
WO2018088615A1 (ko) 차량 주행 제어 장치 및 방법
WO2019066477A1 (en) AUTONOMOUS VEHICLE AND ITS CONTROL METHOD
WO2017222299A1 (en) Vehicle control device mounted on vehicle and method for controlling the vehicle
WO2019035652A1 (en) DRIVING ASSISTANCE SYSTEM AND VEHICLE COMPRISING THE SAME
WO2017138702A1 (ko) 차량용 사용자 인터페이스 장치 및 차량
WO2018056538A1 (en) Vehicle control device mounted on vehicle and method for controlling the vehicle
WO2018088647A1 (en) Vehicle control device mounted on vehicle and method for controlling the vehicle
WO2018079919A1 (ko) 자율 주행 차량 및 자율 주행 차량의 동작 방법
WO2017183797A1 (ko) 차량용 운전 보조 장치
WO2019117333A1 (ko) 차량에 구비된 디스플레이 장치 및 디스플레이 장치의 제어방법
WO2019098434A1 (ko) 차량에 구비된 차량 제어 장치 및 차량의 제어방법
EP3475134A1 (en) Vehicle control device mounted on vehicle and method for controlling the vehicle
WO2018097465A1 (en) Vehicle control device mounted on vehicle and method for controlling the vehicle
WO2018230768A1 (ko) 차량에 구비된 차량 제어 장치 및 차량의 제어방법
WO2018110789A1 (en) Vehicle controlling technology
WO2018088614A1 (ko) 차량용 사용자 인터페이스 장치 및 차량
WO2018110762A1 (ko) 차량에 구비된 차량 제어 장치 및 차량의 제어방법
EP3545380A1 (en) Vehicle control device mounted on vehicle and method for controlling the vehicle
WO2019132078A1 (ko) 차량에 구비된 디스플레이 장치
WO2022154369A1 (ko) 차량과 연동된 디스플레이 장치 및 그것의 동작방법
WO2018235979A1 (ko) 차량에 구비된 차량 제어 장치 및 차량의 제어방법
WO2019054719A1 (ko) 차량 운전 보조 장치 및 차량
WO2018093192A1 (ko) 차량용 레이더 장치 및 차량
WO2017155199A1 (ko) 차량에 구비된 차량 제어 장치 및 차량의 제어방법
WO2018097424A1 (ko) 차량용 통신 장치 및 차량

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18861545

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18861545

Country of ref document: EP

Kind code of ref document: A1