CN108446644A - A kind of virtual display system for New-energy electric vehicle - Google Patents
A kind of virtual display system for New-energy electric vehicle Download PDFInfo
- Publication number
- CN108446644A CN108446644A CN201810249691.7A CN201810249691A CN108446644A CN 108446644 A CN108446644 A CN 108446644A CN 201810249691 A CN201810249691 A CN 201810249691A CN 108446644 A CN108446644 A CN 108446644A
- Authority
- CN
- China
- Prior art keywords
- central processing
- processing unit
- module
- image
- data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000012545 processing Methods 0.000 claims abstract description 57
- 238000000034 method Methods 0.000 claims abstract description 29
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 16
- 238000000605 extraction Methods 0.000 claims abstract description 13
- 230000001815 facial effect Effects 0.000 claims abstract description 10
- 238000009499 grossing Methods 0.000 claims description 28
- 230000003287 optical effect Effects 0.000 claims description 13
- 230000033001 locomotion Effects 0.000 claims description 8
- 230000008859 change Effects 0.000 claims description 7
- 230000003993 interaction Effects 0.000 claims description 4
- 208000003464 asthenopia Diseases 0.000 claims description 3
- 238000004364 calculation method Methods 0.000 claims description 3
- 238000001514 detection method Methods 0.000 claims description 3
- 238000006073 displacement reaction Methods 0.000 claims description 3
- 230000005611 electricity Effects 0.000 claims description 3
- 230000007613 environmental effect Effects 0.000 claims description 3
- 230000006870 function Effects 0.000 claims description 3
- 238000013178 mathematical model Methods 0.000 claims description 3
- 210000003205 muscle Anatomy 0.000 claims description 3
- 230000000284 resting effect Effects 0.000 claims description 3
- 239000011800 void material Substances 0.000 claims 1
- 238000005516 engineering process Methods 0.000 description 8
- 230000008447 perception Effects 0.000 description 4
- 230000006399 behavior Effects 0.000 description 2
- 238000002485 combustion reaction Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000000052 comparative effect Effects 0.000 description 1
- 238000005094 computer simulation Methods 0.000 description 1
- 230000001276 controlling effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 230000008921 facial expression Effects 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 210000003128 head Anatomy 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000007689 inspection Methods 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003387 muscular Effects 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000001105 regulatory effect Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000014860 sensory perception of taste Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000016776 visual perception Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- G06T5/70—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/269—Analysis of motion using gradient-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/59—Context or environment of the image inside of a vehicle, e.g. relating to seat occupancy, driver state or inner lighting conditions
- G06V20/597—Recognising the driver's state or behaviour, e.g. attention or drowsiness
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
- G06V40/176—Dynamic expression
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/193—Preprocessing; Feature extraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/197—Matching; Classification
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B21/00—Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
- G08B21/02—Alarms for ensuring the safety of persons
- G08B21/06—Alarms for ensuring the safety of persons indicating a condition of sleep, e.g. anti-dozing alarms
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
- G06T2207/30201—Face
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30248—Vehicle exterior or interior
- G06T2207/30252—Vehicle exterior; Vicinity of vehicle
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30248—Vehicle exterior or interior
- G06T2207/30268—Vehicle interior
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2210/00—Indexing scheme for image generation or computer graphics
- G06T2210/21—Collision detection, intersection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2210/00—Indexing scheme for image generation or computer graphics
- G06T2210/61—Scene description
Abstract
A kind of virtual display system for New-energy electric vehicle is claimed in the present invention, and as the ancillary equipment of entity display, the virtual display system includes:Image capture module, data acquisition module, the second central processing unit, virtual reality experience unit and alarm module, second central processing unit is connected with image capture module, data acquisition module respectively, and second central processing unit is connected with virtual reality experience unit;Wherein described second central processing unit includes image processing module, risk of collision computing unit and memory, described image acquisition module is used to obtain the information including vehicle ambient enviroment image, motorist's facial information, copilot personnel face, and using feature extraction algorithm extraction driver eye, the characteristic information of copilot eye, it is transferred to the second central processing unit.This method processing data capability is strong, management is quick, decision accurately facilitates.
Description
Technical field
The invention belongs to electric vehicle engineering field more particularly to a kind of virtual display systems for New-energy electric vehicle
System.
Background technology
Existing electric vehicle refers to driving wheels travel with motor using vehicle power supply as power, meets road traffic, peace
The vehicle of full regulation requirements.It is started using the electricity stored in the battery.Use 12 or 24 sometimes when driving automobile
Block battery then needs more sometimes.The composition of electric vehicle includes:The machineries such as electric drive and control system, driving force transmission
System, the equipment etc. for completing assigned tasks.Electric drive and control system are the core of electric vehicle, and are different from interior
The maximum difference of combustion engine automobile.Electric drive and control system are filled by the speed regulating control of drive motor, power supply and motor
The compositions such as set.Other devices of electric vehicle are substantially identical as internal-combustion engines vehicle.
Electric vehicle design does not need machine driven system, and does not need the direction of ABS brake system and mechanical linkage
Disc system.This kind of electric vehicle is commonly equipped with system below:Computer control system, with small, reliability is high, the service life
The advantages of long and low cost;Integrated wheel system, possesses the steering of power and computer long-distance control.Because having used electricity
Sub- technology can arbitrarily change the position of steering wheel or so driving, to be suitble to the needs of country variant, driver can also be allowed direction
Disk moves, reverse to drive towards rear.Virtual reality technology is a kind of Computer Simulation that can be created with the experiencing virtual world
System, it generates a kind of simulated environment using computer, be a kind of Multi-source Information Fusion, interactive Three-Dimensional Dynamic what comes into a driver's and
The system emulation of entity behavior makes user be immersed in the environment, simulated environment be generated by computer, in real time dynamic three
The three-dimensional photorealism of dimension.Perception, which refers to ideal VR, should have perception possessed by all people.Except computer graphics techniques institute
Outside the visual perception of generation, also the sense of hearing, tactile, power feel, the perception such as movement, or even further include smell and sense of taste etc., also referred to as
More perception.Natural technical ability refers to the head rotation of people, eyes, gesture or other human body behavior acts, by computer come handle with
The data that the action of participant is adapted, and real-time response is made to the input of user, and the face of user are fed back to respectively.It passes
It refers to three-dimension interaction equipment to feel equipment.But the technology is not used in existing vehicle, and not corresponding document report,
And the system that the safety of corresponding vehicle carries out virtual controlling.
Invention content
Present invention seek to address that the above problem of the prior art.Propose it is a kind of it is intelligent, improve safety coefficient and be used for
The virtual display system of New-energy electric vehicle.
Technical scheme is as follows:
A kind of virtual display system for New-energy electric vehicle, as the ancillary equipment of entity display, the reality
Body display includes wireless data interface, wired data interface, the first central processing unit and display screen, the virtual display system
Including:Image capture module, data acquisition module, the second central processing unit, virtual reality experience unit and alarm module, it is described
Second central processing unit is connected with image capture module, data acquisition module respectively, second central processing unit with it is virtual
Experience of reality unit is connected;Wherein described second central processing unit include image processing module, risk of collision computing unit and
Memory, described image acquisition module include vehicle ambient enviroment image, motorist's facial information, copilot personnel's face for obtaining
Information including portion, and using feature extraction algorithm extraction driver eye, the characteristic information of copilot eye, it is transferred to second
Central processing unit;Data acquisition module, for obtain include the battery usage amount data of electric vehicle, engine temperature data,
Data including battery temperature are transferred to the second central processing unit;Virtual reality experience unit, using Web3D technologies, for carrying
For a simple virtual reality scenario, there is the function including user's displacement, the interaction of ten finger keyboards;It alarms for receiving
The alarm command that module is sent;The data that described image processing module is used to obtain image capture module are stored with memory module
Tired eye feature information carry out matching comparison, if matching result be driver's fatigue driving if immediately start alarm module into
Row alarm, and notify copilot personnel;Transmit information to driver if copilot is matched as resting state, it is automatic or
Manually adjust the seat of copilot personnel;It is additionally operable to the battery usage amount data of electric vehicle, engine temperature data, battery
Data including temperature, and it is transferred to the second central processing unit;These data are carried out at prediction using Secondary Exponential Smoothing Method
Reason, the calculation formula that Secondary Exponential Smoothing Method obtains double smoothing on the basis of single exponential smoothing are:
In formula:St (2)--- the double smoothing value in t periods;St (1)--- the single exponential smoothing value in t periods;
St-1 (2)--- the double smoothing value in t-1 periods;A --- weighting coefficient is also referred to as smoothing factor;Establish the mathematics of prediction
Then model determines predicted value with mathematical model, alarm module is notified if prediction result is beyond the threshold value of setting, and will report
Alert result is transferred to driver;The vehicle environmental view that the risk of collision judging unit is used to that image processing module to be combined to obtain
Picture, motorist's facial information, judge risk of collision, alarm module are sent to if being judged as danger, and send out by alarm module
It gives wheel tire cohesion device to be braked, and is virtually shown.
Further, the data acquisition module includes camera, Velocity-acceleration sensor, ultrasonic radar sensing
Device, GPS positioning sensor, for obtaining electric vehicle itself and ambient condition information.
Further, described image acquisition module uses feature extraction algorithm using the optical flow method based on dynamic image
Feature extracting method specifically includes:
Optical flow method is to reflect the method for respective objects grey scale change between different frame in dynamic image, uses successive frame first
Between optical flow field and gradient fields, indicate the change in time and space of image respectively, realize the expression area tracking per frame facial image;So
Afterwards by the variation of the characteristic area direction of motion, the movement of face muscle, and then corresponding different expression are indicated;Combining people
After face biological information, the detection to human face region is realized using improved Ratio Template algorithms;Then light is used
Stream method calculates the gray level model by all kinds of means of face, completes the tracking to human face region;Finally expression point is realized with SVM algorithm
Class.
Further, spare each other between first central processing unit and the second central processing unit, second center
Processor is set to around the first central processing unit.
It advantages of the present invention and has the beneficial effect that:
Spare each other between first central processing unit and the second central processing unit of the invention, second central processing unit is set
It is placed in around the first central processing unit, improves the safety coefficient of equipment in this way, reduces failure rate, even if breaking down
When can not also be disturbed.In addition, image capture module of the present invention uses feature extraction algorithm using based on dynamic image
Optical flow method feature extracting method, optical flow method dynamic image reflects the process of human face expression generation, therefore the table of dynamic image
Feelings feature is mainly manifested on the continuing deformation of face and the muscular movement of facial different zones, the driver's dynamic extracted in this way
Image is just more more acurrate than using static nature extraction algorithm, after combining face biological information, uses improved Ratio
Template (scale model) algorithm realizes that the detection to human face region, improvement are:Comparative example is used and is carried out from high to low
Ratio is configured similar in the driver of arrangement, selection percentage and realization setting;To completing classification and knowledge to face characteristic
Not;Prediction processing is carried out to these data using Secondary Exponential Smoothing Method, rather than uses Single Exponential Smoothing, is because of fortune
With Secondary Exponential Smoothing Method to the smooth again of single exponential smoothing.The original that it is suitable for having the time series of linear trend
Reason, prediction are more acurrate.
Description of the drawings
Fig. 1 is the structural schematic diagram that the present invention provides that preferred embodiment states virtual display system.
Fig. 2 is the structure diagram that the present invention provides the second central processing unit of preferred embodiment.
Specific implementation mode
Following will be combined with the drawings in the embodiments of the present invention, and technical solution in the embodiment of the present invention carries out clear, detailed
Carefully describe.Described embodiment is only a part of the embodiment of the present invention.
The present invention solve above-mentioned technical problem technical solution be:
It is as shown in Figs. 1-2 a kind of virtual display system for New-energy electric vehicle, as the auxiliary of entity display
It includes wireless data interface, wired data interface, the first central processing unit and display screen, institute to help equipment, the entity display
Stating virtual display system includes:Image capture module, data acquisition module, the second central processing unit, virtual reality experience unit
And alarm module, second central processing unit is connected with image capture module, data acquisition module respectively, in described second
Central processor is connected with virtual reality experience unit;Wherein described second central processing unit includes image processing module, collision
Risk Calculation unit and memory, described image acquisition module include vehicle ambient enviroment image, motorist face letter for obtaining
Information including breath, copilot personnel face, and using the feature of feature extraction algorithm extraction driver eye, copilot eye
Information is transferred to the second central processing unit;Data acquisition module, for obtain include electric vehicle battery usage amount data,
Data including engine temperature data, battery temperature are transferred to the second central processing unit;Virtual reality experience unit uses
Web3D technologies have for providing a simple virtual reality scenario including user's displacement, the interaction of ten finger keyboards
Function;The alarm command sent for receiving alarm module;What described image processing module was used to obtain image capture module
Data match comparing with the tired eye feature information that memory module stores, if matching result is driver's fatigue driving
Start alarm module immediately to alarm, and notifies copilot personnel;Information is passed if copilot is matched as resting state
It is defeated by driver, either automatically or manually adjusts the seat of copilot personnel;Be additionally operable to by the battery usage amount data of electric vehicle,
Data including engine temperature data, battery temperature, and it is transferred to the second central processing unit;Using Secondary Exponential Smoothing Method pair
These data carry out prediction processing, and Secondary Exponential Smoothing Method obtains the calculating of double smoothing on the basis of single exponential smoothing
Formula is:
In formula:St (2)--- the double smoothing value in t periods;St (1)--- the single exponential smoothing value in t periods;
St-1 (2)--- the double smoothing value in t-1 periods;A --- weighting coefficient is also referred to as smoothing factor;Establish the mathematics of prediction
Then model determines predicted value with mathematical model, alarm module is notified if prediction result is beyond the threshold value of setting, and will report
Alert result is transferred to driver;The vehicle environmental view that the risk of collision judging unit is used to that image processing module to be combined to obtain
Picture, motorist's facial information, judge risk of collision, alarm module are sent to if being judged as danger, and send out by alarm module
It gives wheel tire cohesion device to be braked, and is virtually shown.
Preferably, the data acquisition module include camera, Velocity-acceleration sensor, ultrasonic radar sensor,
GPS positioning sensor, for obtaining electric vehicle itself and ambient condition information.
Preferably, described image acquisition module is special using the optical flow method based on dynamic image using feature extraction algorithm
Extracting method is levied, is specifically included:Optical flow method is to reflect the method for respective objects grey scale change between different frame in dynamic image, first
The optical flow field and gradient fields between successive frame are first used, indicates the change in time and space of image respectively, realizes the table per frame facial image
Feelings area tracking;Then by the variation of the characteristic area direction of motion, the movement of face muscle, and then corresponding different table are indicated
Feelings;After combining face biological information, the inspection to human face region is realized using improved Ratio Template algorithms
It surveys;Then it uses optical flow method to calculate the gray level model by all kinds of means of face, completes the tracking to human face region;Finally use SVM algorithm
Realize expression classification.Preferably, spare each other between first central processing unit and the second central processing unit, described second
Central processing unit is set to around the first central processing unit.
The preferred embodiment of the present invention has been described in detail above.It should be appreciated that those skilled in the art without
It needs creative work according to the present invention can conceive and makes many modifications and variations.Therefore, all technologies in the art
Personnel are available by logical analysis, reasoning, or a limited experiment on the basis of existing technology under this invention's idea
Technical solution, all should be in the protection domain being defined in the patent claims.
Claims (4)
1. a kind of virtual display system for New-energy electric vehicle, as the ancillary equipment of entity display, the entity
Display includes wireless data interface, wired data interface, the first central processing unit and display screen, which is characterized in that the void
Quasi- display system includes:Image capture module, data acquisition module, the second central processing unit, virtual reality experience unit and report
Alert module, second central processing unit are connected with image capture module, data acquisition module respectively, second centre
Reason device is connected with virtual reality experience unit;Wherein described second central processing unit includes image processing module, risk of collision
Computing unit and memory, described image acquisition module include vehicle ambient enviroment image, motorist's facial information, pair for obtaining
Information including driver face, and the characteristic information of feature extraction algorithm extraction driver eye, copilot eye is used,
It is transferred to the second central processing unit;Data acquisition module, for obtain include electric vehicle battery usage amount data, engine
Data including temperature data, battery temperature are transferred to the second central processing unit;Virtual reality experience unit, using Web3D skills
Art has the function including user's displacement, the interaction of ten finger keyboards for providing a simple virtual reality scenario;With
In the alarm command that receiving alarm module is sent;The data that described image processing module is used to obtain image capture module with deposit
The tired eye feature information of storage module storage carries out matching comparison, starts immediately if matching result is driver's fatigue driving
Alarm module is alarmed, and notifies copilot personnel;Driving is transmitted information to if copilot is matched as resting state
Person either automatically or manually adjusts the seat of copilot personnel;It is additionally operable to battery usage amount data, the engine temperature of electric vehicle
Data including degrees of data, battery temperature, and it is transferred to the second central processing unit;Using Secondary Exponential Smoothing Method to these data
Prediction processing is carried out, the calculation formula that Secondary Exponential Smoothing Method obtains double smoothing on the basis of single exponential smoothing is:
In formula:St (2)--- the double smoothing value in t periods;St (1)--- the single exponential smoothing value in t periods;
St-1 (2)--- the double smoothing value in t-1 periods;A --- weighting coefficient is also referred to as smoothing factor;Establish the mathematics of prediction
Then model determines predicted value with mathematical model, alarm module is notified if prediction result is beyond the threshold value of setting, and will report
Alert result is transferred to driver;The vehicle environmental view that the risk of collision judging unit is used to that image processing module to be combined to obtain
Picture, motorist's facial information, judge risk of collision, alarm module are sent to if being judged as danger, and send out by alarm module
It gives wheel tire cohesion device to be braked, and is virtually shown.
2. the virtual display system according to claim 1 for New-energy electric vehicle, which is characterized in that the data
Acquisition module includes camera, Velocity-acceleration sensor, ultrasonic radar sensor, GPS positioning sensor, for obtaining electricity
Electrical automobile itself and ambient condition information.
3. the virtual display system according to claim 1 for New-energy electric vehicle, which is characterized in that described image
Acquisition module, using the optical flow method feature extracting method based on dynamic image, is specifically included using feature extraction algorithm:
Optical flow method is to reflect the method for respective objects grey scale change between different frame in dynamic image, first using between successive frame
Optical flow field and gradient fields, indicate the change in time and space of image respectively, realize the expression area tracking per frame facial image;Then lead to
The variation for crossing the characteristic area direction of motion indicates the movement of face muscle, and then corresponding different expression;Combining face life
After object information, the detection to human face region is realized using improved Ratio Template algorithms;Then optical flow method is used
The gray level model by all kinds of means of face is calculated, the tracking to human face region is completed;Finally expression classification is realized with SVM algorithm.
4. the virtual display system according to claim 1 for New-energy electric vehicle, which is characterized in that described first
Spare each other between central processing unit and the second central processing unit, second central processing unit is set to the first central processing unit
Around.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810249691.7A CN108446644A (en) | 2018-03-26 | 2018-03-26 | A kind of virtual display system for New-energy electric vehicle |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810249691.7A CN108446644A (en) | 2018-03-26 | 2018-03-26 | A kind of virtual display system for New-energy electric vehicle |
Publications (1)
Publication Number | Publication Date |
---|---|
CN108446644A true CN108446644A (en) | 2018-08-24 |
Family
ID=63196464
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810249691.7A Pending CN108446644A (en) | 2018-03-26 | 2018-03-26 | A kind of virtual display system for New-energy electric vehicle |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108446644A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110641478A (en) * | 2019-10-15 | 2020-01-03 | 深圳市英博超算科技有限公司 | Automobile domain controller display method and device, automobile and readable storage medium |
CN116901975A (en) * | 2023-09-12 | 2023-10-20 | 深圳市九洲卓能电气有限公司 | Vehicle-mounted AI security monitoring system and method thereof |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1801730A1 (en) * | 2005-12-23 | 2007-06-27 | Delphi Technologies, Inc. | Method of detecting vehicle-operator state |
EP1867281A1 (en) * | 2006-06-13 | 2007-12-19 | Delphi Technologies, Inc. | Improved dynamic eye tracking system |
CN101299296A (en) * | 2007-05-02 | 2008-11-05 | 韩晓刚 | Virtual traffic display system |
CN101578547A (en) * | 2006-11-20 | 2009-11-11 | 约翰逊控制技术公司 | Apparatus and system for providing a virtual display for a vehicle |
CN101941425A (en) * | 2010-09-17 | 2011-01-12 | 上海交通大学 | Intelligent recognition device and method for fatigue state of driver |
CN103479367A (en) * | 2013-09-09 | 2014-01-01 | 广东工业大学 | Driver fatigue detection method based on facial action unit recognition |
CN104881956A (en) * | 2015-06-17 | 2015-09-02 | 上海大学 | Fatigue driving early warning system |
CN106291924A (en) * | 2015-05-28 | 2017-01-04 | 拓蒂电子(上海)有限公司 | A kind of vehicle-mounted head-up display system based on DLP technology |
CN106530623A (en) * | 2016-12-30 | 2017-03-22 | 南京理工大学 | Fatigue driving detection device and method |
-
2018
- 2018-03-26 CN CN201810249691.7A patent/CN108446644A/en active Pending
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1801730A1 (en) * | 2005-12-23 | 2007-06-27 | Delphi Technologies, Inc. | Method of detecting vehicle-operator state |
EP1867281A1 (en) * | 2006-06-13 | 2007-12-19 | Delphi Technologies, Inc. | Improved dynamic eye tracking system |
CN101578547A (en) * | 2006-11-20 | 2009-11-11 | 约翰逊控制技术公司 | Apparatus and system for providing a virtual display for a vehicle |
CN101299296A (en) * | 2007-05-02 | 2008-11-05 | 韩晓刚 | Virtual traffic display system |
CN101941425A (en) * | 2010-09-17 | 2011-01-12 | 上海交通大学 | Intelligent recognition device and method for fatigue state of driver |
CN103479367A (en) * | 2013-09-09 | 2014-01-01 | 广东工业大学 | Driver fatigue detection method based on facial action unit recognition |
CN106291924A (en) * | 2015-05-28 | 2017-01-04 | 拓蒂电子(上海)有限公司 | A kind of vehicle-mounted head-up display system based on DLP technology |
CN104881956A (en) * | 2015-06-17 | 2015-09-02 | 上海大学 | Fatigue driving early warning system |
CN106530623A (en) * | 2016-12-30 | 2017-03-22 | 南京理工大学 | Fatigue driving detection device and method |
Non-Patent Citations (4)
Title |
---|
徐亮: "航空发动机气路性能参数趋势预测研究", 《中国优秀硕士学位论文全文数据库 工程科技Ⅱ辑》 * |
时召会: "《中国人工智能进展》", 31 December 2007 * |
蒋斌: "人脸表情识别的研究进展", 《计算机科学》 * |
韩志艳: "《面向语音与面部表情信号的多模式情感识别技术研究》", 31 January 2017 * |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110641478A (en) * | 2019-10-15 | 2020-01-03 | 深圳市英博超算科技有限公司 | Automobile domain controller display method and device, automobile and readable storage medium |
CN110641478B (en) * | 2019-10-15 | 2021-02-19 | 英博超算(南京)科技有限公司 | Automobile domain controller display method and device, automobile and readable storage medium |
CN116901975A (en) * | 2023-09-12 | 2023-10-20 | 深圳市九洲卓能电气有限公司 | Vehicle-mounted AI security monitoring system and method thereof |
CN116901975B (en) * | 2023-09-12 | 2023-11-21 | 深圳市九洲卓能电气有限公司 | Vehicle-mounted AI security monitoring system and method thereof |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11688203B2 (en) | Systems and methods for providing visual allocation management | |
Morrell et al. | Design and evaluation of a vibrotactile seat to improve spatial awareness while driving | |
US20220203996A1 (en) | Systems and methods to limit operating a mobile phone while driving | |
CN106652643A (en) | Driving training simulation system based on virtual reality technology | |
CN110291479A (en) | Rollback operation is executed based on instantaneous autonomous operation situation | |
Li et al. | Multi-sensor soft-computing system for driver drowsiness detection | |
CN111915159A (en) | Personalized takeover early warning method and system based on dynamic time budget | |
Wang et al. | Automobile driver posture monitoring systems: A review | |
WO2022110737A1 (en) | Vehicle anticollision early-warning method and apparatus, vehicle-mounted terminal device, and storage medium | |
CN108446644A (en) | A kind of virtual display system for New-energy electric vehicle | |
CN109017444A (en) | Electrombile control method, device, electric vehicle control system and electric vehicle | |
US20220001894A1 (en) | System for controlling autonomous vehicle for reducing motion sickness | |
CN114030475A (en) | Vehicle driving assisting method and device, vehicle and storage medium | |
Rong et al. | Artificial intelligence methods in in-cabin use cases: a survey | |
CN108583569A (en) | A kind of collision warning device based on double moving average algorithm | |
Qiu et al. | Incorporating gaze behavior using joint embedding with scene context for driver takeover detection | |
Baltodano et al. | Eliciting driver stress using naturalistic driving scenarios on real roads | |
Louie et al. | Towards a driver monitoring system for estimating driver situational awareness | |
US20230347903A1 (en) | Sensor-based in-vehicle dynamic driver gaze tracking | |
Lashkov et al. | A multimodal approach to psycho-emotional state detection of a vehicle driver | |
CN115320626B (en) | Danger perception capability prediction method and device based on human-vehicle state and electronic equipment | |
Wang et al. | Fatigue driving detection of urban road at night based on multimodal information fusion. | |
Gomes et al. | Continuous-time feedback device to enhance situation awareness during take-over requests in automated driving conditions | |
Liu et al. | Deep Learning Based Take-Over Performance Prediction and Its Application on Intelligent Vehicles | |
Alomari | Human-centric detection and mitigation approach for various levels of cell phone-based driver distractions |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20180824 |