TWI619099B - Intelligent multifunctional driving assisted driving recording method and system - Google Patents
Intelligent multifunctional driving assisted driving recording method and system Download PDFInfo
- Publication number
- TWI619099B TWI619099B TW105134417A TW105134417A TWI619099B TW I619099 B TWI619099 B TW I619099B TW 105134417 A TW105134417 A TW 105134417A TW 105134417 A TW105134417 A TW 105134417A TW I619099 B TWI619099 B TW I619099B
- Authority
- TW
- Taiwan
- Prior art keywords
- image
- driving
- feature
- vehicle
- road condition
- Prior art date
Links
Landscapes
- Image Analysis (AREA)
- Traffic Control Systems (AREA)
Abstract
本發明係揭露一種智慧多功能行車輔助駕駛記錄方法及系統,其係以影像擷取裝置擷取車輛行駛中的行車路況影像。並於物件特徵資料庫建立有物件特徵資料,再於每一物件特徵資料設定物件名稱。資訊處理單元於物件特徵資料庫辨識出與行車路況影像的其中至少一個物件之影像符合的物件名稱,再將已辨識出物件名稱的物件於顯示幕所顯示的行車路況影像中予以標記,俾能以標記作為行車預警資訊,藉以提升車輛於行車時的安全性。 The invention discloses a smart multi-functional driving assisted driving recording method and system, which uses an image capturing device to capture an image of driving road conditions during running of a vehicle. And the object feature data is established in the object feature database, and the object name is set in each object feature data. The information processing unit identifies the object name corresponding to the image of at least one of the objects of the driving road condition image in the object feature database, and then marks the object whose name has been identified in the driving road condition image displayed on the display screen. Marking as driving warning information to improve the safety of vehicles while driving.
Description
本發明係有關一種智慧多功能行車輔助駕駛記錄方法及系統,尤指一種可以藉由行車預警而提升行車安全的影像辨識技術。 The invention relates to a smart multi-functional driving assisted driving recording method and system, in particular to an image recognition technology capable of improving driving safety by driving warning.
按,為使車輛具備交通事故的舉證功效,一般車主大多會於車輛內裝設行車記錄器,於行車時,除了可以擷取車輛前方的行車路況影像之外,並可記錄對應的時間資訊,以作為時光回溯的證據比對依據。雖然習知行車記錄器設置具備擷取前方行車路況影像及時間資訊等功能;惟,其無法藉由檢測得知是否有穿越快車道的行人;或是移動中之前方車輛及對向來車的動態資訊,以致無法判定車輛與上述物件之間的相對運動關係,若是不幸地發生碰撞交通事故時,習知行車記錄器所記錄的行車路況影像確實較無法提供足夠的舉證資訊來釐清交通事故的肇事原因,因而造成行車記錄使用上的不便與極大的困擾。 In order to make the vehicle have the evidence-proven effect of traffic accidents, most of the owners will install a driving recorder in the vehicle. In addition to the driving road image in front of the vehicle, the corresponding time information can be recorded. Used as a basis for evidence of time backtracking. Although the conventional driving recorder is equipped with functions such as capturing traffic image and time information in front of the road; however, it cannot detect whether there is a pedestrian crossing the fast lane; or the movement of the preceding vehicle and the oncoming vehicle. Information, so that the relative motion relationship between the vehicle and the above objects cannot be determined. If a traffic accident occurs unfortunately, the traffic image recorded by the conventional driving recorder is actually less able to provide sufficient evidence to clarify the accident of the traffic accident. The reason, thus causing inconvenience and great troubles in the use of driving records.
為改善上述缺失,相關技術領域業者已然一種如中華民國新型第M414577號『具影像辨識功能之行車導航記錄器』所示的專利;其係依據車輛之經緯度資訊取得車輛之車速信號並傳予處理單元;影像擷取單元係取得車輛周圍道路影像並傳予處理單元;影像辨識單元分析道路影像以產生分析結果。儲存單元係設置於行車導航記錄器主機,係儲存該車速信號、道路影像及分析結果;藉由影像擷取單元擷取道路影像並傳輸至處 理單元及影像辨識單元以處理影像辨識工作,而處理單元進一步連接衛星定位單元以取得當時之車速信號,以供處理單元辨識並判斷是否為危險駕駛行為。 In order to improve the above-mentioned deficiencies, the related art field has already obtained a patent as shown in the new Republic of China No. M414577 "Driving Navigation Recorder with Image Recognition Function"; it obtains the vehicle speed signal according to the latitude and longitude information of the vehicle and transmits it to the processing. The image capturing unit acquires the road image around the vehicle and transmits it to the processing unit; the image recognition unit analyzes the road image to generate an analysis result. The storage unit is installed in the driving navigation recorder host to store the vehicle speed signal, road image and analysis result; the image capturing unit captures the road image and transmits it everywhere The processing unit and the image recognition unit process the image recognition work, and the processing unit further connects the satellite positioning unit to obtain the current vehicle speed signal for the processing unit to recognize and determine whether it is a dangerous driving behavior.
該專利影像辨識單元雖然可以分析道路影像以產生分析結果;惟,並未對『分析結果』的具體技術內容以及是否作為『危險駕駛行為』的判定因子之一提出任何的具體論述;除此之外,該專利並無深度學習的機能建制,以致無法透過深度學習來強化影像比對樣本的分類功能,故而影像影像比對辨識的精確度較低而無法被有效應用。 Although the patent image recognition unit can analyze the road image to produce the analysis result; however, it does not provide any specific discussion on the specific technical content of the "analysis result" and whether it is one of the determining factors of "dangerous driving behavior"; In addition, the patent does not have the function of deep learning, so that it is impossible to enhance the classification function of the image comparison sample through deep learning. Therefore, the image image comparison accuracy is low and cannot be effectively applied.
另有一種如中華民國發明第I495343號『行車記錄系統及其位置判斷方法』所示的專利;其係包含曲率影像鏡頭、運算模組、處理模組顯示模組及儲存模組。係曲率影像鏡頭擷取其周圍環境之曲率影像;運算模組係將曲率影像還原為還原影像。座標定位模組係接收還原影像,並以座標定位還原影像,以產生經座標定位之還原影像;處理模組係接收還原影像,並產生時間資訊;顯示模組係顯示還原影像及時間資訊。儲存模組係儲存還原影像及時間資訊。 There is also a patent as shown in the Republic of China Invention No. I495343 "Driving Recording System and Position Judgment Method"; it includes a curvature image lens, a computing module, a processing module display module, and a storage module. The curvature image captures the curvature image of the surrounding environment; the computation module restores the curvature image to the restored image. The coordinate positioning module receives the restored image, and restores the image by coordinates to generate a restored image by coordinate positioning; the processing module receives the restored image and generates time information; and the display module displays the restored image and time information. The storage module stores the restored image and time information.
該專利雖然可由還原影像計算出車輛與其中至少一個特定物件之間的速度及相對距離的相對運動關係,以此作為發出警示訊號的依據;惟,其並無深度學習的機能建制,以致無法透過深度學習來強化影像比對樣本的分類功能,因而影像影像比對辨識的精確度;除此之外,該專利僅是對移動中之前方車輛或是對向來車進行辨識計算,以得到速度與距離等資訊;至於行人方面則是無法有效監視,所以無法藉由檢測得知是否有欲穿越快車道的行人出現,以致無法有效降低開車撞擊行人的情事;或 是移動中之前方車輛及對向來車的速度與距離等資訊,因而造成行車記錄使用上的不便與即極大困擾。 Although the patent can calculate the relative motion relationship between the speed and the relative distance between the vehicle and at least one specific object from the restored image, as the basis for issuing the warning signal; however, it has no deep learning function, so that it cannot pass through. Deep learning to enhance the classification function of the image comparison sample, so the accuracy of the image image comparison identification; in addition, the patent only identifies the previous vehicle or the oncoming vehicle in motion to obtain the speed and Information such as distance; as far as pedestrians are concerned, it is impossible to monitor effectively, so it is impossible to detect whether there is a pedestrian who wants to cross the fast lane, so that it is impossible to effectively reduce the impact of driving pedestrians; or It is the information such as the speed and distance of the preceding vehicle and the oncoming vehicle during the movement, which causes great inconvenience in the use of the driving record.
有鑑於此,尚未有一種具備較佳深度學習演算的物件影像識別以強化資料庫分類功能之行車記錄技術的專利被提出,而且基於相關產業的迫切需求之下,本發明人等乃經不斷的努力研發之下,終於研發出一套有別於上述習知技術與專利的本發明。 In view of this, there is no patent for driving recording technology with object recognition for better deep learning calculus to enhance the classification function of the database, and the inventors have been continually based on the urgent needs of related industries. Under the circumstance of research and development, a set of inventions different from the above-mentioned prior art and patents has finally been developed.
本發明第一目的,在於提供一種智慧多功能行車輔助駕駛記錄方法及系統,主要是可對車道附近的行人或是移動中之前方車輛及對向來車的動向進行監控,藉此依據監控結果而做出相應的行車預警,而可藉由行車預警來提升行車的安全性。達成本發明第一目的採用之技術手段,係以影像擷取裝置擷取車輛行駛中的行車路況影像。於物件特徵資料庫建立有物件特徵資料,並於每一物件特徵資料設定物件名稱。資訊處理單元於物件特徵資料庫辨識出與行車路況影像的其中至少一個物件之影像符合的物件名稱,再將已辨識出物件名稱的物件於顯示幕所顯示的行車路況影像中予以標記,俾能以標記作為行車預警資訊。 A first object of the present invention is to provide a smart multi-functional driving assistance driving recording method and system, which can mainly monitor pedestrians in a vicinity of a lane or a moving vehicle in front of a moving vehicle and an oncoming vehicle, thereby monitoring results according to the monitoring result. Corresponding driving warnings can be made, and driving safety can be improved by driving warning. The technical means adopted for achieving the first object of the present invention is to capture an image of the driving road condition during running of the vehicle by using the image capturing device. The object feature data is created in the object feature database, and the object name is set in each object feature data. The information processing unit identifies the object name corresponding to the image of at least one of the objects of the driving road condition image in the object feature database, and then marks the object whose name has been identified in the driving road condition image displayed on the display screen. Mark as a driving warning information.
本發明第二目的,在於提供一種具備深度學習功能的智慧多功能行車輔助駕駛記錄方法及系統,主要是藉由深度學習之機能建制,除了可以強化特徵樣本資料庫分類功能之外,並可將前段的特徵擷取部分利用深度學習方式來實現自我學習修正,因而可以達到高度的辨識率,並可避免傳統影像辨識會因前段特徵擷取容易受到環境變化而失去辨識準確率的缺失產生。達成本發明第二目的採用之技術手段,係以影像擷取裝置擷取車輛行駛中的行車路況影像。於物件特徵資料庫建立有物件特徵資料,並於每一物件特徵資料設定物件名稱。資訊處理單元於物件特徵資料庫辨 識出與行車路況影像的其中至少一個物件之影像符合的物件名稱,再將已辨識出物件名稱的物件於顯示幕所顯示的行車路況影像中予以標記,俾能以標記作為行車預警資訊。其中,其更包含一具備深度學習訓練功能以執行該影像辨識步驟的深度學習演算法,執行該深度學習演算法則包含下列之步驟:一訓練階段步驟,係輸入距離資料及巨量的該行車路況影像以建立一深度學習模型,並由該深度學習模型測試影像辨識的正確率,再判斷影像辨識正確率是否足夠,當判斷結果為是,則將辨識結果輸出及儲存;當判斷結果為否,則使該深度學習模型自我修正學習;及一運行預測階段步驟,係於該深度學習模型輸入該距離資料及即時擷取的該行車路況影像,並由該深度學習模型進行預測性影像辨識,以得到至少一個辨識結果的該物件名稱及距離值,再將該物件名稱及該距離值輸出。 A second object of the present invention is to provide a smart multi-functional driving assisted driving recording method and system with deep learning function, which is mainly constructed by deep learning, in addition to strengthening the feature sample database classification function, and The feature extraction part of the previous paragraph uses the deep learning method to realize the self-learning correction, so that the high recognition rate can be achieved, and the traditional image recognition can avoid the loss of the recognition accuracy due to the environmental change of the previous segment. The technical means adopted for achieving the second object of the present invention is to use an image capturing device to capture an image of the driving road condition during running of the vehicle. The object feature data is created in the object feature database, and the object name is set in each object feature data. Information processing unit is identified in the object feature database The object name that matches the image of at least one of the objects of the road traffic image is recognized, and the object whose name has been identified is marked in the driving road condition image displayed on the display screen, and the marking can be used as the driving warning information. The method further includes a deep learning algorithm with a deep learning training function to perform the image recognition step, and the performing the deep learning algorithm includes the following steps: a training phase step, inputting distance data and a huge amount of the driving road condition The image is used to establish a deep learning model, and the depth learning model tests the correct rate of image recognition, and then determines whether the image recognition correct rate is sufficient. When the judgment result is yes, the identification result is output and stored; when the judgment result is no, And the depth learning model self-correction learning; and a running prediction phase step, the depth learning model inputs the distance data and the immediately captured driving road image, and the deep learning model performs predictive image recognition, Obtaining at least one object name and distance value of the identification result, and outputting the object name and the distance value.
10‧‧‧影像擷取裝置 10‧‧‧Image capture device
20‧‧‧物件特徵資料庫 20‧‧‧ Object characterization database
21‧‧‧車輛特徵資料 21‧‧‧ Vehicle characteristics data
22‧‧‧機車/人特徵資料 22‧‧‧Motorcycle/person profile data
23‧‧‧行人特徵資料 23‧‧‧Pedestrian characteristics
24‧‧‧交通號誌特徵資料 24‧‧‧ Traffic Signature Characteristics
25‧‧‧座標參數資料庫 25‧‧‧ Coordinate Parameter Database
30‧‧‧資訊處理單元 30‧‧‧Information Processing Unit
31‧‧‧深度學習演算模組 31‧‧‧Deep learning calculus module
310‧‧‧深度學習模型 310‧‧‧Deep learning model
32‧‧‧危險警告評估模組 32‧‧‧Danger Warning Evaluation Module
33‧‧‧行車危險等級預警模組 33‧‧‧ Driving hazard warning module
50‧‧‧機殼 50‧‧‧Chassis
40‧‧‧顯示幕 40‧‧‧ display screen
41‧‧‧標記 41‧‧‧ mark
51‧‧‧記憶裝置 51‧‧‧ memory device
52‧‧‧音頻播放模組 52‧‧‧Audio playback module
ECU‧‧‧行車電腦 ECU‧‧‧Driving computer
GPS‧‧‧全球衛星定位模組 GPS‧‧‧Global Satellite Positioning Module
Ob‧‧‧物件影像 Ob‧‧‧ Object image
圖1係本發明具體架構的功能方塊示意圖。 1 is a functional block diagram of a specific architecture of the present invention.
圖2係本發明深度學習模型的訓練階段的實施示意圖。 2 is a schematic diagram of the implementation of the training phase of the deep learning model of the present invention.
圖3係本發明深度學習模型的運行預測階段的實施示意圖。 3 is a schematic diagram of the implementation of the operational prediction phase of the deep learning model of the present invention.
圖4係本發明於畫面物件框選為行車預警的實施示意圖。 FIG. 4 is a schematic diagram of the implementation of the screen object selection in the present invention.
圖5係本發明物件影像套入已知座標定位的實施示意圖。 FIG. 5 is a schematic view showing the implementation of positioning the image of the object of the present invention into a known coordinate.
圖6係本發明物件影像套入已知座標定位的另一實施示意圖。 Fig. 6 is a schematic view showing another embodiment of the image of the object of the present invention nested in a known coordinate.
圖7係本發明物件影像的路徑預測實施示意圖。 Fig. 7 is a schematic diagram showing the path prediction implementation of the image of the object of the present invention.
圖8係本發明卷積神經網路的具體實施架構示意圖。 FIG. 8 is a schematic diagram of a specific implementation architecture of a convolutional neural network of the present invention.
圖9係本發明具體的外觀實施示意圖。 Figure 9 is a schematic view showing the specific appearance of the present invention.
為讓 貴審查委員能進一步瞭解本發明整體的技術特徵與達成本發明目的之技術手段,玆以具體實施例並配合圖式加以詳細說明: In order to allow the reviewing committee to further understand the technical features of the present invention and the technical means for achieving the object of the present invention, detailed description will be made with specific embodiments and drawings:
請配合參看圖1、4所示,為達成本發明第一目的之實施例,而且本實施是一種透過影像辨識方式實現的實施例,係包括提供至少一影像擷取裝置10、一物件特徵資料庫20、一資訊處理單元30及一顯示幕40等技術特徵。影像擷取裝置10及顯示幕40皆設於車輛,用以擷取及顯示車輛行駛中的前方行車路況影像。物件特徵資料庫20建立有包含複數個不同物件特徵資料,並於每一物件特徵資料設定有一特徵資料,再於每一特徵資料設定有一物件名稱。資訊處理單元30執行時則依序包含下列之處理步驟: As shown in FIG. 1 and FIG. 4, in order to achieve the first object of the present invention, the present embodiment is an embodiment implemented by the image recognition method, which includes providing at least one image capturing device 10 and an object feature data. Technical features such as library 20, an information processing unit 30, and a display screen 40. The image capturing device 10 and the display screen 40 are both disposed in the vehicle for capturing and displaying the image of the driving road condition during the running of the vehicle. The object feature database 20 is configured to include a plurality of different object feature data, and set a feature data for each object feature data, and then set an object name for each feature data. When the information processing unit 30 executes, the following processing steps are sequentially included:
(a)影像辨識步驟:係以影像辨識分析出經過影像前處理後之行車路況影像的物件影像Ob之特徵值,再於物件特徵資料庫20比對出與特徵值大致符合的(如辨形相似度約百分之七十以上)的特徵資料,並由特徵資料得到對應的物件名稱;若是影像辨識成功率不高,則可提升辨形相似度,直到達到所需的影像辨識成功率為止。 (a) Image recognition step: analyzing the feature value of the object image Ob of the driving road condition image after image pre-processing by image recognition, and then comparing the object feature database 20 with the feature value (such as the shape recognition) Characteristic data with a similarity of about 70% or more, and the corresponding object name is obtained from the feature data; if the image recognition success rate is not high, the similarity of the shape can be improved until the desired image recognition success rate is reached. .
(b)位置距離計算步驟:係將車輛與物件(如行人、機車與人的組合或是車輛)之間的距離或是相對位置進行計算,以輸出相應的距離訊號或是位置訊號。 (b) Position distance calculation step: Calculate the distance or relative position between the vehicle and the object (such as a pedestrian, a combination of a locomotive and a person or a vehicle) to output a corresponding distance signal or position signal.
(c)行車預警輸出步驟:如圖4所示,將已辨識出物件名稱的物件於顯示幕40所顯示的行車路況影像中予以標記41,並以此標記41作為行駛中的行車預警資訊;具體來說,此標記41係為用以框選物件的框格,而且不同物件係以不同形狀的框格框選,以圖4為例,長矩形框代表車輛種類,短矩形框代表機車與人組合種類,橢圓形框代表行人種類;但是不以此為限,亦可以不同顏色的框格來代表物件名稱。除此之外,本發明可以音頻、語音、顯示或是記錄的方式將距離訊號或是位置訊號輸出 合併為上述的行車預警資訊,此行車預警資訊除了包括上述的標記41之外,更包含物件名稱資訊(如車輛、機車與人的組合、行人以及交通號誌)、物件位置資訊、物件距離資訊、交通號誌即時狀態資訊、車速資訊以及危險警告資訊。 (c) driving warning output step: as shown in FIG. 4, the object whose name has been identified is marked 41 in the driving road condition image displayed on the display screen 40, and the marking 41 is used as the driving warning information during driving; Specifically, the mark 41 is a sash for arranging objects, and different objects are selected by framing boxes of different shapes. Taking FIG. 4 as an example, a long rectangular frame represents a vehicle type, and a short rectangular frame represents a locomotive and The type of person combination, the oval frame represents the type of pedestrian; but not limited to this, the sash of different colors can also represent the object name. In addition, the present invention can output distance signals or position signals in an audio, voice, display or recording manner. Combined into the above-mentioned driving warning information, this driving warning information includes the object name information (such as vehicle, locomotive and person combination, pedestrian and traffic number), object location information, and object distance information, in addition to the above-mentioned mark 41. Traffic status, status information, speed information and hazard warning information.
承上所述,交通號誌即時狀態資訊係透過影像辨識得知交通號誌即時所處狀態為綠燈或是紅燈。至於車速資訊的取得如圖1所示,係透過全球衛星定位模組(GPS)所提供之位置訊號於一時間單位之位置變化量所計算得到;此外,亦可透過訊號擷取方式(如can bus)擷取車輛的行車電腦(ECU),於此同樣可以得到車輛車速資訊,如圖1所示。 According to the above, the traffic status real-time information is obtained through image recognition to know whether the traffic signal is in a green light or a red light. As far as the speed information is obtained, as shown in Figure 1, the position signal provided by the Global Positioning System (GPS) is calculated from the position change of a time unit; in addition, the signal acquisition method (such as can Bus) Take the vehicle's driving computer (ECU), and the vehicle speed information can also be obtained here, as shown in Figure 1.
請配合參看圖4~6所示的運作實施例中,當車輛於行駛中時,影像擷取模組10則連續擷取至少二張行車路況影像,並預先建立一與行車路況影像影像對應且包含複數座標資料的座標參數資料庫25,之後,再執行影像定位法求得物件影像Ob之重心位置,再與座標參數資料庫25之座標資料(即圖5~6中所示的座標框格,每一座標框格皆有各自定義的座標值)進行比對,並逐一計算出此物件影像Ob所處的座標位置,進而得到物件影像Ob的行進軌跡路徑資訊,以作為物件影像Ob是否會與車輛之行車路徑於一時點形成交會的判斷依據。由圖5所示的實施例中得知,本發明可由物件影像Ob出現在座標框格的位置以及面積大小,即可得到此物件影像Ob的體積大小及距離為何(可透過預先的參數設定得知;或是以二個影像擷取裝置10針對物件進行三角定位量測法而可得知與物件之間的距離資訊為何?至於物件速度資訊可由物件影像Ob移動至下一個座標框格的時間來推算,即可獲得上述的速度資訊為何?接著,資訊處理單元30再透過文字疊加技術將行車預警資訊之文字疊加於行車路況影像之中,除了由顯示幕40顯示顯示出疊加有行車預警資訊之文字的行車路況影像之外,並可由記憶裝置51來記錄疊加有行車預警資訊的行車路況影像。此外,必須 說明的是文字疊加於影像的技術確實已為非常習知的技術,故不對文字疊加技術做贅述。 Referring to the operation example shown in FIG. 4 to FIG. 6 , when the vehicle is running, the image capturing module 10 continuously captures at least two driving road image images, and pre-establishes a corresponding image of the traffic road image and The coordinate parameter database 25 containing the plurality of coordinate data, and then performing the image localization method to obtain the position of the center of gravity of the object image Ob, and then the coordinates of the coordinate parameter database 25 (ie, the coordinate frame shown in Figures 5-6) Each frame has its own defined coordinate value for comparison, and the coordinates of the object image Ob are calculated one by one, and the path information of the object image Ob is obtained as the object image Ob. Form a judgment basis for the intersection with the vehicle's driving route at a certain point. It can be seen from the embodiment shown in FIG. 5 that the present invention can obtain the position and the size of the object frame Ob by the object image Ob, and the volume size and distance of the object image Ob can be obtained (can be set by pre-determined parameters) Knowing; or using the two image capturing device 10 to perform the triangulation measurement method on the object to know the distance information between the object and the object; as for the time when the object speed information can be moved from the object image Ob to the next coordinate frame After the calculation, the speed information can be obtained. Then, the information processing unit 30 superimposes the text of the driving warning information into the road traffic image through the text superposition technology, except that the display screen 40 displays the superimposed driving warning information. In addition to the road traffic image of the text, the traffic device 51 can record the road traffic image superimposed with the driving warning information. It is explained that the technique of superimposing text on images is indeed a very well-known technique, so the text superposition technique is not described.
舉例來說,圖4~6所示為四個物件,其中,物件1(Ob1)為同向的前車,物件2(Ob2)為前車後的機車騎士,物件3(Ob3)為車道右側的等待穿越道路的行人,物件4(Ob4)為對向車輛,當資訊處理單元30影像辨識及計算出各個物件名稱、位置及距離時,資訊處理單元30則透過顯示幕40或音頻播放模組52發出行車預警資訊,行車預警資訊除了於行車路況影像顯示框選物件的標記41之外,更包括有以顯示或語音方式輸出『前方距離5公尺處有同向前車』、『右前側距離2公尺處有機車騎士』以及『慢車道右側距離25公尺處有行人,請小心駕駛』等之行車預警資訊。假設車輛與前車之距離低於5公尺時,如圖5所示,資訊處理單元30則透過顯示幕40或語音假設車輛與同向前車低於5公尺時,資訊處理單元30則透過顯示幕40或音頻播放模組52輸出『請與前車保持安全距離』。假設行人穿梭道路且經資訊處理單元30計算後得知會撞到行人時,資訊處理單元30則透過顯示幕40或音頻播放模組52輸出『請放慢車速,否則會撞到行人』之危險警告資訊。 For example, Figures 4-6 show four objects, where object 1 (Ob1) is the same car in front, object 2 (Ob2) is the locomotive knight in front of the car, and object 3 (Ob3) is the right side of the lane. The pedestrian waiting for the crossing of the road, the object 4 (Ob4) is the opposite vehicle. When the image processing unit 30 recognizes and calculates the name, position and distance of each object, the information processing unit 30 passes through the display screen 40 or the audio playing module. 52 issued the road warning information, the driving warning information in addition to the markings 41 of the object in the driving road image display box, including the display or voice output "the front distance of 5 meters at the same place forward", "right front side There is a warning message such as the 2" organic car knight" and "Pedestrians at 25 meters from the right side of the slow lane, please drive carefully". Assuming that the distance between the vehicle and the preceding vehicle is less than 5 meters, as shown in FIG. 5, the information processing unit 30 transmits the information processing unit 30 through the display screen 40 or the voice hypothesis that the vehicle and the forward vehicle are less than 5 meters. Output "Please keep a safe distance from the preceding vehicle" through the display screen 40 or the audio playback module 52. Assuming that the pedestrian shuttles the road and is informed by the information processing unit 30 that the pedestrian will be hit, the information processing unit 30 outputs a danger warning of "Please slow down the speed, otherwise it will hit the pedestrian" through the display screen 40 or the audio playback module 52. News.
請配合參看圖1~4所示,為達成本發明第二目的之第二實施例,係包括至少一影像擷取裝置10、一物件特徵資料庫20、一資訊處理單元30及一顯示幕40等技術特徵。影像擷取裝置10設於車輛內,用以擷取車輛行駛中的前方行車路況影像。物件特徵資料庫20建立有包含複數個物件特徵資料,並於每一物件特徵資料設定有一物件名稱。具體來說,深度學習演算模組31可以是一種卷積類神經網路(CNN)演算法、專家系統演算法以及隨機森林演算法等諸多的人工智慧演算法。此外,深度學習演算模組31執行時更包含下列之處理步驟: Referring to FIG. 1 to FIG. 4, in order to achieve the second embodiment of the second object of the present invention, at least one image capturing device 10, an object feature database 20, an information processing unit 30, and a display screen 40 are included. And other technical features. The image capturing device 10 is disposed in the vehicle for capturing an image of the driving road condition during running of the vehicle. The object feature database 20 is constructed to include a plurality of object feature data, and an object name is set for each object feature data. Specifically, the deep learning calculus module 31 can be a human intelligence algorithm such as a convolutional neural network (CNN) algorithm, an expert system algorithm, and a random forest algorithm. In addition, the deep learning calculus module 31 further includes the following processing steps when executed:
(a)影像辨識步驟:係於物件特徵資料庫20比對出與經影像 前處理之行車路況影像的其中至少一個物件影像符合的物件名稱。具體來說,資訊處理單元30內建一具備深度學習訓練功能以執行影像辨識步驟的深度學習演算法,用以對經影像前處理之行車路況影像依序進行特徵擷取、特徵選擇、推理及預測性影像辨識,並將行車路況影像中之至少一個特徵擷取部分影像處理為物件特徵資料,再將物件特徵資料依據物件之類別而累積至物件特徵資料庫20中,以供自我訓練學習,藉以得到至少一個物件名稱,以累積特徵值資料來強化物件特徵資料庫20的分類功能以及強化物件特徵擷取的功能。具體來說,執行深度學習演算模組31時則包含下列之階段步驟: (a) Image recognition step: the object feature database 20 is compared with the image The name of the object to which at least one object image of the pre-processed road traffic image conforms. Specifically, the information processing unit 30 has a deep learning algorithm with a deep learning training function to perform an image recognition step, and sequentially performs feature extraction, feature selection, reasoning, and image processing on the road image processed by the image pre-processing. Predictive image recognition, and at least one feature captured image of the driving road condition image is processed into object feature data, and then the object feature data is accumulated into the object feature database 20 according to the category of the object for self-training learning. Thereby, at least one object name is obtained, and the feature value data is accumulated to enhance the classification function of the object feature database 20 and the function of enhancing the object feature extraction. Specifically, when the deep learning calculus module 31 is executed, the following phase steps are included:
(a-1)訓練階段步驟,如圖2所示,係建立一深度學習模型310,以於深度學習模型310輸入距離資料及巨量的行車路況影像,並由深度學習模型310測試影像辨識的正確率,再判斷影像辨識正確率是否足夠,當判斷結果為是,則將辨識結果輸出及儲存;當判斷結果為否,則使深度學習模型310自我修正學習。 (a-1) The training phase step, as shown in FIG. 2, establishes a deep learning model 310 for inputting distance data and a large amount of driving road condition image in the deep learning model 310, and testing the image recognition by the deep learning model 310. The correct rate is determined whether the image recognition correctness rate is sufficient. When the judgment result is yes, the identification result is output and stored; when the judgment result is negative, the deep learning model 310 is self-corrected and learned.
(a-2)運行預測階段步驟,如圖3所示,係於深度學習模型310輸入距離資料及即時擷取的行車路況影像,並由深度學習模型310進行預測性影像辨識,以得到至少一個辨識結果的物件名稱及距離值,再將物件名稱及距離值輸出。 (a-2) running the prediction phase step, as shown in FIG. 3, inputting the distance data and the immediately captured driving road condition image in the deep learning model 310, and performing predictive image recognition by the deep learning model 310 to obtain at least one Identify the object name and distance value of the result, and then output the object name and distance value.
(b)位置距離計算步驟:係將車輛與物件之間的距離或是相對位置進行計算,以輸出相應的距離訊號或是位置訊號。 (b) Position distance calculation step: Calculate the distance or relative position between the vehicle and the object to output the corresponding distance signal or position signal.
(c)行車預警輸出步驟:將已辨識出物件名稱的物件於顯示幕40所顯示的行車路況影像中予以標記41,並以此標記41作為行駛中的行車預警資訊;具體來說,上述標記41係為用以框選物件的框格,不同物件係以不同形狀的框格來框選;或是以不同顏色的框格來框選,於此,駕駛人即可從框格的形狀或是顏色辨識出物件的名稱種類;不僅如此,亦 可以不同形狀或是不同顏色的框格作為行車預警等級的顯示資訊;或是以不同音頻來代表物件名稱;於此,駕駛人即可從框格的形狀或是顏色辨識出物件相對車輛的預警等級為何? (c) driving warning output step: marking the object whose name has been identified in the driving road condition image displayed on the display screen 40, and using the marking 41 as the driving warning information during driving; specifically, the above marking The 41 series is a sash for selecting objects, and the different objects are framed by sashes of different shapes; or by sashes of different colors, the driver can take the shape of the sash or Is the name of the object that the color recognizes; not only that, but also The sash of different shapes or different colors can be used as the display information of the driving warning level; or the object name can be represented by different audio; here, the driver can recognize the object relative warning from the shape or color of the sash. What is the rating?
如圖1、9所示的實施例中,更包含一用以容裝影像擷取裝置10、資訊處理單元30及顯示幕40的機殼50;此機殼50前面設置影像擷取裝置10,其背面設有顯示幕40,顯示幕40顯示有行車路況影像以及疊加在行車路況影像上的行車預警資訊。另外,機殼50設有一用以記錄行車路況影像及行車預警資訊的記憶裝置51。此外,必須說明的是,上述影像擷取裝置10係為一種可以作為測距及記錄之用的複鏡頭影像擷取裝置,於此,即可透過三角定位量測法來求出物件與車輛之間的距離。進一步來說,三角影像定位量測法可以是一種習知的Epipolar平面三角測距法,主要是以複鏡頭影像擷取裝置10來擷取左側行車路況影像以及右側行車路況影像,資訊處理單元10則將左側與右側之行車路況影像進行Epipolar平面之三角測距法的演算,進而求出車輛與物件Ob之間的距離,由於Epipolar平面三角測距法係為非常習知的技術,故不再予以贅述具體的運算內容。 The embodiment shown in FIG. 1 and FIG. 9 further includes a casing 50 for housing the image capturing device 10, the information processing unit 30, and the display screen 40. The image capturing device 10 is disposed in front of the casing 50. The back side is provided with a display screen 40, and the display screen 40 displays driving road condition images and driving warning information superimposed on the driving road condition image. In addition, the casing 50 is provided with a memory device 51 for recording driving road condition images and driving warning information. In addition, it should be noted that the image capturing device 10 is a complex lens image capturing device that can be used for ranging and recording. Here, the object and the vehicle can be obtained through the triangulation measurement method. The distance between them. Further, the triangular image localization measurement method may be a conventional Epipolar planar triangulation method, which mainly uses the multi-lens image capturing device 10 to capture the left driving road condition image and the right driving road condition image, and the information processing unit 10 Then the left and right road traffic images are calculated by the triangulation method of the Epipolar plane, and then the distance between the vehicle and the object Ob is obtained. Since the Epipolar plane triangulation method is a very well-known technique, it is no longer The specific operation content will be described.
請參看圖1、7所示的實施例中,資訊處理單元30包含一危險警告評估模組32,而危險警告資訊則是危險警告評估模組32所計算評估,其係判斷物件之預測路徑D1是否與車輛之行車路徑D2於一個時點形成交會點,判斷結果為是,則發出危險警告資訊。再者,圖1所示之物件特徵資料庫20係包含有一儲存複數車輛特徵樣本的車輛特徵資料21、一儲存複數機車與人組合特徵樣本的機車/人特徵資料22、一儲存複數行人特徵樣本的行人特徵資料23以及一儲存複數交通號誌特徵樣本的交通號誌特徵資料24。 Referring to the embodiment shown in FIG. 1 and FIG. 7, the information processing unit 30 includes a hazard warning evaluation module 32, and the hazard warning information is the evaluation calculated by the hazard warning evaluation module 32, which is the predicted path D of the object. 1 Whether or not a crossing point is formed at a time point with the vehicle's driving route D 2 , and if the judgment result is yes, a danger warning message is issued. Furthermore, the object feature database 20 shown in FIG. 1 includes a vehicle feature data 21 storing a plurality of vehicle feature samples, a locomotive/person profile data storing a plurality of locomotive and human combination feature samples, and a stored plurality of pedestrian feature samples. The pedestrian characteristic data 23 and a traffic sign feature data 24 for storing a plurality of traffic sign feature samples.
請參看圖1所示,資訊處理單元30係包含一行車危險等級預警模組33,此行車危險等級預警模組33針對物件之移動路徑是否與車輛 之行車動線於一個時點形成交會點的可能性進行預測評估,並依據各物件所評估的危險等級排序而輸出行車預警資訊。舉圖4、7為例來說明,圖4中的機車則列為最高等級的行車預警資訊,如以語音及記錄方式輸出:『小心前方機車騎士的行車動線』;其次,圖4中的同向汽車則列為第二等級的行車預警資訊,如以語音及記錄方式輸出:『小心保持與前方車輛的安全車距』;圖4中之行人若是處停止等待狀態則將行人列為最低等級的行車預警資訊,如以語音及記錄方式輸出:『請持續注意前方右側行人的動線』。 Referring to FIG. 1 , the information processing unit 30 includes a row of vehicle danger level warning module 33 , and the driving danger level warning module 33 determines whether the moving path of the object is related to the vehicle. The driving line analyzes the possibility of forming a meeting point at a certain point in time, and outputs the driving warning information according to the ranking of the hazard levels evaluated by the objects. Take 4 and 7 as an example to illustrate, the locomotive in Figure 4 is listed as the highest level of driving warning information, such as voice and recording output: "Be careful of the driving line of the front locomotive knight"; secondly, in Figure 4 The same-direction car is listed as the second-level driving warning information, such as voice and recording output: "Be careful to maintain the safe distance from the vehicle in front"; if the pedestrian in Figure 4 stops waiting, the pedestrian is listed as the lowest. The level of driving warning information, such as voice and record output: "Please continue to pay attention to the right side of the pedestrian line."
經由上述具體實施例的說明后得知,本發明第二實施例是採用深度學習技術為基底,以運用於先進駕駛之輔助系統的前方辨識,藉以辨識行車路況影像中的汽車、行人、機車以及交通號誌等物件,進而達到行車預警的功效,並將深度學習演算法建構於行動裝置專用的晶片組(含圖形處理器GPU)中,以達到即時偵測行車動態之目的,最終包裝成先進駕駛輔助系統之前方影像辨識解決方案提供給相關業者。再者,深度學習中的卷積類神經網路(Convolutional Neural Network;CNN)在2012年之後,即大舉突破早期分類器與方法的辨識率。一般傳統影像辨識往往會因為前段的特徵擷取演算法容易受到環境變化而失去辨識準確率。為此,本發明即是改善此一缺失,係利用卷積類神經網路(Convolutional Neural Network;CNN))將前段的特徵擷取部分利用深度學習方式來自我學習修正,只要建立數量充足的物件特徵資料,讓深度學習模型自我學習修正,於此,既可達到高度的辨識率,並可大舉突破一般傳統的機器學習方法。 Through the description of the above specific embodiments, it is known that the second embodiment of the present invention uses a deep learning technique as a base for the front identification of an auxiliary system for advanced driving, thereby identifying cars, pedestrians, locomotives, and the like in driving road conditions images. Traffic signs and other objects, in order to achieve the effect of driving warning, and the deep learning algorithm is built in the mobile device-specific chipset (including graphics processor GPU), in order to achieve the purpose of real-time detection of driving dynamics, and finally packaged into advanced The driver identification system's previous image recognition solution is provided to the relevant industry. Furthermore, the Convolutional Neural Network (CNN) in deep learning has broken through the recognition rates of early classifiers and methods after 2012. In general, traditional image recognition often loses the recognition accuracy because the feature extraction algorithm in the previous segment is susceptible to environmental changes. To this end, the present invention is to improve this deficiency, using the Convolutional Neural Network (CNN) to use the deep learning method from the learning section of the previous paragraph to learn the correction, as long as a sufficient number of objects are established. The feature data allows the deep learning model to self-learn and correct. In this way, it can achieve a high recognition rate and can break through the traditional traditional machine learning methods.
一般而言,卷積神經網路從影像擷取裝置10獲得行車路況影像後,經過影像前處理(即預處理)、特徵擷取、特徵選擇,再到推理以及做出預測性辨識。另一方面,卷積神經網路的深度學習實質,是通過構建具有多個隱層的機器學習模型及海量訓練數據,來達到學習更有用的特徵,從而最終提升分類或預測的準確性。卷積神經網路利用海量訓練數 據來學習特徵辨識,於此方能刻畫出數據的豐富內在訊息。由於卷積神經網路為一種權值共享的網路結構,所以除了可以降低網路模型的複雜度之外,並可減少權值的數量。此優點在網路的輸入是多維圖像時表現的更為明顯,使圖像可以直接作為網路的輸入,避免了傳統影像辨識演算法中複雜的特徵擷取與數據重建過程。物件分類方式幾乎都是基於統計特徵的,這就意味著在進行分辨前必須提取某些特徵。然而,顯式的特徵擷取並不容易,在一些應用問題中也並非總是可靠的。卷積神經網路可避免顯式的特徵取樣,隱式地從訓練數據中進行學習。這使得卷積神經網路明顯有別於其他基於神經網路的分類器,通過結構重組和減少權值將特徵擷取功能融合進多層感知器。它可以直接處理灰度圖片,能夠直接用於處理基於圖像的分類。卷積網路較一般神經網路在圖像處理方面有如下優點:輸入圖像與網路的拓撲結構能很好的吻合;特徵擷取與模式分類同時進行,並同時在訓練中產生;權重共享可以減少網路的訓練參數,使神經網路結構變得更為簡單,適應性更強。 Generally, the convolutional neural network obtains the driving road condition image from the image capturing device 10, and then undergoes image pre-processing (ie, pre-processing), feature extraction, feature selection, and then reasoning and predictive identification. On the other hand, the deep learning essence of convolutional neural networks is to build more useful features by constructing machine learning models with multiple hidden layers and massive training data, so as to improve the accuracy of classification or prediction. Convolutional neural networks use massive training numbers According to the study of feature identification, it is possible to portray the rich internal information of the data. Since the convolutional neural network is a weight-sharing network structure, in addition to reducing the complexity of the network model, the number of weights can be reduced. This advantage is more obvious when the input of the network is a multi-dimensional image, so that the image can be directly used as the input of the network, avoiding the complicated feature extraction and data reconstruction process in the traditional image recognition algorithm. Object classification is almost always based on statistical features, which means that certain features must be extracted before resolving. However, explicit feature extraction is not easy and is not always reliable in some application problems. The convolutional neural network avoids explicit feature sampling and implicitly learns from the training data. This makes the convolutional neural network distinct from other neural network-based classifiers, merging feature extraction functions into multilayer perceptrons through structural reorganization and reduced weights. It can process grayscale images directly and can be used directly to process image-based classifications. The convolutional network has the following advantages in image processing compared with the general neural network: the input image and the network topology can be well matched; the feature extraction is performed simultaneously with the pattern classification, and simultaneously generated in the training; Sharing can reduce the training parameters of the network, making the neural network structure simpler and more adaptable.
請參看圖8所示,係為卷積神經網路的具體模型架構示意,輸入行車路況影像通過與三個可訓練的濾波器與可加偏置進行卷積處理,卷積處理後在C1層產生三個特徵映射圖,然後此特徵映射圖中每組的四個像素再進行求和,加權值,加偏置;接著,透過一個函數(Sigmoid)得到三個S2層的特徵映射圖,此映射圖再進過濾波得到C3層。這個層級結構再和S2一樣產生S4...Sn。最後,這些像素值被光柵化,並連接成一個向量輸入到傳統的神經網路,進而得到輸出。 Please refer to FIG. 8 , which is a schematic diagram of a specific model structure of the convolutional neural network. The input driving road condition image is convoluted by the three trainable filters and the addable bias, and the convolution processing is performed on the C1 layer. Three feature maps are generated, and then four pixels of each group in the feature map are summed, weighted, and offset; then, a feature map of three S2 layers is obtained through a function (Sigmoid). The map is filtered again to obtain the C3 layer. This hierarchical structure produces S4...Sn as well as S2. Finally, these pixel values are rasterized and concatenated into a vector input to a traditional neural network to obtain an output.
一般而言,C層為特徵擷取層,每個神經元的輸入與前一層的局部感受野相連,並擷取該局部的特徵,一旦該局部特徵被擷取後,它與其他特徵間的位置關係也隨之確定下來。至於S層則是特徵映射層,網路的每個計算層由多個特徵映射組成,每個特徵映射為一個平面,平面上 所有神經元的權值相等。特徵映射結構採用影響函數核小的sigmoid函數作為卷積網路的激活函數(Activation function;(1)Tan H,(2)Rectifier Relu),(3)Parameteric Rect,fied Lineor Unit(Prelu),使得特徵映射具有位移不變性。此外,由於一個映射面上的神經元共享權值,因而減少了網路自由參數的個數,降低了網路參數選擇的複雜度。卷積神經網路中的每一個特徵擷取層(C-層)都緊跟著一個用來求局部平均與二次提取的計算層(S-層),這種特有的兩次特徵擷取結構使網路在識別時對輸入樣本有較高的畸變容忍能力。 In general, the C layer is a feature extraction layer, and the input of each neuron is connected to the local receptive field of the previous layer, and the local feature is extracted, and once the local feature is captured, it is interposed with other features. The positional relationship is also determined. As for the S layer, it is a feature mapping layer. Each computing layer of the network is composed of multiple feature maps, each of which is mapped to a plane, on a plane. All neurons have equal weights. The feature mapping structure uses a small sigmoid function that affects the function kernel as the activation function of the convolution network (Activation function; (1) Tan H, (2) Rectifier Relu), (3) Parameteric Rect, fied Lineor Unit (Prelu), The feature map has displacement invariance. In addition, since the neurons on a mapped surface share weights, the number of network free parameters is reduced, which reduces the complexity of network parameter selection. Each feature extraction layer (C-layer) in the convolutional neural network is followed by a computational layer (S-layer) for local averaging and secondary extraction. This unique feature extraction is performed twice. The structure allows the network to have high distortion tolerance for input samples when it is identified.
卷積神經網路主要用來識別位移、縮放及其他形式扭曲不變性的二維圖形。由於卷積神經網路的特徵檢測層通過訓練數據進行學習,所以在使用卷積神經網路時,避免了顯式的特徵擷取,而隱式地從訓練數據中進行學習;再者由於同一特徵映射面上的神經元權值相同,所以網路可以並行學習,這也是卷積神經網路相對於神經元彼此相連網路的一大優勢。卷積神經網路以其局部權值共享的特殊結構在語音識別和圖像處理方面有著獨特的優越性,其布局更接近於實際的生物神經網路,權值共享降低了網路的複雜性,特別是多維輸入向量的圖像可以直接輸入網路這一特點避免了特徵擷取與分類過程中數據重建的複雜度。 Convolutional neural networks are primarily used to identify two-dimensional graphics of displacement, scaling, and other forms of distortion invariance. Since the feature detection layer of the convolutional neural network learns through the training data, when using the convolutional neural network, explicit feature extraction is avoided, and learning is implicitly learned from the training data; The weights of the neurons on the feature mapping surface are the same, so the network can learn in parallel, which is also a major advantage of the convolutional neural network relative to the neural network connected to each other. The convolutional neural network has unique advantages in speech recognition and image processing with its special structure of local weight sharing. Its layout is closer to the actual biological neural network, and weight sharing reduces the complexity of the network. In particular, the image of the multi-dimensional input vector can be directly input into the network, which avoids the complexity of data reconstruction in the feature extraction and classification process.
以上所述,僅為本發明之可行實施例,並非用以限定本發明之專利範圍,凡舉依據下列請求項所述之內容、特徵以及其精神而為之其他變化的等效實施,皆應包含於本發明之專利範圍內。本發明所具體界定於請求項之結構特徵,未見於同類物品,且具實用性與進步性,已符合發明專利要件,爰依法具文提出申請,謹請 鈞局依法核予專利,以維護本申請人合法之權益。 The above is only a possible embodiment of the present invention, and is not intended to limit the scope of the patents of the present invention, and the equivalent implementations of other changes according to the contents, features and spirits of the following claims should be It is included in the patent of the present invention. The invention is specifically defined in the structural features of the request item, is not found in the same kind of articles, and has practicality and progress, has met the requirements of the invention patent, and has filed an application according to law, and invites the bureau to approve the patent according to law to maintain the present invention. The legal rights of the applicant.
Claims (9)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
TW105134417A TWI619099B (en) | 2016-10-25 | 2016-10-25 | Intelligent multifunctional driving assisted driving recording method and system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
TW105134417A TWI619099B (en) | 2016-10-25 | 2016-10-25 | Intelligent multifunctional driving assisted driving recording method and system |
Publications (2)
Publication Number | Publication Date |
---|---|
TWI619099B true TWI619099B (en) | 2018-03-21 |
TW201816745A TW201816745A (en) | 2018-05-01 |
Family
ID=62189420
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
TW105134417A TWI619099B (en) | 2016-10-25 | 2016-10-25 | Intelligent multifunctional driving assisted driving recording method and system |
Country Status (1)
Country | Link |
---|---|
TW (1) | TWI619099B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111464797A (en) * | 2020-06-03 | 2020-07-28 | 上海商汤临港智能科技有限公司 | Driving assistance apparatus, method, vehicle, and storage medium |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10748033B2 (en) | 2018-12-11 | 2020-08-18 | Industrial Technology Research Institute | Object detection method using CNN model and object detection apparatus using the same |
TWI717102B (en) * | 2019-11-14 | 2021-01-21 | 黃玄 | Traffic condition system for internet of vehicles based on image recognition |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2002029349A (en) * | 2000-07-13 | 2002-01-29 | Nissan Motor Co Ltd | Device for recognizing vehicular circumference |
JP2005329779A (en) * | 2004-05-19 | 2005-12-02 | Daihatsu Motor Co Ltd | Method and device for recognizing obstacle |
JP2006279114A (en) * | 2005-03-28 | 2006-10-12 | Clarion Co Ltd | Circumference monitor for vehicle |
CN102236966A (en) * | 2010-04-21 | 2011-11-09 | 昆达电脑科技(昆山)有限公司 | System and method for combining traveling record and early warning device with portable electronic device |
CN103879404A (en) * | 2012-12-19 | 2014-06-25 | 财团法人车辆研究测试中心 | Moving-object-traceable anti-collision warning method and device thereof |
TW201437985A (en) * | 2013-03-18 | 2014-10-01 | Create Electronic Optical Co Ltd | Tachograph with road detection and warning function in safe range |
US9430946B2 (en) * | 2011-12-28 | 2016-08-30 | Toyota Jidosha Kabushiki Kaisha | Obstacle determination device |
TW201643836A (en) * | 2015-06-05 | 2016-12-16 | Papago Inc | Dashboard camera and living object detection method applying the same |
-
2016
- 2016-10-25 TW TW105134417A patent/TWI619099B/en active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2002029349A (en) * | 2000-07-13 | 2002-01-29 | Nissan Motor Co Ltd | Device for recognizing vehicular circumference |
JP2005329779A (en) * | 2004-05-19 | 2005-12-02 | Daihatsu Motor Co Ltd | Method and device for recognizing obstacle |
JP2006279114A (en) * | 2005-03-28 | 2006-10-12 | Clarion Co Ltd | Circumference monitor for vehicle |
CN102236966A (en) * | 2010-04-21 | 2011-11-09 | 昆达电脑科技(昆山)有限公司 | System and method for combining traveling record and early warning device with portable electronic device |
US9430946B2 (en) * | 2011-12-28 | 2016-08-30 | Toyota Jidosha Kabushiki Kaisha | Obstacle determination device |
CN103879404A (en) * | 2012-12-19 | 2014-06-25 | 财团法人车辆研究测试中心 | Moving-object-traceable anti-collision warning method and device thereof |
TW201437985A (en) * | 2013-03-18 | 2014-10-01 | Create Electronic Optical Co Ltd | Tachograph with road detection and warning function in safe range |
TW201643836A (en) * | 2015-06-05 | 2016-12-16 | Papago Inc | Dashboard camera and living object detection method applying the same |
Non-Patent Citations (1)
Title |
---|
MOBILEYE C2-270,< http://www.zingher.com.tw/ProductSpec/Mobileye/>,上傳時間:2012年12月9日 * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111464797A (en) * | 2020-06-03 | 2020-07-28 | 上海商汤临港智能科技有限公司 | Driving assistance apparatus, method, vehicle, and storage medium |
Also Published As
Publication number | Publication date |
---|---|
TW201816745A (en) | 2018-05-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106845547B (en) | A kind of intelligent automobile positioning and road markings identifying system and method based on camera | |
JP6599986B2 (en) | Hyperclass expansion and regularization deep learning for fine-grained image classification | |
CN107730904A (en) | Multitask vehicle driving in reverse vision detection system based on depth convolutional neural networks | |
JP2016062610A (en) | Feature model creation method and feature model creation device | |
CN113139470B (en) | Glass identification method based on Transformer | |
CN111899515B (en) | Vehicle detection system based on wisdom road edge calculates gateway | |
TWI619099B (en) | Intelligent multifunctional driving assisted driving recording method and system | |
CN114802261B (en) | Parking control method, obstacle recognition model training method and device | |
CN107315998A (en) | Vehicle class division method and system based on lane line | |
CN111256693A (en) | Pose change calculation method and vehicle-mounted terminal | |
Zhang et al. | Road marking segmentation based on siamese attention module and maximum stable external region | |
CN113052071B (en) | Method and system for rapidly detecting distraction behavior of driver of hazardous chemical substance transport vehicle | |
Reza et al. | An ai-based real-time roadway-environment perception for autonomous driving | |
Zhao et al. | Improving autonomous vehicle visual perception by fusing human gaze and machine vision | |
Rocky et al. | Review of Accident Detection Methods Using Dashcam Videos for Autonomous Driving Vehicles | |
Qin et al. | An improved deep learning algorithm for obstacle detection in complex rail transit environments | |
CN113449629B (en) | Lane line false and true identification device, method, equipment and medium based on driving video | |
TWM538142U (en) | Smart multifunctional driver-assisting driving recording system | |
CN115424225A (en) | Three-dimensional real-time target detection method for automatic driving system | |
Park et al. | Roadway image preprocessing for deep learning-based driving scene understanding | |
Rakesh et al. | Machine Learning and Internet of Things-based Driver Safety and Support System | |
Ramakrishnan et al. | Autonomous Vehicle Image Classification using Deep Learning | |
Du et al. | A Lightweight and Accurate Method for Detecting Traffic Flow in Real Time | |
US20240282072A1 (en) | Method for generating improved training data video and apparatus thereof | |
Huang et al. | Driving safety prediction and safe route mapping using in-vehicle and roadside data |