WO2023109640A1 - Interpretability method and system for deep reinforcement learning model in driverless scene - Google Patents

Interpretability method and system for deep reinforcement learning model in driverless scene Download PDF

Info

Publication number
WO2023109640A1
WO2023109640A1 PCT/CN2022/137511 CN2022137511W WO2023109640A1 WO 2023109640 A1 WO2023109640 A1 WO 2023109640A1 CN 2022137511 W CN2022137511 W CN 2022137511W WO 2023109640 A1 WO2023109640 A1 WO 2023109640A1
Authority
WO
WIPO (PCT)
Prior art keywords
reinforcement learning
model
unmanned driving
learning model
deep reinforcement
Prior art date
Application number
PCT/CN2022/137511
Other languages
French (fr)
Chinese (zh)
Inventor
周纪民
刘延东
张中劲
王鲁佳
王洋
须成忠
Original Assignee
深圳先进技术研究院
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳先进技术研究院 filed Critical 深圳先进技术研究院
Publication of WO2023109640A1 publication Critical patent/WO2023109640A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F30/00Computer-aided design [CAD]
    • G06F30/20Design optimisation, verification or simulation
    • G06F30/27Design optimisation, verification or simulation using machine learning, e.g. artificial intelligence, neural networks, support vector machines [SVM] or training a model
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2119/00Details relating to the type or aim of the analysis or the optimisation
    • G06F2119/02Reliability analysis or reliability optimisation; Failure analysis, e.g. worst case scenario performance, failure mode and effects analysis [FMEA]

Definitions

  • the invention belongs to the field of model learning, and in particular relates to an interpretability method and system for a deep reinforcement learning model in an unmanned driving scene.
  • the interpretability technology of the deep reinforcement learning model in the unmanned driving scenario realizes the interpretation of the opaque model in the unmanned driving scenario, and has built-in deep reinforcement learning model problems and interpretation problems in the unmanned driving scenario.
  • Solving algorithms and optimization schemes which explains and visualizes to the user the important factors in the model's operation and decision-making process in an unmanned driving environment.
  • the deep reinforcement learning model mainly involves the selection of various deep reinforcement learning algorithms (Deep Reinforcement Learning, DRL) for the autonomous decision-making of agents in an unmanned driving environment.
  • DRL Deep Reinforcement Learning
  • explainable artificial intelligence is mainly to explain and visualize various AI algorithm models.
  • DRL is an algorithm combining deep learning and reinforcement learning. It combines the perception ability of deep learning and the decision-making ability of reinforcement learning, and has been developed at a deeper level, including value-based DRL, policy-based (policy- based) DRL, model-based DRL, hierarchical-based DRL, etc.
  • XRL is interpretable reinforcement learning, which is used to explain and visualize DRL.
  • XRL technologies are classified as follows:
  • intrinsic interpretability intrinsic interpretability
  • intrinsic interpretation is constructed to be intrinsically interpretable or itself interpretable during training, such as decision trees
  • post-hoc interpretability post-hoc interpretability
  • Alternative models or saliency maps are typical by providing an explanation for the original model after training by creating a second simpler model or other operations such as perturbation.
  • Interpretability technology of deep reinforcement learning models is a key issue in the field of unmanned driving and computers.
  • XRL has not been widely studied.
  • the current research direction of deep reinforcement learning is post-hoc interpretability, both global interpretability and local interpretability.
  • post-hoc interpretability is post-hoc interpretability, both global interpretability and local interpretability.
  • XRL started late, there are currently several typical studies on post-hoc interpretability.
  • the explanatory algorithms for other artificial intelligence models in XAI can also be used in the interpretation of DRL.
  • the saliency map algorithm (Saliencymap) explained by disturbing the input needs to perform regional Gaussian blurring on the input image at a certain interval, and input the blurred image into the network every time it is blurred, and the obtained value and the value obtained by entering the original image into the network Take the difference to get the degree of influence of the area on the model decision. In this way, it is not easy to obtain the influence of a specific feature in the picture on the model decision through uniform blurring.
  • the LIME algorithm which trains a simple model to approximate a complex model, uses a simple model to approximate a complex classification network, and uses a simple one-dimensional linear model to perform one-dimensional quantization and disturbance on the input image to approximate the original model. Finally, the model can be explained by looking at the magnitude of the coefficients of the linear model. This method can well explain the influence of the characteristics of the input image on the model's decision.
  • LIME can only explain one sample at a time, and needs to build a new model each time. Although this algorithm is more general and accurate, it takes a long time to use, and it is difficult to use the data to update the network. And it is not very suitable for scenes with fast scene changes and high speed requirements.
  • Embodiments of the present invention provide an interpretability method and system for a deep reinforcement learning model in an unmanned driving scene, to at least solve the technical problem that the prior art cannot accurately analyze the influence of each feature of a picture on model decision-making.
  • an interpretability method for a deep reinforcement learning model in an unmanned driving scene comprising the following steps:
  • Input the pictures taken in the unmanned driving scene to the reinforcement learning model divide the features of the pictures and perform quantitative analysis of the influence of the features, calculate the degree of influence of each feature on the decision of the model, and obtain the corresponding difference matrix, and get the improved model network model.
  • the resulting improved network model includes:
  • the state image is obtained through the interaction between the model and the environment, and the image is divided into a fixed number of blocks according to the characteristics through superpixel segmentation, and the image set is obtained by sequentially blurring the Gaussian blur method of the irregular area;
  • the difference matrix is up-sampled so that the size of the matrix is equal to the size of the input image, and the value of the difference matrix is multiplied by a preset multiple and superimposed on the original image.
  • A3C in deep reinforcement learning is selected as the algorithm for the autonomous decision-making of the agent in unmanned driving.
  • the unmanned driving environment selects the carla simulation environment, selects a suitable scene, and selects a picture as input.
  • the reinforcement learning model before inputting the pictures taken in the unmanned driving scene to the reinforcement learning model, it also includes: preprocessing the pictures taken in the unmanned driving scene.
  • preprocessing the pictures taken in unmanned driving scenarios includes:
  • image segmentation forms adjacent pixels with similar texture, color, and brightness characteristics into visually meaningful irregular pixel blocks, and replaces a large number of pixels with a small number of pixels; where image blurring is the average value of surrounding pixels for each pixel .
  • the method also includes:
  • an interpretability system for a deep reinforcement learning model in an unmanned driving scenario including:
  • the network model module is used to select a suitable simulation environment and a suitable deep reinforcement learning algorithm, and obtain a convergent reinforcement learning model through training;
  • the explanatory algorithm module is used to input the pictures taken in the unmanned driving scene to the reinforcement learning model, divide the features of the pictures and perform quantitative analysis on the influence of features, calculate the degree of influence of each feature on the model decision, and obtain the corresponding
  • the difference matrix is used to obtain an improved network model.
  • the interpretability method and system of the deep reinforcement learning model in the unmanned driving scene in the embodiment of the present invention select a suitable simulation environment and a suitable depth reinforcement learning algorithm, obtain a convergent reinforcement learning model through training, and input the reinforcement learning model
  • the features of the pictures are divided and the quantitative analysis of the influence of the features is carried out, the degree of influence of each feature on the model decision is calculated, and the corresponding difference matrix is obtained, and the improved network model is obtained.
  • Fig. 1 is the frame diagram of the overall design of the interpretability method and system of the deep reinforcement learning model in the unmanned driving scene of the present invention
  • Fig. 2 is a working flow chart of the interpretability method and system of the deep reinforcement learning model in the unmanned driving scene of the present invention.
  • the invention is a novel XRL algorithm, aiming at the deep reinforcement learning model, providing a fast and accurate solution for the interpretation and visualization of its decision-making; it quantifies the influence of the features determined in the input picture on the decision-making of the model; in order to improve Speed, reducing the number of superpixel blocks of each picture; in order to design a general interpretability algorithm that adapts to various actual scenarios, it does not depend on a specific model in the design process, so the XRL proposed by the present invention does not depend on A certain model (Model-free), and adapt to the actual scene problems, XRL should also have a certain degree of flexibility and scalability, so as to be able to adapt to various scenes with different numbers of features.
  • the problem to be solved by the present invention is to use algorithms such as Superpixel segmentation, Gaussian Blur, Saliencymap, and Deep reinforcement learning to solve the unknowable problem of deep reinforcement learning models in unmanned driving scenarios and Its extended problem enables users to understand the favorable and unfavorable factors in the decision-making process through explanatory algorithms, and presents them to users through good human-computer interaction. Present the decision-making basis of the intelligent body to the user, and increase the user's trust in the unmanned driving model.
  • algorithms such as Superpixel segmentation, Gaussian Blur, Saliencymap, and Deep reinforcement learning
  • the interpretability method and the overall system design framework of the deep reinforcement learning model in the unmanned driving scenario are composed of three parts: the network model part, the explanatory algorithm part, and the network improvement part, as shown in Figure 1.
  • the network model part includes the selection of deep reinforcement learning algorithms, scene design and model training in unmanned driving scenarios.
  • the present invention needs to select a suitable simulation environment and a suitable deep reinforcement learning algorithm in advance.
  • A3C Asynchronous advantageous actor-critic
  • the unmanned driving environment selects the carla simulation environment, selects a suitable scene, and selects a picture as input: then through training, a convergent reinforcement learning model is finally obtained, and at this time, the model that is explained next is obtained.
  • the explanatory algorithm part includes image preprocessing, saliency map operation (solution of difference matrix), visualization and other modules.
  • the results of image preprocessing will be beneficial to the operation of the Saliencymap module and the division of features, and will be conducive to the quantitative analysis of the influence of explanatory algorithms on features;
  • the Saliencymap module will calculate the impact of each feature on The degree of influence of model decision-making, and obtain the corresponding difference matrix, so as to obtain the important factors in the decision-making process of the model;
  • the visualization module presents the explained content to users in a form that users can easily understand.
  • Unmanned driving scene The scene is as rich as possible, close to the real situation, and a convergent deep reinforcement learning model is obtained.
  • the preprocessing part tries to make the number of features separated from the picture appropriate.
  • the area of the saliency map should be as convergent as possible, and not too scattered.
  • the present invention designs the interpretability method and system workflow of the deep reinforcement learning model in the unmanned driving scene, as shown in FIG. 2 .
  • the present invention When the present invention obtains the required model, it starts to interpret the model.
  • the state image is obtained through the interaction between the model and the environment, and the image is divided into a fixed number of blocks according to the characteristics through superpixel segmentation.
  • the method of Gaussian Blur (GaussianBlur) in the regular area blurs the image set separately in turn.
  • the image set and the original image are input into the network separately, so that the decision value of the original image and the blurred image are obtained, and the difference between the two is obtained to obtain the difference matrix.
  • the difference matrix is up-sampled so that the size of the matrix is equal to the size of the input image, and the value of the difference matrix is multiplied by a certain multiple and superimposed on the original image, so as to be displayed to the user in the form of a saliency map. Afterwards, the area of the meaningful part is significantly enhanced, and the purpose of improving the network model is obtained.
  • the image preprocessing problem of XRL is the premise of XRL analysis. Converting the input image into the form required for the explanation of the present invention can be described as: find out the appropriate image features in the unmanned driving environment and segment them, which can be Some unimportant or relatively small features are not divided, and the minimum number of segmentation blocks is used to include the features required in the unmanned driving environment, thereby greatly reducing the time spent and achieving the desired effect.
  • the preprocessing algorithm based on superpixel segmentation and Gaussian blur used in the present invention can well realize the segmentation of the main features of the input image and the Gaussian blur of irregular features, and is a better preprocessing method that can be used for deep reinforcement learning model interpretation .
  • Image preprocessing follows the traditional image processing algorithm process: image segmentation forms irregular pixel blocks with certain visual significance from adjacent pixels with similar texture, color, brightness and other characteristics, and replaces a large number of pixels with a small number of pixels. Image blurring can be understood as taking the average value of surrounding pixels for each pixel.
  • the blurred part of the picture is to eliminate some features, so that the image is different from the original image in terms of features, which is convenient for subsequent comparison with the strategy obtained from the original image.
  • the blurred parts it should be noted that the blurred parts should be as far as possible Smooth transition with the unblurred part, so as not to affect the model's decision-making due to the obvious boundary between the blurred part and the unblurred part.
  • Feature interpretation Process the picture according to the feature area, and obtain and quantify the influence of the feature area on the model decision-making. Through normalization processing, each feature is contrasted, and the positive contribution of each feature to the decision-making is calculated. Influence and negative influence, the obtained data is conducive to the next update of the model.
  • the obtained explanation can know which are the favorable factors and which are the unfavorable factors during the normal operation of the model.
  • the model can also know which feature of the input image caused the failure of the system's decision-making. This information can be used to improve the model.
  • Key point of the present invention and want to protect point are at least:
  • the present invention aims at the interpretability scene of the deep reinforcement learning model in the unmanned driving scene to solve the opaque problem of the DRL model in the scene, explain its decision-making, and provide a visualized human-computer interaction interface, which explains the depth to a certain extent.
  • the interpretability of the reinforcement learning model increases the user's trust, and at the same time provides a basis for the improvement of the model.
  • the present invention mainly embodies the following advantages:
  • image features can be smoothed and blurred in irregular areas, so that the blurred area and the unblurred area can be smoothly handed over;
  • Deep reinforcement learning models can be further improved through the content of the explanation, which is currently not covered by explanation systems and model improvement systems.
  • the XRL and model improvement schemes are verified, and the XRL algorithm is verified.
  • the visualization algorithm of the simulation platform the user understands the decision-making basis of the model, and improves the model based on this basis.
  • the alternative scheme of the present invention is at least:
  • the XRL system is scalable, and expansion modules can be combined arbitrarily to meet customer needs. For example, adding or changing the image preprocessing process, changing the perturbation method of the image, changing the difference calculation method, etc.
  • a unit described as a separate component may or may not be physically separated, and a component shown as a unit may or may not be a physical unit, that is, it may be located in one place, or may be distributed over multiple units. Part or all of the units can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
  • each functional unit in each embodiment of the present invention may be integrated into one processing unit, each unit may exist separately physically, or two or more units may be integrated into one unit.
  • the above-mentioned integrated units can be implemented in the form of hardware or in the form of software functional units.
  • the integrated unit is realized in the form of a software function unit and sold or used as an independent product, it can be stored in a computer-readable storage medium.
  • the computer software product is stored in a storage medium , including several instructions to make a computer device (which may be a personal computer, a server or a network device, etc.) execute all or part of the steps of the methods in various embodiments of the present invention.
  • the foregoing storage medium comprises: U disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), mobile hard disk, magnetic disk or optical disk and other media that can store program codes.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Artificial Intelligence (AREA)
  • Computer Hardware Design (AREA)
  • Geometry (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Image Analysis (AREA)

Abstract

The present invention relates to the field of model learning, and in particular to an interpretability method and system for a deep reinforcement learning model in a driverless scene. According to the method and system, a suitable simulation environment and a suitable deep reinforcement learning algorithm are selected to obtain a converged reinforcement learning model by training; a picture captured in a driverless scene is input into the reinforcement learning model, feature division is performed on the picture and quantitative analysis of feature influence is performed, the degrees of influence of features on model decision are calculated, a corresponding difference matrix is obtained, and an improved network model is obtained. Thus, the technical problem in the prior art that the influence of features of a picture on model decision cannot be accurately analyzed is at least solved.

Description

无人驾驶场景下深度强化学习模型的可解释性方法及系统Method and system for interpretability of deep reinforcement learning model in unmanned driving scene 技术领域technical field
本发明属于模型学习领域,尤其涉及一种无人驾驶场景下深度强化学习模型的可解释性方法及系统。The invention belongs to the field of model learning, and in particular relates to an interpretability method and system for a deep reinforcement learning model in an unmanned driving scene.
背景技术Background technique
无人驾驶场景下深度强化学习模型的可解释性技术实现了对无人驾驶场景下的不透明模型的解释,内置了无人驾驶场景下的深度强化学习模型问题和解释问题的解决算法和优化方案,向用户解释并可视化了无人驾驶环境中模型的运行与决策过程中的重要因素。其中深度强化学习模型主要涉及在无人驾驶环境下,为了智能体进行自主决策而进行的各种深度强化学习算法(DeepReinforcementLearning,DRL)的选择。可解释的人工智能(ExplainableAI,XAI)作为一项人工智能新兴领域,主要是对各种AI算法模型进行解释与可视化。而可解释强化学习(ExplainableReinforcementLearning,XRL)则是XAI技术中的一个分支,通过一系列手段对强化学习模型进行解释,其中包括对目前和深度学习相结合的深度强化学习的解释,并以用户可理解的文本或者图片的格式进行可视化,呈现给用户。The interpretability technology of the deep reinforcement learning model in the unmanned driving scenario realizes the interpretation of the opaque model in the unmanned driving scenario, and has built-in deep reinforcement learning model problems and interpretation problems in the unmanned driving scenario. Solving algorithms and optimization schemes , which explains and visualizes to the user the important factors in the model's operation and decision-making process in an unmanned driving environment. Among them, the deep reinforcement learning model mainly involves the selection of various deep reinforcement learning algorithms (Deep Reinforcement Learning, DRL) for the autonomous decision-making of agents in an unmanned driving environment. As an emerging field of artificial intelligence, explainable artificial intelligence (ExplainableAI, XAI) is mainly to explain and visualize various AI algorithm models. Explainable reinforcement learning (Explainable Reinforcement Learning, XRL) is a branch of XAI technology, which explains the reinforcement learning model through a series of means, including the interpretation of the current deep reinforcement learning combined with deep learning, and the user can The understood text or image format is visualized and presented to the user.
DRL是深度学习与强化学习结合的算法,结合了深度学习的感知能力和强化学习的决策能力,并得到了更深层次的发展,包括基于值函数(value-based)的DRL、基于策略(policy-based)的DRL、基于模型的(model-based)DRL、基于分层(hierarchical-based)的DRL等。DRL is an algorithm combining deep learning and reinforcement learning. It combines the perception ability of deep learning and the decision-making ability of reinforcement learning, and has been developed at a deeper level, including value-based DRL, policy-based (policy- based) DRL, model-based DRL, hierarchical-based DRL, etc.
XRL则是可解释强化学习,用该技术对DRL进行解释与可视化,XRL技术分类如下:XRL is interpretable reinforcement learning, which is used to explain and visualize DRL. XRL technologies are classified as follows:
(1)根据提取信息的时间分为:内在解释的(intrinsicinterpretability),内在解释在训练时被构造成内在可解释或本身就是可解释的,比如决策树;事后解释(post-hocinterpretability),事后解释是通过在训练后通过创建第二个更简单的模型或者扰动法等其他操作,从而为原始模型提供解释,替代模型或显著图比较典型。(1) According to the time of extracting information, it is divided into: intrinsic interpretability (intrinsic interpretability), intrinsic interpretation is constructed to be intrinsically interpretable or itself interpretable during training, such as decision trees; post-hoc interpretability (post-hoc interpretability), post-hoc interpretation Alternative models or saliency maps are typical by providing an explanation for the original model after training by creating a second simpler model or other operations such as perturbation.
(2)根据解释的范围分为:全局解释(globalinterpretability)和局部解释(localinterpretability),全局解释整个、一般的模型行为,而局部解释为特定的决策提供解释。在实践中很难实现全局模型的可解释性,尤其是对于超出少数参数的模型。因此局部可解释性可以更容易地应用。解释做出特定决定或单一预测的原因意味着可解释性在局部发生。通常,使用这种可解释性的方式来生成单独的解释,以说明为什么模型为实例做出特定决策的理由。(2) According to the scope of explanation, it is divided into: global interpretability and local interpretability. Global interpretability explains the entire and general model behavior, while local interpretation provides explanations for specific decisions. Interpretability of global models is difficult to achieve in practice, especially for models beyond a few parameters. Thus local interpretability can be more easily applied. Explaining why a particular decision or single prediction was made means that interpretability happens locally. Typically, this approach to interpretability is used to generate individual explanations for why the model made a particular decision for an instance.
深度强化学习模型的可解释性技术是无人驾驶、计算机领域的重点问题。XRL作为XAI的子领域,尚未得到广泛研究。现在对深度强化学习的研究方向是事后解释(post-hocinterpretability),既有全局解释(globalinterpretability),也有局部解释(localinterpretability)。虽然XRL起步较晚,但是目前也有几个典型研究在事后解释(post-hocinterpretability)也有几个典型研究。同时XAI中针对其他的人工智能模型的解释性算法也可以利用到DRL的解释中。Interpretability technology of deep reinforcement learning models is a key issue in the field of unmanned driving and computers. As a subfield of XAI, XRL has not been widely studied. The current research direction of deep reinforcement learning is post-hoc interpretability, both global interpretability and local interpretability. Although XRL started late, there are currently several typical studies on post-hoc interpretability. At the same time, the explanatory algorithms for other artificial intelligence models in XAI can also be used in the interpretation of DRL.
事后解释(post-hocinterpretability):2018年,Greydanus等人在ICML论文上提出显著图法(Saliencymap)这是基于扰动的方法,对输入直接进行扰动。通过对区域性的高斯模糊,对比正常图片与模糊图片经过网络时的差值,模糊区域在图片上进行滑动,从而遍历整张图片,得到多个差值。其中差值大的区域对智能体决策起到重要作用,得出DRL学习的关键区域。但是现有的基于扰动的计算显着性的方法通常会突出显示与代理所采取的动作无关的输入区域。2020年,Nikaash等人在ICLR论文中提出的方法SARFA(特定和相关特征归因)通过平衡两个方面(特殊性和相关性)来生成更加突出的显着性图,这两个方面捕获了不同的显着性需求。第一个记录了摄动对将要解释的动作的相对预期回报的影响。第二部分权衡了不相关的特征,这些特征改变了将要解释的动作以外的动作的相对预期回报。也可以通过训练第二个可解释模型来逼近原来的黑匣子模型,2016年,Ribeiro等人在SIGKDD论文中提出了LIME算法,训练了一个线性可解释的模型来逼近原来的分类网络,从而对卷积(CNN)分类网络进行了解释,同时该方法也用到了对图片进行的扰动。Post-hoc interpretability: In 2018, Greydanus et al. proposed a salient map method (Saliencymap) on the ICML paper. This is a perturbation-based method that directly perturbs the input. Through the regional Gaussian blur, compare the difference between the normal picture and the blurred picture when passing through the network, and slide the blurred area on the picture to traverse the entire picture to obtain multiple differences. Among them, the area with a large difference plays an important role in the decision-making of the agent, and the key area of DRL learning is obtained. But existing perturbation-based methods for computing saliency often highlight regions of the input that are not relevant to the actions taken by the agent. In 2020, the method SARFA (Specific and Relevant Feature Attribution) proposed by Nikaash et al. in the ICLR paper generates more salient saliency maps by balancing two aspects (specificity and relevance), which capture Different salience needs. The first documents the effect of perturbations on the relative expected rewards of actions to be explained. The second part weighs irrelevant features that change the relative expected rewards of actions other than the one to be explained. It is also possible to approach the original black box model by training a second interpretable model. In 2016, Ribeiro et al. proposed the LIME algorithm in the SIGKDD paper, and trained a linear interpretable model to approximate the original classification network. CNN classification network is explained, and the method also uses the perturbation of the image.
现有的关于深度强化学习模型的可解释性技术都是基于扰动技术和训练可解释模型逼近来实现,但是目前提出的方法都有局限性并有极大的改进空间,并且在解释范围不能量化或解释速度低,不能更好地为深度强化学习模型提供合理的解释。现有的算法局限性也不能很好地满足实际应用对解释性算法的需求。Existing interpretability techniques for deep reinforcement learning models are based on perturbation techniques and training interpretable model approximations, but the methods currently proposed have limitations and great room for improvement, and cannot be quantified in the scope of interpretation Or the interpretation speed is low, and it cannot better provide reasonable explanations for deep reinforcement learning models. The limitations of existing algorithms cannot well meet the needs of practical applications for explanatory algorithms.
通过扰动输入进行解释的显著图算法(Saliencymap)需要将输入图片按照一定间隔对图片进行区域高斯模糊,每模糊一次都要将模糊的图片输入网络,将得到的值与原图片进入网络得到的值取差值,从而得到该区域对模型决策的影响程度,这样通过均匀模糊不容易得到图片中特定特征对模型决策的影响,当模糊范围小,时不能覆盖整个特征,只能是该特征某一部分对决策的影响;当模糊范围比较大时,容易覆盖多个特征,不能得到某一个特征对模型决策的影响,不利于精准分析图片各个特征对模型决策的影响;The saliency map algorithm (Saliencymap) explained by disturbing the input needs to perform regional Gaussian blurring on the input image at a certain interval, and input the blurred image into the network every time it is blurred, and the obtained value and the value obtained by entering the original image into the network Take the difference to get the degree of influence of the area on the model decision. In this way, it is not easy to obtain the influence of a specific feature in the picture on the model decision through uniform blurring. When the blur range is small, the entire feature cannot be covered, only a certain part of the feature Impact on decision-making; when the fuzzy range is relatively large, it is easy to cover multiple features, and the impact of a certain feature on model decision-making cannot be obtained, which is not conducive to accurate analysis of the impact of each feature of the image on model decision-making;
而训练简单模型逼近复杂模型的LIME算法,通过使用简单的模型来对复杂的分类网络进行逼近,利用简单的一维线性模型,将输入图片进行一维量化以及扰动后,对原模型进行逼近。最后可以通过查看线性模型的系数大小来对模型进行解释。该方法可以很好地解释输入图片的特征对模型决策的影响。但是LIME一次只会对一个样本进行解释,并且每次都需要新建立一个模型,这种算法虽然比较通用且准确,但是用起来花费时间较长,并且很难将数据利用在网络的更新上,并且在场景变换快、速度要求高的场景中不是非常适用。The LIME algorithm, which trains a simple model to approximate a complex model, uses a simple model to approximate a complex classification network, and uses a simple one-dimensional linear model to perform one-dimensional quantization and disturbance on the input image to approximate the original model. Finally, the model can be explained by looking at the magnitude of the coefficients of the linear model. This method can well explain the influence of the characteristics of the input image on the model's decision. However, LIME can only explain one sample at a time, and needs to build a new model each time. Although this algorithm is more general and accurate, it takes a long time to use, and it is difficult to use the data to update the network. And it is not very suitable for scenes with fast scene changes and high speed requirements.
技术问题technical problem
本发明实施例提供了一种无人驾驶场景下深度强化学习模型的可解释性方法及系统,以至少解决现有技术不能精准分析图片各个特征对模型决策的影响的技术问题。Embodiments of the present invention provide an interpretability method and system for a deep reinforcement learning model in an unmanned driving scene, to at least solve the technical problem that the prior art cannot accurately analyze the influence of each feature of a picture on model decision-making.
技术解决方案technical solution
根据本发明的一实施例,提供了一种无人驾驶场景下深度强化学习模型的可解释性方法,包括以下步骤:According to an embodiment of the present invention, an interpretability method for a deep reinforcement learning model in an unmanned driving scene is provided, comprising the following steps:
选择合适的仿真环境以及适合的深度强化学习算法,通过训练得到收敛的强化学习模型;Select a suitable simulation environment and a suitable deep reinforcement learning algorithm, and obtain a convergent reinforcement learning model through training;
对强化学习模型输入无人驾驶场景下拍摄的图片,对图片进行特征的划分及进行特征影响力的量化分析,计算各个特征对模型决策的影响程度,并得到相应的差值矩阵,得到改进型网络模型。Input the pictures taken in the unmanned driving scene to the reinforcement learning model, divide the features of the pictures and perform quantitative analysis of the influence of the features, calculate the degree of influence of each feature on the decision of the model, and obtain the corresponding difference matrix, and get the improved model network model.
进一步地,对强化学习模型输入无人驾驶场景下拍摄的图片,对图片进行特征的划分及进行特征影响力的量化分析,计算各个特征对模型决策的影响程度,并得到相应的差值矩阵,得到改进型网络模型包括:Further, input the pictures taken in unmanned driving scenarios to the reinforcement learning model, divide the features of the pictures and perform quantitative analysis on the influence of features, calculate the degree of influence of each feature on the model decision, and obtain the corresponding difference matrix, The resulting improved network model includes:
首先通过模型与环境交互得到状态图像,通过超像素分割将图片根据特征分为固定的块数,通过对不规则区域的高斯模糊的方法依次分别模糊得到图像集;First, the state image is obtained through the interaction between the model and the environment, and the image is divided into a fixed number of blocks according to the characteristics through superpixel segmentation, and the image set is obtained by sequentially blurring the Gaussian blur method of the irregular area;
之后将图像集与原图分别输入网络,得到原图与模糊图像的决策值,两者做差,得到差值矩阵;Then input the image set and the original image into the network separately to obtain the decision value of the original image and the blurred image, and make a difference between the two to obtain the difference matrix;
将差值矩阵通过上采样,使得矩阵大小等于输入图像的大小,并将差值矩阵的值乘以预设的倍数叠加到原图中。The difference matrix is up-sampled so that the size of the matrix is equal to the size of the input image, and the value of the difference matrix is multiplied by a preset multiple and superimposed on the original image.
进一步地,选择深度强化学习中的A3C作为无人驾驶中的智能体自主决策的算法。Furthermore, A3C in deep reinforcement learning is selected as the algorithm for the autonomous decision-making of the agent in unmanned driving.
进一步地,无人驾驶环境选择carla仿真环境,选择合适的场景,选择图片作为输入。Further, the unmanned driving environment selects the carla simulation environment, selects a suitable scene, and selects a picture as input.
进一步地,对强化学习模型输入无人驾驶场景下拍摄的图片之前还包括:对无人驾驶场景下拍摄的图片进行预处理。Further, before inputting the pictures taken in the unmanned driving scene to the reinforcement learning model, it also includes: preprocessing the pictures taken in the unmanned driving scene.
进一步地,对无人驾驶场景下拍摄的图片进行预处理包括:Further, preprocessing the pictures taken in unmanned driving scenarios includes:
将输入图片转换成进行解释时所需要的形式:将无人驾驶环境中合适的图像特征找出来并进行分割,利用最小的分割块数来囊括无人驾驶环境中所需要的特征。Convert the input image into the form required for interpretation: find and segment the appropriate image features in the unmanned driving environment, and use the minimum number of segmentation blocks to include the features required in the unmanned driving environment.
进一步地,图像分割将具有相似纹理、颜色、亮度特征的相邻像素构成有视觉意义的不规则像素块,并用少量像素来代替大量像素;其中图像模糊为每一个像素都取周边像素的平均值。Further, image segmentation forms adjacent pixels with similar texture, color, and brightness characteristics into visually meaningful irregular pixel blocks, and replaces a large number of pixels with a small number of pixels; where image blurring is the average value of surrounding pixels for each pixel .
进一步地,使用显著图算法对图片进行特征的划分及进行特征影响力的量化分析。Further, use the saliency map algorithm to divide the features of the picture and perform quantitative analysis of the influence of the features.
进一步地,该方法还包括:Further, the method also includes:
以用户易理解的形式将解释的内容展现给用户。Present the explained content to the user in a form that the user can easily understand.
根据本发明的另一实施例,提供了一种无人驾驶场景下深度强化学习模型的可解释性系统,包括:According to another embodiment of the present invention, an interpretability system for a deep reinforcement learning model in an unmanned driving scenario is provided, including:
网络模型模块,用于选择合适的仿真环境以及适合的深度强化学习算法,通过训练得到收敛的强化学习模型;The network model module is used to select a suitable simulation environment and a suitable deep reinforcement learning algorithm, and obtain a convergent reinforcement learning model through training;
解释性算法模块,用于对强化学习模型输入无人驾驶场景下拍摄的图片,对图片进行特征的划分及进行特征影响力的量化分析,计算各个特征对模型决策的影响程度,并得到相应的差值矩阵,得到改进型网络模型。The explanatory algorithm module is used to input the pictures taken in the unmanned driving scene to the reinforcement learning model, divide the features of the pictures and perform quantitative analysis on the influence of features, calculate the degree of influence of each feature on the model decision, and obtain the corresponding The difference matrix is used to obtain an improved network model.
有益效果Beneficial effect
本发明实施例中的无人驾驶场景下深度强化学习模型的可解释性方法及系统,选择合适的仿真环境以及适合的深度强化学习算法,通过训练得到收敛的强化学习模型,对强化学习模型输入无人驾驶场景下拍摄的图片,对图片进行特征的划分及进行特征影响力的量化分析,计算各个特征对模型决策的影响程度,并得到相应的差值矩阵,得到改进型网络模型。The interpretability method and system of the deep reinforcement learning model in the unmanned driving scene in the embodiment of the present invention, select a suitable simulation environment and a suitable depth reinforcement learning algorithm, obtain a convergent reinforcement learning model through training, and input the reinforcement learning model For the pictures taken in the unmanned driving scene, the features of the pictures are divided and the quantitative analysis of the influence of the features is carried out, the degree of influence of each feature on the model decision is calculated, and the corresponding difference matrix is obtained, and the improved network model is obtained.
附图说明Description of drawings
此处所说明的附图用来提供对本发明的进一步理解,构成本申请的一部分,本发明的示意性实施例及其说明用于解释本发明,并不构成对本发明的不当限定。在附图中:The accompanying drawings described here are used to provide a further understanding of the present invention and constitute a part of the application. The schematic embodiments of the present invention and their descriptions are used to explain the present invention and do not constitute improper limitations to the present invention. In the attached picture:
图1为本发明无人驾驶场景下深度强化学习模型的可解释性方法及系统的整体设计框架图;Fig. 1 is the frame diagram of the overall design of the interpretability method and system of the deep reinforcement learning model in the unmanned driving scene of the present invention;
图2为本发明无人驾驶场景下深度强化学习模型的可解释性方法及系统的工作流程图。Fig. 2 is a working flow chart of the interpretability method and system of the deep reinforcement learning model in the unmanned driving scene of the present invention.
本发明的实施方式Embodiments of the present invention
为了使本技术领域的人员更好地理解本发明方案,下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分的实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都应当属于本发明保护的范围。In order to enable those skilled in the art to better understand the solutions of the present invention, the technical solutions in the embodiments of the present invention will be clearly and completely described below in conjunction with the drawings in the embodiments of the present invention. Obviously, the described embodiments are only It is an embodiment of a part of the present invention, but not all embodiments. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts shall fall within the protection scope of the present invention.
需要说明的是,本发明的说明书和权利要求书及上述附图中的术语“第一”、“第二”等是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。应该理解这样使用的数据在适当情况下可以互换,以便这里描述的本发明的实施例能够以除了在这里图示或描述的那些以外的顺序实施。此外,术语“包括”和“具有”以及他们的任何变形,意图在于覆盖不排他的包含,例如,包含了一系列步骤或单元的过程、方法、系统、产品或设备不必限于清楚地列出的那些步骤或单元,而是可包括没有清楚地列出的或对于这些过程、方法、产品或设备固有的其它步骤或单元。It should be noted that the terms "first" and "second" in the description and claims of the present invention and the above drawings are used to distinguish similar objects, but not necessarily used to describe a specific sequence or sequence. It is to be understood that the data so used are interchangeable under appropriate circumstances such that the embodiments of the invention described herein can be practiced in sequences other than those illustrated or described herein. Furthermore, the terms "comprising" and "having", as well as any variations thereof, are intended to cover a non-exclusive inclusion, for example, a process, method, system, product or device comprising a sequence of steps or elements is not necessarily limited to the expressly listed instead, may include other steps or elements not explicitly listed or inherent to the process, method, product or apparatus.
本发明为一款新颖的XRL算法,针对深度强化学习模型,为其决策的解释与可视化提供的在快速与准确的解决方案;对输入图片中确定的特征对模型决策的影响进行量化;为提高速度,减少每张图片的超像素块数;为设计出通用的适应各种实际场景的可解释性算法,在设计过程中不依赖某种特定的模型,所以本发明提出的XRL是不依赖于某一模型(Model-free),并且适应实际出现的场景问题,XRL还应具备一定的灵活性和扩展性,以便能适应特征数不同的各个场景。The invention is a novel XRL algorithm, aiming at the deep reinforcement learning model, providing a fast and accurate solution for the interpretation and visualization of its decision-making; it quantifies the influence of the features determined in the input picture on the decision-making of the model; in order to improve Speed, reducing the number of superpixel blocks of each picture; in order to design a general interpretability algorithm that adapts to various actual scenarios, it does not depend on a specific model in the design process, so the XRL proposed by the present invention does not depend on A certain model (Model-free), and adapt to the actual scene problems, XRL should also have a certain degree of flexibility and scalability, so as to be able to adapt to various scenes with different numbers of features.
本发明要解决的问题是利用超像素(Superpixel)分割、高斯模糊(GaussianBlur)、显著性图(Saliencymap)、深度强化学习(Deepreinforcementlearning)等算法解决无人驾驶场景下深度强化学习模型不可知问题及其延伸问题,通过解释性算法使用户了解决策过程中的有利因素以及不利因素,并通过良好的人机交互形式展现给用户。为用户呈现智能体的决策依据,增加用户对无人驾驶模型的信任度。The problem to be solved by the present invention is to use algorithms such as Superpixel segmentation, Gaussian Blur, Saliencymap, and Deep reinforcement learning to solve the unknowable problem of deep reinforcement learning models in unmanned driving scenarios and Its extended problem enables users to understand the favorable and unfavorable factors in the decision-making process through explanatory algorithms, and presents them to users through good human-computer interaction. Present the decision-making basis of the intelligent body to the user, and increase the user's trust in the unmanned driving model.
本发明技术方案的基本内容包括如下:The basic contents of the technical solution of the present invention include as follows:
1.设计通用的解释性算法,对深度强化学习每个系列算法都适用;1. Design a general-purpose explanatory algorithm, which is applicable to every series of algorithms in deep reinforcement learning;
2.针对解释性算法过程中的输入图像预处理的设计;2. For the design of input image preprocessing in the process of interpreting algorithms;
3.通过网络与预处理图像进行对决策影响的解释;3. Interpretation of the decision-making impact through the network and pre-processed images;
4.将解释内容通过可视化技术对用户展现。4. Present the explanation content to the user through visualization technology.
无人驾驶场景下深度强化学习模型的可解释性方法及系统整体设计框架由3部分组成:网络模型部分、解释性算法部分、网络改进部分,如图1所示。The interpretability method and the overall system design framework of the deep reinforcement learning model in the unmanned driving scenario are composed of three parts: the network model part, the explanatory algorithm part, and the network improvement part, as shown in Figure 1.
(1)网络模型部分(1) Network model part
网络模型部分包含了在无人驾驶场景下的深度强化学习算法的选择、场景的设计与模型训练。The network model part includes the selection of deep reinforcement learning algorithms, scene design and model training in unmanned driving scenarios.
本发明需要事先选择合适的仿真环境以及适合的深度强化学习算法,通过比较,选择深度强化学习中的A3C(Asynchronousadvantageactor-critic)作为无人驾驶中的智能体自主决策的算法。无人驾驶环境选择carla仿真环境,选择合适的场景,选择图片作为输入:然后通过训练,最后得到收敛的强化学习模型,此时就得到了本发明所需的接下来进行解释的模型。The present invention needs to select a suitable simulation environment and a suitable deep reinforcement learning algorithm in advance. Through comparison, A3C (Asynchronous advantageous actor-critic) in deep reinforcement learning is selected as the algorithm for the autonomous decision-making of the agent in unmanned driving. The unmanned driving environment selects the carla simulation environment, selects a suitable scene, and selects a picture as input: then through training, a convergent reinforcement learning model is finally obtained, and at this time, the model that is explained next is obtained.
(2)解释性算法部分(2) Explanatory algorithm part
解释性算法部分包括图片预处理、显著图运算(求解差值矩阵)、可视化等模块。The explanatory algorithm part includes image preprocessing, saliency map operation (solution of difference matrix), visualization and other modules.
其中,图片预处理的结果将有利于显著图算法(Saliencymap)模块的运行以及特征的划分,有利于解释性算法对特征影响力的量化分析;显著图算法(Saliencymap)模块则将计算各个特征对模型决策的影响程度,并得到相应的差值矩阵,从而得到模型的决策过程中的重要因素;可视化模块通过以用户易理解的形式将解释的内容展现给用户。Among them, the results of image preprocessing will be beneficial to the operation of the Saliencymap module and the division of features, and will be conducive to the quantitative analysis of the influence of explanatory algorithms on features; the Saliencymap module will calculate the impact of each feature on The degree of influence of model decision-making, and obtain the corresponding difference matrix, so as to obtain the important factors in the decision-making process of the model; the visualization module presents the explained content to users in a form that users can easily understand.
(3)网络改进部分(3) Network improvement part
通过解释的信息,加强有用的信息,隔离不重要的信息,使得网络效果更好,并进一步验证解释的效果。Through the explained information, strengthen the useful information, isolate the unimportant information, make the network effect better, and further verify the effect of the explanation.
无人驾驶场景下深度强化学习模型的可解释性方法及系统需要满足三项基本要求:Interpretability methods and systems for deep reinforcement learning models in unmanned driving scenarios need to meet three basic requirements:
(1)无人驾驶场景。场景尽量丰富,接近真实情况,并得到收敛的深度强化学习模型。(1) Unmanned driving scene. The scene is as rich as possible, close to the real situation, and a convergent deep reinforcement learning model is obtained.
(2)解释性算法,预处理部分尽可能使得图片分出的特征数量合适。(2) Interpretive algorithm, the preprocessing part tries to make the number of features separated from the picture appropriate.
(3)显著图的区域要尽量收敛,不要过于分散。(3) The area of the saliency map should be as convergent as possible, and not too scattered.
基于此三项要求,本发明设计了无人驾驶场景下深度强化学习模型的可解释性方法及系统的工作流程,如图2所示。Based on these three requirements, the present invention designs the interpretability method and system workflow of the deep reinforcement learning model in the unmanned driving scene, as shown in FIG. 2 .
当本发明得到所需要的模型后,便要开始对模型进行解释,首先通过模型与环境交互得到状态图像,通过超像素(Superpixel)分割将图片根据特征分为固定的块数,之后通过对不规则区域的高斯模糊(GaussianBlur)的方法依次分别模糊得到图像集。之后将图像集与原图分别输入网络,这样就得到了原图与模糊图像的决策值,两者做差,得到差值矩阵。将差值矩阵通过上采样,使得矩阵大小等于输入图像的大小,并将差值矩阵的值乘以一定的倍数叠加到原图中,从而以显著图的形式展现给用户。之后得到显著性增强有意义部分的区域,得到改进网络模型的目的。When the present invention obtains the required model, it starts to interpret the model. First, the state image is obtained through the interaction between the model and the environment, and the image is divided into a fixed number of blocks according to the characteristics through superpixel segmentation. The method of Gaussian Blur (GaussianBlur) in the regular area blurs the image set separately in turn. Afterwards, the image set and the original image are input into the network separately, so that the decision value of the original image and the blurred image are obtained, and the difference between the two is obtained to obtain the difference matrix. The difference matrix is up-sampled so that the size of the matrix is equal to the size of the input image, and the value of the difference matrix is multiplied by a certain multiple and superimposed on the original image, so as to be displayed to the user in the form of a saliency map. Afterwards, the area of the meaningful part is significantly enhanced, and the purpose of improving the network model is obtained.
其中XRL的图像预处理问题,是XRL进行分析的前提,将输入图片转换成本发明进行解释时所需要的形式,可以描述为:将无人驾驶环境中合适的图像特征找出来并进行分割,可以将一些不重要或者说比较小的特征不分割,利用最小的分割块数来囊括无人驾驶环境中所需要的特征,从而大大减少耗费的时间,从而达到了所效果。本发明使用的基于超像素分割以及高斯模糊预处理算法,很好地实现了对输入图像主要特征的分割以及不规则特征的高斯模糊,是较好的可用于深度强化学习模型解释的预处理方法。Among them, the image preprocessing problem of XRL is the premise of XRL analysis. Converting the input image into the form required for the explanation of the present invention can be described as: find out the appropriate image features in the unmanned driving environment and segment them, which can be Some unimportant or relatively small features are not divided, and the minimum number of segmentation blocks is used to include the features required in the unmanned driving environment, thereby greatly reducing the time spent and achieving the desired effect. The preprocessing algorithm based on superpixel segmentation and Gaussian blur used in the present invention can well realize the segmentation of the main features of the input image and the Gaussian blur of irregular features, and is a better preprocessing method that can be used for deep reinforcement learning model interpretation .
图像预处理遵循传统的图像处理算法过程:图像分割将具有相似纹理、颜色、亮度等特征的相邻像素构成的有一定视觉意义的不规则像素块,并用少量像素来代替大量像素。图像模糊可以理解为每一个像素都取周边像素的平均值。Image preprocessing follows the traditional image processing algorithm process: image segmentation forms irregular pixel blocks with certain visual significance from adjacent pixels with similar texture, color, brightness and other characteristics, and replaces a large number of pixels with a small number of pixels. Image blurring can be understood as taking the average value of surrounding pixels for each pixel.
该解释性算法特点是在每个解释过程中对特征的重要性进行了量化,并分别对决策有正负两个,因此,将只描述与传统解释性算法不同的部分:The feature of this explanatory algorithm is that the importance of features is quantified in each interpretation process, and there are positive and negative two for decision-making respectively. Therefore, only the parts that are different from traditional explanatory algorithms will be described:
不规则区域的模糊:图片的模糊部分是消除部分特征,使得图像与原图在特征上有所区别,方便后续进行与原图得到的策略进行对比,关于模糊的地方需要注意的是模糊处尽量与未模糊处平滑过渡,以免模糊部分与未模糊之间过于明显的边界对模型的决策造成影响。Blur in irregular areas: The blurred part of the picture is to eliminate some features, so that the image is different from the original image in terms of features, which is convenient for subsequent comparison with the strategy obtained from the original image. Regarding the blurred parts, it should be noted that the blurred parts should be as far as possible Smooth transition with the unblurred part, so as not to affect the model's decision-making due to the obvious boundary between the blurred part and the unblurred part.
特征解释:将图片按照特征区域进行处理,得到了特征区域对模型决策影响程度并进行了量化,通过归一化处理,使得各个特征有了对比性,并且计算每个特征对决策的正向的影响和负向的影响,得到的数据有利于接下来对模型的更新。Feature interpretation: Process the picture according to the feature area, and obtain and quantify the influence of the feature area on the model decision-making. Through normalization processing, each feature is contrasted, and the positive contribution of each feature to the decision-making is calculated. Influence and negative influence, the obtained data is conducive to the next update of the model.
根据解释对模型进行改进:得到的解释,可以知道模型正常运行过程中,有利因素是哪些,不利因素是哪些。同时当模型出错时,也可以了解到具体是输入图像的哪一个特征造成了系统决策失败。通过这些信息从而可以对模型进行改进。Improve the model based on the explanation: the obtained explanation can know which are the favorable factors and which are the unfavorable factors during the normal operation of the model. At the same time, when the model makes an error, it can also know which feature of the input image caused the failure of the system's decision-making. This information can be used to improve the model.
本发明的关键点和欲保护点至少在于:Key point of the present invention and want to protect point are at least:
1.XRL整体设计方案;1. The overall design of XRL;
2.图像预处理方法可视化;2. Visualization of image preprocessing methods;
3.基于可解释性的深度强化学习模型改进算法。3. Improved algorithms for deep reinforcement learning models based on interpretability.
本发明针对无人驾驶场景下深度强化学习模型的可解释性场景,以解决该场景下DRL模型不透明问题,对其决策进行解释,并提供可视化得人机交互界面,在一定程度上解释了深度强化学习模型的可解释性问题,增加用户的信任度问题,同时为模型的改进提供了依据。本发明主要体现了以下优点:The present invention aims at the interpretability scene of the deep reinforcement learning model in the unmanned driving scene to solve the opaque problem of the DRL model in the scene, explain its decision-making, and provide a visualized human-computer interaction interface, which explains the depth to a certain extent. The interpretability of the reinforcement learning model increases the user's trust, and at the same time provides a basis for the improvement of the model. The present invention mainly embodies the following advantages:
对输入图片特征对模型决策的影响进行量化以及对比,得到重要特征;Quantify and compare the impact of input image features on model decision-making to obtain important features;
解释系统以为由模块化组成,具有灵活性高,扩展性好等特点;Explain that the system is composed of modules and has the characteristics of high flexibility and good scalability;
在图像预处理阶段可以将图像特征进行不规则区域进行平滑模糊化,使得模糊区域与未模糊区域平滑交接;In the image preprocessing stage, image features can be smoothed and blurred in irregular areas, so that the blurred area and the unblurred area can be smoothly handed over;
可以通过解释的内容,对深度强化学习模型进行进一步的改进,这是目前解释系统和模型改进系统所未涉及的部分。Deep reinforcement learning models can be further improved through the content of the explanation, which is currently not covered by explanation systems and model improvement systems.
以无人驾驶环境为实验平台,验证了XRL及模型改进方案,验证了XRL算法,并通过仿真平台可视化算法使得用户了解了模型决策依据,并根据该依据进行模型的改进。Taking the unmanned driving environment as the experimental platform, the XRL and model improvement schemes are verified, and the XRL algorithm is verified. Through the visualization algorithm of the simulation platform, the user understands the decision-making basis of the model, and improves the model based on this basis.
本发明替换方案至少为:The alternative scheme of the present invention is at least:
1.XRL系统具有可扩展性,可任意组合扩展模块,以便符合客户需求。例如增加或者变更图片预处理过程、改变图片的扰动方式、改变差值计算方法等。1. The XRL system is scalable, and expansion modules can be combined arbitrarily to meet customer needs. For example, adding or changing the image preprocessing process, changing the perturbation method of the image, changing the difference calculation method, etc.
2.提出通过解释性改进模型,增强正向特征,抑制负面特征,实现模型改进。2. Propose to improve the model through explanatory, enhance positive features, suppress negative features, and achieve model improvement.
上述本发明实施例序号仅仅为了描述,不代表实施例的优劣。The serial numbers of the above embodiments of the present invention are for description only, and do not represent the advantages and disadvantages of the embodiments.
在本发明的上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述的部分,可以参见其他实施例的相关描述。In the above-mentioned embodiments of the present invention, the descriptions of each embodiment have their own emphases, and for parts not described in detail in a certain embodiment, reference may be made to relevant descriptions of other embodiments.
在本申请所提供的几个实施例中,应该理解到,所揭露的技术内容,可通过其它的方式实现。其中,以上所描述的系统实施例仅仅是示意性的,例如单元的划分,可以为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,单元或模块的间接耦合或通信连接,可以是电性或其它的形式。In the several embodiments provided in this application, it should be understood that the disclosed technical content can be realized in other ways. Wherein, the system embodiments described above are only illustrative, for example, the division of units can be divided into a logical function, and there may be other division methods in actual implementation, for example, multiple units or components can be combined or integrated into Another system, or some features may be ignored, or not implemented. In another point, the mutual coupling or direct coupling or communication connection shown or discussed may be through some interfaces, and the indirect coupling or communication connection of units or modules may be in electrical or other forms.
作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。A unit described as a separate component may or may not be physically separated, and a component shown as a unit may or may not be a physical unit, that is, it may be located in one place, or may be distributed over multiple units. Part or all of the units can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
另外,在本发明各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。In addition, each functional unit in each embodiment of the present invention may be integrated into one processing unit, each unit may exist separately physically, or two or more units may be integrated into one unit. The above-mentioned integrated units can be implemented in the form of hardware or in the form of software functional units.
集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本发明的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的全部或部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可为个人计算机、服务器或者网络设备等)执行本发明各个实施例方法的全部或部分步骤。而前述的存储介质包括:U盘、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、移动硬盘、磁碟或者光盘等各种可以存储程序代码的介质。If the integrated unit is realized in the form of a software function unit and sold or used as an independent product, it can be stored in a computer-readable storage medium. Based on such an understanding, the essence of the technical solution of the present invention or the part that contributes to the prior art or all or part of the technical solution can be embodied in the form of a software product, and the computer software product is stored in a storage medium , including several instructions to make a computer device (which may be a personal computer, a server or a network device, etc.) execute all or part of the steps of the methods in various embodiments of the present invention. And the foregoing storage medium comprises: U disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), mobile hard disk, magnetic disk or optical disk and other media that can store program codes.
以上所述仅是本发明的优选实施方式,应当指出,对于本技术领域的普通技术人员来说,在不脱离本发明原理的前提下,还可以做出若干改进和润饰,这些改进和润饰也应视为本发明的保护范围。The above is only a preferred embodiment of the present invention, it should be pointed out that, for those of ordinary skill in the art, without departing from the principle of the present invention, some improvements and modifications can also be made, and these improvements and modifications can also be made. It should be regarded as the protection scope of the present invention.

Claims (10)

  1. 一种无人驾驶场景下深度强化学习模型的可解释性方法,其特征在于,包括以下步骤:An interpretability method for a deep reinforcement learning model in an unmanned driving scene, characterized in that it comprises the following steps:
    选择合适的仿真环境以及适合的深度强化学习算法,通过训练得到收敛的强化学习模型;Select a suitable simulation environment and a suitable deep reinforcement learning algorithm, and obtain a convergent reinforcement learning model through training;
    对强化学习模型输入无人驾驶场景下拍摄的图片,对图片进行特征的划分及进行特征影响力的量化分析,计算各个特征对模型决策的影响程度,并得到相应的差值矩阵,得到改进型网络模型。Input the pictures taken in the unmanned driving scene to the reinforcement learning model, divide the features of the pictures and perform quantitative analysis of the influence of the features, calculate the degree of influence of each feature on the decision of the model, and obtain the corresponding difference matrix, and get the improved model network model.
  2. 根据权利要求1所述的无人驾驶场景下深度强化学习模型的可解释性方法,其特征在于,对强化学习模型输入无人驾驶场景下拍摄的图片,对图片进行特征的划分及进行特征影响力的量化分析,计算各个特征对模型决策的影响程度,并得到相应的差值矩阵,得到改进型网络模型包括:The interpretability method of the deep reinforcement learning model under the unmanned driving scene according to claim 1, characterized in that, input the picture taken under the unmanned driving scene to the reinforcement learning model, and divide the features of the picture and perform feature influence Quantitative analysis of power, calculate the degree of influence of each feature on model decision-making, and obtain the corresponding difference matrix, the improved network model includes:
    首先通过模型与环境交互得到状态图像,通过超像素分割将图片根据特征分为固定的块数,通过对不规则区域的高斯模糊的方法依次分别模糊得到图像集;First, the state image is obtained through the interaction between the model and the environment, and the image is divided into a fixed number of blocks according to the characteristics through superpixel segmentation, and the image set is obtained by sequentially blurring the Gaussian blur method of the irregular area;
    之后将图像集与原图分别输入网络,得到原图与模糊图像的决策值,两者做差,得到差值矩阵;Then input the image set and the original image into the network separately to obtain the decision value of the original image and the blurred image, and make a difference between the two to obtain the difference matrix;
    将差值矩阵通过上采样,使得矩阵大小等于输入图像的大小,并将差值矩阵的值乘以预设的倍数叠加到原图中。The difference matrix is up-sampled so that the size of the matrix is equal to the size of the input image, and the value of the difference matrix is multiplied by a preset multiple and superimposed on the original image.
  3. 根据权利要求1所述的无人驾驶场景下深度强化学习模型的可解释性方法,其特征在于,选择深度强化学习中的A3C作为无人驾驶中的智能体自主决策的算法。The interpretability method of a deep reinforcement learning model in an unmanned driving scene according to claim 1, wherein A3C in deep reinforcement learning is selected as an algorithm for autonomous decision-making of an agent in unmanned driving.
  4. 根据权利要求1所述的无人驾驶场景下深度强化学习模型的可解释性方法,其特征在于,无人驾驶环境选择carla仿真环境,选择合适的场景,选择图片作为输入。The interpretability method of the deep reinforcement learning model under the unmanned driving scene according to claim 1, characterized in that, the unmanned driving environment selects a carla simulation environment, selects a suitable scene, and selects a picture as input.
  5. 根据权利要求1所述的无人驾驶场景下深度强化学习模型的可解释性方法,其特征在于,对强化学习模型输入无人驾驶场景下拍摄的图片之前还包括:对无人驾驶场景下拍摄的图片进行预处理。The interpretability method of the deep reinforcement learning model under the unmanned driving scene according to claim 1, wherein, before inputting the picture taken under the unmanned driving scene to the reinforcement learning model, it also includes: taking pictures under the unmanned driving scene images are preprocessed.
  6. 根据权利要求5所述的无人驾驶场景下深度强化学习模型的可解释性方法,其特征在于,对无人驾驶场景下拍摄的图片进行预处理包括:The interpretability method of the deep reinforcement learning model under the unmanned driving scene according to claim 5, wherein the preprocessing of the pictures taken under the unmanned driving scene comprises:
    将输入图片转换成进行解释时所需要的形式:将无人驾驶环境中合适的图像特征找出来并进行分割,利用最小的分割块数来囊括无人驾驶环境中所需要的特征。Convert the input image into the form required for interpretation: find and segment the appropriate image features in the unmanned driving environment, and use the minimum number of segmentation blocks to include the features required in the unmanned driving environment.
  7. 根据权利要求6所述的无人驾驶场景下深度强化学习模型的可解释性方法,其特征在于,图像分割将具有相似纹理、颜色、亮度特征的相邻像素构成有视觉意义的不规则像素块,并用少量像素来代替大量像素;其中图像模糊为每一个像素都取周边像素的平均值。According to claim 6, the interpretability method of deep reinforcement learning model under unmanned driving scene is characterized in that, image segmentation forms adjacent pixels with similar texture, color and brightness characteristics into irregular pixel blocks with visual significance , and replace a large number of pixels with a small number of pixels; where image blurring takes the average value of surrounding pixels for each pixel.
  8. 根据权利要求1所述的无人驾驶场景下深度强化学习模型的可解释性方法,其特征在于,使用显著图算法对图片进行特征的划分及进行特征影响力的量化分析。The interpretability method of a deep reinforcement learning model in an unmanned driving scene according to claim 1, wherein a saliency map algorithm is used to divide the features of the picture and perform a quantitative analysis of the influence of the features.
  9. 根据权利要求1所述的无人驾驶场景下深度强化学习模型的可解释性方法,其特征在于,所述方法还包括:The interpretability method of the deep reinforcement learning model under the unmanned driving scene according to claim 1, it is characterized in that, described method also comprises:
    以用户易理解的形式将解释的内容展现给用户。Present the explained content to the user in a form that the user can easily understand.
  10. 一种无人驾驶场景下深度强化学习模型的可解释性系统,其特征在于,包括:An interpretability system for a deep reinforcement learning model in an unmanned driving scenario, characterized in that it includes:
    网络模型模块,用于选择合适的仿真环境以及适合的深度强化学习算法,通过训练得到收敛的强化学习模型;The network model module is used to select a suitable simulation environment and a suitable deep reinforcement learning algorithm, and obtain a convergent reinforcement learning model through training;
    解释性算法模块,用于对强化学习模型输入无人驾驶场景下拍摄的图片,对图片进行特征的划分及进行特征影响力的量化分析,计算各个特征对模型决策的影响程度,并得到相应的差值矩阵,得到改进型网络模型。The explanatory algorithm module is used to input the pictures taken in the unmanned driving scene to the reinforcement learning model, divide the features of the pictures and perform quantitative analysis on the influence of features, calculate the degree of influence of each feature on the model decision, and obtain the corresponding The difference matrix is used to obtain an improved network model.
PCT/CN2022/137511 2021-12-14 2022-12-08 Interpretability method and system for deep reinforcement learning model in driverless scene WO2023109640A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202111527231.4A CN114330109A (en) 2021-12-14 2021-12-14 Interpretability method and system of deep reinforcement learning model under unmanned scene
CN202111527231.4 2021-12-14

Publications (1)

Publication Number Publication Date
WO2023109640A1 true WO2023109640A1 (en) 2023-06-22

Family

ID=81051066

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/137511 WO2023109640A1 (en) 2021-12-14 2022-12-08 Interpretability method and system for deep reinforcement learning model in driverless scene

Country Status (2)

Country Link
CN (1) CN114330109A (en)
WO (1) WO2023109640A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116704208A (en) * 2023-08-04 2023-09-05 南京理工大学 Local interpretable method based on characteristic relation

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114330109A (en) * 2021-12-14 2022-04-12 深圳先进技术研究院 Interpretability method and system of deep reinforcement learning model under unmanned scene

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110221611A (en) * 2019-06-11 2019-09-10 北京三快在线科技有限公司 A kind of Trajectory Tracking Control method, apparatus and automatic driving vehicle
CN112590792A (en) * 2020-12-18 2021-04-02 的卢技术有限公司 Vehicle convergence control method based on deep reinforcement learning algorithm
CN113139024A (en) * 2021-05-10 2021-07-20 杭州电子科技大学 Visual analysis method for MADDPG multi-agent reinforcement learning model
CN113353102A (en) * 2021-07-08 2021-09-07 重庆大学 Unprotected left-turn driving control method based on deep reinforcement learning
US20210312282A1 (en) * 2020-04-01 2021-10-07 Royal Bank Of Canada System and method for facilitating explainability in reinforcement machine learning
CN114330109A (en) * 2021-12-14 2022-04-12 深圳先进技术研究院 Interpretability method and system of deep reinforcement learning model under unmanned scene

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110221611A (en) * 2019-06-11 2019-09-10 北京三快在线科技有限公司 A kind of Trajectory Tracking Control method, apparatus and automatic driving vehicle
US20210312282A1 (en) * 2020-04-01 2021-10-07 Royal Bank Of Canada System and method for facilitating explainability in reinforcement machine learning
CN112590792A (en) * 2020-12-18 2021-04-02 的卢技术有限公司 Vehicle convergence control method based on deep reinforcement learning algorithm
CN113139024A (en) * 2021-05-10 2021-07-20 杭州电子科技大学 Visual analysis method for MADDPG multi-agent reinforcement learning model
CN113353102A (en) * 2021-07-08 2021-09-07 重庆大学 Unprotected left-turn driving control method based on deep reinforcement learning
CN114330109A (en) * 2021-12-14 2022-04-12 深圳先进技术研究院 Interpretability method and system of deep reinforcement learning model under unmanned scene

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116704208A (en) * 2023-08-04 2023-09-05 南京理工大学 Local interpretable method based on characteristic relation
CN116704208B (en) * 2023-08-04 2023-10-20 南京理工大学 Local interpretable method based on characteristic relation

Also Published As

Publication number Publication date
CN114330109A (en) 2022-04-12

Similar Documents

Publication Publication Date Title
CN111292337B (en) Image background replacement method, device, equipment and storage medium
US11967151B2 (en) Video classification method and apparatus, model training method and apparatus, device, and storage medium
WO2023109640A1 (en) Interpretability method and system for deep reinforcement learning model in driverless scene
CN109741318B (en) Real-time detection method of single-stage multi-scale specific target based on effective receptive field
Li et al. Cross-layer feature pyramid network for salient object detection
CN109598279B (en) Zero sample learning method based on self-coding countermeasure generation network
CN108459785A (en) A kind of video multi-scale visualization method and exchange method
CN112001939B (en) Image foreground segmentation algorithm based on edge knowledge conversion
Miao et al. Balanced single-shot object detection using cross-context attention-guided network
CN113269089B (en) Real-time gesture recognition method and system based on deep learning
CN112257665A (en) Image content recognition method, image recognition model training method, and medium
CN111402126A (en) Video super-resolution method and system based on blocks
CN110390327A (en) Foreground extracting method, device, computer equipment and storage medium
CN113850135A (en) Dynamic gesture recognition method and system based on time shift frame
CN115761574A (en) Weak surveillance video target segmentation method and device based on frame labeling
Fang et al. M2RNet: Multi-modal and multi-scale refined network for RGB-D salient object detection
Liu et al. Toward visual quality enhancement of dehazing effect with improved Cycle-GAN
Guo et al. Saliency detection on sampled images for tag ranking
Zhang et al. Multiple classifiers fusion for facial expression recognition
Dai et al. Feature extraction and fusion network for salient object detection
Horita et al. SSA-GAN: End-to-end time-lapse video generation with spatial self-attention
JP7265686B1 (en) Information processing device, information processing method and information processing program
JP7265690B2 (en) Information processing device, information processing method and information processing program
CN117793151B (en) Distributed network security monitoring system and method based on cloud computing
CN116756232B (en) Data visualization method and system for intelligently generating dynamic network diagram

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22906388

Country of ref document: EP

Kind code of ref document: A1