CN107544516A - Automated driving system and method based on relative entropy depth against intensified learning - Google Patents
Automated driving system and method based on relative entropy depth against intensified learning Download PDFInfo
- Publication number
- CN107544516A CN107544516A CN201710940590.XA CN201710940590A CN107544516A CN 107544516 A CN107544516 A CN 107544516A CN 201710940590 A CN201710940590 A CN 201710940590A CN 107544516 A CN107544516 A CN 107544516A
- Authority
- CN
- China
- Prior art keywords
- driving
- road information
- strategy
- relative entropy
- client
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
Landscapes
- Engineering & Computer Science (AREA)
- Aviation & Aerospace Engineering (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Automation & Control Theory (AREA)
- Traffic Control Systems (AREA)
- Control Of Driving Devices And Active Controlling Of Vehicle (AREA)
Abstract
本发明涉及一种基于相对熵深度逆强化学习的自动驾驶系统,包括:(1)客户端:显示驾驶策略;(2)驾驶基础数据采集子系统:采集道路信息;(3)存储模块:与客户端及驾驶基础数据采集子系统连接并存储驾驶基础数据采集子系统所采集到的道路信息;其中,驾驶基础数据采集子系统采集道路信息并将所述道路信息传输给客户端及存储模块,存储模块接收道路信息,并将持续的一段道路信息存储为历史轨迹,根据历史轨迹进行分析计算模拟出驾驶策略,存储模块将所述驾驶策略传输至客户端以供用户选择,客户端接收道路信息并根据用户选择实施自动驾驶。本发明系统采用相对熵的深度逆强化学习算法实现无模型下自动驾驶。
The invention relates to an automatic driving system based on relative entropy deep inverse reinforcement learning, including: (1) client: display driving strategy; (2) driving basic data collection subsystem: collect road information; (3) storage module: with The client terminal and the driving basic data acquisition subsystem are connected to and store the road information collected by the driving basic data acquisition subsystem; wherein, the driving basic data acquisition subsystem collects road information and transmits the road information to the client terminal and the storage module, The storage module receives the road information, and stores a continuous section of road information as a historical track, analyzes and calculates the driving strategy based on the historical track, and the storage module transmits the driving strategy to the client for user selection, and the client receives the road information And implement automatic driving according to the user's choice. The system of the present invention adopts a deep inverse reinforcement learning algorithm of relative entropy to realize automatic driving without models.
Description
技术领域technical field
本发明涉及一种基于相对熵深度逆强化学习的自动驾驶系统及方法,属于自动驾驶技术领域。The invention relates to an automatic driving system and method based on relative entropy deep inverse reinforcement learning, belonging to the technical field of automatic driving.
背景技术Background technique
随着我国汽车持有量的增加,道路交通拥堵现象越来越严重,每年发生的交通事故也在不断上升,为了更好的解决这一问题,研究和开发汽车自动驾驶系统很有必要。且随着人们对生活质量追求的提升,人们希望从疲劳的驾驶活动中得到解放,自动驾驶技术应运而生。With the increase of the number of cars in our country, road traffic congestion is becoming more and more serious, and the annual traffic accidents are also rising. In order to better solve this problem, it is necessary to research and develop auto-driving systems. And with the improvement of people's pursuit of quality of life, people hope to be liberated from tiring driving activities, and automatic driving technology has emerged as the times require.
现有的一种汽车自动驾驶系统是由装在驾驶室的摄像机和图像识别系统辨别驾驶环境,然后由车载主控计算机、GPS定位系统和路径规划软件根据预先存好的道路地图等信息对车辆进行导航,在车辆的当前位置和目的地之间规划出合理的行驶路径将车辆导向目的地。An existing auto-driving system for a vehicle uses a camera installed in the driver's cab and an image recognition system to identify the driving environment, and then the vehicle's main control computer, GPS positioning system, and path planning software provide information on the vehicle based on pre-stored road maps and other information. Carry out navigation, plan a reasonable driving path between the current position of the vehicle and the destination, and guide the vehicle to the destination.
上述汽车自动驾驶系统中,由于道路地图是预存于车辆内,其数据的更新依赖于驾驶员的人工操作,更新频率不能够保证,并且,即使驾驶员能够做到及时更新,也可能由于现有资源里没有关于道路的最新信息而使得最终得到的资料不能够反应当下的道路情况,最终造成行车路线不合理,导航准确率不高,给行车带来不便。并且,目前在自动驾驶技术领域的大部分汽车自动驾驶系统还需要人工进行干预,并不能达到完全的自动驾驶的地步。In the above autopilot system, since the road map is pre-stored in the vehicle, the update of its data depends on the manual operation of the driver, and the update frequency cannot be guaranteed. There is no latest information about roads in the resources, so the final data cannot reflect the current road conditions, which eventually leads to unreasonable driving routes and low navigation accuracy, which brings inconvenience to driving. Moreover, most of the current autopilot systems in the field of autopilot technology still require manual intervention, and cannot achieve complete autopilot.
发明内容Contents of the invention
本发明的目的在于提供一种基于相对熵深度逆强化学习的自动驾驶系统及方法,利用深度神经网络结构并输入用户驾驶员的历史驾驶轨迹信息,获取多种代表个性驾驶习惯的驾驶策略,通过这些驾驶策略进行个性、智能的自动驾驶。The purpose of the present invention is to provide an automatic driving system and method based on relative entropy deep inverse reinforcement learning, which uses a deep neural network structure and inputs the historical driving track information of the user driver to obtain a variety of driving strategies representing individual driving habits. These driving strategies carry out personalized and intelligent automatic driving.
为达到上述目的,本发明提供如下技术方案:一种基于相对熵深度逆强化学习的自动驾驶系统,所述系统包括:In order to achieve the above object, the present invention provides the following technical solution: an automatic driving system based on relative entropy deep inverse reinforcement learning, the system includes:
客户端:显示驾驶策略;Client: display driving strategy;
驾驶基础数据采集子系统:采集道路信息;Driving basic data acquisition subsystem: collect road information;
存储模块:与所述客户端及驾驶基础数据采集子系统连接并存储所述驾驶基础数据采集子系统所采集到的道路信息;Storage module: connected with the client and the driving basic data collection subsystem and storing the road information collected by the driving basic data collection subsystem;
其中,所述驾驶基础数据采集子系统采集道路信息并将所述道路信息传输给所述客户端及存储模块,所述存储模块接收所述道路信息,并将持续的一段道路信息存储为历史轨迹,根据所述历史轨迹进行分析计算模拟出驾驶策略,所述存储模块将所述驾驶策略传输至客户端以供用户选择,所述客户端接受并根据所述道路信息和用户个性选择的所述驾驶策略实施自动驾驶。Wherein, the driving basic data collection subsystem collects road information and transmits the road information to the client and the storage module, and the storage module receives the road information and stores a continuous section of road information as a historical track , analyze and calculate the driving strategy according to the historical trajectory, the storage module transmits the driving strategy to the client for selection by the user, and the client accepts and selects the driving strategy based on the road information and user personality Driving strategy implements autonomous driving.
进一步地,所述存储模块包括用于存储历史驾驶轨迹的驾驶轨迹库、根据驾驶轨迹及驾驶习惯计算并模拟出驾驶策略的轨迹信息处理子系统及存储驾驶策略的驾驶策略库;所述驾驶轨迹库将驾驶轨迹数据传输给所述轨迹信息处理子系统,所述轨迹信息处理子系统根据所述驾驶轨迹数据分析计算并模拟出驾驶策略并传输给所述驾驶策略库,所述驾驶策略库接收并存储所述驾驶策略。Further, the storage module includes a driving trajectory library for storing historical driving trajectories, a trajectory information processing subsystem for calculating and simulating a driving strategy according to driving trajectories and driving habits, and a driving strategy library for storing driving strategies; the driving trajectory The library transmits the driving trajectory data to the trajectory information processing subsystem, and the trajectory information processing subsystem analyzes and calculates and simulates a driving strategy based on the driving trajectory data and transmits it to the driving strategy library, and the driving strategy library receives And store the driving strategy.
进一步地,所述轨迹信息处理子系统采用多目标的相对熵深度逆强化学习算法计算并模拟驾驶策略。Further, the trajectory information processing subsystem adopts a multi-objective relative entropy deep inverse reinforcement learning algorithm to calculate and simulate the driving strategy.
进一步地,所述多目标的逆强化学习算法采用EM算法框架嵌套相对熵深度逆强化学习计算多奖赏函数的参数。Further, the multi-objective inverse reinforcement learning algorithm uses the EM algorithm framework to nest the relative entropy depth inverse reinforcement learning to calculate the parameters of the multi-reward function.
进一步地,所述驾驶基础数据采集子系统包括用于采集道路信息的传感器。Further, the driving basic data collection subsystem includes sensors for collecting road information.
本发明还提供了一种基于相对熵深度逆强化学习的自动驾驶的方法,所述方法包括如下步骤:The present invention also provides a method for automatic driving based on relative entropy deep inverse reinforcement learning, said method comprising the following steps:
包括如下步骤:Including the following steps:
S1:采集道路信息并将所述道路信息传输给客户端及存储模块;S1: collect road information and transmit the road information to the client and the storage module;
S2:所述存储模块接收所述道路信息并将持续的一段道路信息存储为历史轨迹,根据所述历史轨迹分析计算并模拟多种驾驶策略,并将所述驾驶策略传递给所述客户端;S2: The storage module receives the road information and stores a continuous piece of road information as a historical trajectory, analyzes and simulates various driving strategies according to the historical trajectory, and transmits the driving strategy to the client;
S3:所述客户端接收所述道路信息及驾驶策略,并根据用户选择的个性驾驶策略及道路信息实施自动驾驶。S3: The client receives the road information and driving strategy, and implements automatic driving according to the personalized driving strategy and road information selected by the user.
进一步地,所述存储模块包括用于存储历史驾驶轨迹的驾驶轨迹库、根据驾驶规划及驾驶习惯计算并模拟出驾驶策略的轨迹信息处理子系统及存储驾驶策略的驾驶策略库;所述驾驶轨迹库将驾驶轨迹数据传输给所述轨迹信息处理子系统,所述轨迹信息处理子系统根据所述驾驶轨迹数据分析计算并模拟出驾驶策略并传输给所述驾驶策略库,所述驾驶策略库接收并存储所述驾驶策略。Further, the storage module includes a driving trajectory library for storing historical driving trajectories, a trajectory information processing subsystem for calculating and simulating a driving strategy according to driving planning and driving habits, and a driving strategy library for storing driving strategies; the driving trajectory The library transmits the driving trajectory data to the trajectory information processing subsystem, and the trajectory information processing subsystem analyzes and calculates and simulates a driving strategy based on the driving trajectory data and transmits it to the driving strategy library, and the driving strategy library receives And store the driving strategy.
进一步地,所述轨迹信息处理子系统采用多目标的相对熵深度逆强化学习算法计算并模拟驾驶策略。Further, the trajectory information processing subsystem adopts a multi-objective relative entropy deep inverse reinforcement learning algorithm to calculate and simulate the driving strategy.
进一步地,所述多目标的逆强化学习算法采用EM算法框架嵌套相对熵深度逆强化学习计算多奖赏函数的参数。Further, the multi-objective inverse reinforcement learning algorithm uses the EM algorithm framework to nest the relative entropy depth inverse reinforcement learning to calculate the parameters of the multi-reward function.
本发明的有益效果在于:通过在系统中设置驾驶基础数据采集子系统,实时采集道路信息,并将道路信息传递给存储模块,存储模块接收道路信息后并将持续的一段道路信息存储为历史轨迹,根据历史驾驶轨迹模拟驾驶策略,实现个性、智能的自动驾驶。The beneficial effects of the present invention are: by setting the driving basic data collection subsystem in the system, the road information is collected in real time, and the road information is transmitted to the storage module, and the storage module receives the road information and stores a continuous section of road information as a historical track , to simulate driving strategies based on historical driving trajectories to realize personalized and intelligent automatic driving.
上述说明仅是本发明技术方案的概述,为了能够更清楚了解本发明的技术手段,并可依照说明书的内容予以实施,以下以本发明的较佳实施例并配合附图详细说明如后。The above description is only an overview of the technical solutions of the present invention. In order to understand the technical means of the present invention more clearly and implement them according to the contents of the description, the preferred embodiments of the present invention and accompanying drawings are described in detail below.
附图说明Description of drawings
图1为本发明的基于相对熵深度逆强化学习的自动驾驶系统及方法的流程图。FIG. 1 is a flow chart of the automatic driving system and method based on relative entropy deep inverse reinforcement learning of the present invention.
图2为马尔科夫决策过程MDP示意图。Figure 2 is a schematic diagram of a Markov decision process MDP.
具体实施方式detailed description
下面结合附图和实施例,对本发明的具体实施方式作进一步详细描述。以下实施例用于说明本发明,但不用来限制本发明的范围。The specific implementation manners of the present invention will be further described in detail below in conjunction with the accompanying drawings and embodiments. The following examples are used to illustrate the present invention, but are not intended to limit the scope of the present invention.
请参见图1,本发明的一较佳实施例的基于相对熵深度逆强化学习的自动驾驶系统包括:Please refer to Fig. 1, the automatic driving system based on relative entropy deep inverse reinforcement learning of a preferred embodiment of the present invention includes:
客户端1:显示驾驶策略;Client 1: display driving strategy;
驾驶基础数据采集子系统2:采集道路信息;Driving basic data acquisition subsystem 2: collect road information;
存储模块3:与所述客户端1及驾驶基础数据采集子系统2连接并存储所述驾驶基础数据采集子系统2所采集到的道路信息;Storage module 3: connected with the client 1 and the driving basic data collection subsystem 2 and storing the road information collected by the driving basic data collection subsystem 2;
其中,所述驾驶基础数据采集子系统2采集道路信息并将所述道路信息传输给所述客户端1及存储模块3,所述存储模块3接收所述道路信息,并将持续的一段道路信息存储为历史轨迹,根据所述历史轨迹进行分析计算模拟出驾驶策略,所述存储模块3将所述驾驶策略传输至客户端1以供用户选择,所述客户端1接收所述道路信息并根据用户选择的个性驾驶策略实施自动驾驶。在本实施例中,所述存储模块3为云端。Wherein, the driving basic data collection subsystem 2 collects road information and transmits the road information to the client 1 and the storage module 3, and the storage module 3 receives the road information and stores a continuous section of road information It is stored as a historical trajectory, and the driving strategy is simulated according to the analysis and calculation of the historical trajectory. The storage module 3 transmits the driving strategy to the client 1 for the user to choose, and the client 1 receives the road information and according to the The personalized driving strategy selected by the user implements automatic driving. In this embodiment, the storage module 3 is a cloud.
所述客户端1最主要的功能是与用户完成人机交互过程,提供给个性的、智能的多种驾驶策略选择以及服务。客户端1根据用户个性的驾驶策略选择,从云端3驾驶策略库33下载相应的驾驶策略,接着根据驾驶策略和基础数据进行实时的驾驶决策,实现实时的无人驾驶控制。The main function of the client 1 is to complete the human-computer interaction process with the user, and provide personalized and intelligent multiple driving strategy selections and services. The client 1 downloads the corresponding driving strategy from the cloud 3 driving strategy library 33 according to the user's personalized driving strategy selection, and then makes real-time driving decisions according to the driving strategy and basic data to realize real-time unmanned driving control.
所述驾驶基础数据采集子系统2通过传感器(未图示)采集道路信息。采集到的信息有两个用途:将信息传递给客户端1,为当前的驾驶决策提供基础数据;将信息传递到云端3的驾驶轨迹库31,存储为用户驾驶员的历史驾驶轨迹数据。The driving basic data collection subsystem 2 collects road information through sensors (not shown). The collected information has two purposes: transfer the information to the client 1 to provide basic data for the current driving decision; transfer the information to the driving track database 31 in the cloud 3 and store it as the historical driving track data of the user driver.
所述云端3包括用于历史驾驶轨迹的驾驶轨迹库31、根据驾驶规划及驾驶习惯计算并模拟出驾驶策略的轨迹信息处理子系统32及存储驾驶策略的驾驶策略库33;所述驾驶轨迹库31将驾驶轨迹数据传输给所述轨迹信息处理子系统32,所述轨迹信息处理子系统32根据所述驾驶轨迹数据分析计算并模拟出驾驶策略并传输给所述驾驶策略库33,所述驾驶策略库33接收并存储所述驾驶策略。所述轨迹信息处理子系统32采用多目标的相对熵深度逆强化学习算法计算并模拟驾驶策略。在本实施例中,所述多目标的逆强化学习算法采用EM算法框架嵌套相对熵深度逆强化学习计算多奖赏函数的参数。所述历史驾驶轨迹包括专家历史驾驶轨迹及用户的历史轨迹。The cloud 3 includes a driving track library 31 for historical driving tracks, a track information processing subsystem 32 that calculates and simulates a driving strategy according to driving planning and driving habits, and a driving strategy library 33 that stores driving strategies; the driving track library 31 transmits the driving trajectory data to the trajectory information processing subsystem 32, and the trajectory information processing subsystem 32 analyzes and calculates and simulates a driving strategy according to the driving trajectory data and transmits it to the driving strategy library 33. The strategy library 33 receives and stores the driving strategies. The trajectory information processing subsystem 32 uses a multi-objective relative entropy deep inverse reinforcement learning algorithm to calculate and simulate driving strategies. In this embodiment, the multi-objective inverse reinforcement learning algorithm uses the EM algorithm framework to nest relative entropy depth inverse reinforcement learning to calculate the parameters of multiple reward functions. The historical driving trajectories include expert historical driving trajectories and user historical trajectories.
所述逆强化学习IRL是指在环境已知的马尔科夫决策过程MDP中奖赏函数R未知的问题。在一般的强化学习问题RL中,往往利用已知的环境、给定的奖赏函数R以及马尔科夫性质来估计一个状态动作对的值Q(s,a)(也可称为动作累积奖赏值),然后利用收敛的各个状态动作对的值Q(s,a)来求取策略π,智能体(Agent)便可利用策略π进行决策。在现实中,奖赏函数R往往是极难获知的,但是一些优秀的轨迹TN是比较容易获得的。在奖赏函数未知的马尔科夫决策过程MDP/R中,利用优秀的轨迹TN还原奖赏函数R的问题被称为逆强化学习问题IRL。The inverse reinforcement learning IRL refers to the problem that the reward function R is unknown in the environment-known Markov decision process MDP. In the general reinforcement learning problem RL, the known environment, the given reward function R and the Markov property are often used to estimate the value Q(s, a) of a state-action pair (also called the action cumulative reward value ), and then use the converged value Q(s, a) of each state-action pair to obtain the strategy π, and the agent (Agent) can use the strategy π to make decisions. In reality, the reward function R is often extremely difficult to obtain, but some excellent trajectories T N are relatively easy to obtain. In the Markov decision process MDP/R where the reward function is unknown, the problem of using an excellent trajectory T N to restore the reward function R is called the inverse reinforcement learning problem IRL.
在本实施例中,利用所述驾驶轨迹库31中已知的用户历史驾驶轨迹数据,进行相对熵深度逆强化学习,还原出多种用户个性的奖赏函数R,进而模拟出相应的驾驶策略π。相对熵深度逆强化学习算法是一种无模型的算法,无需已知环境模型中的状态转移函数T(s,a,s′),相对熵逆强化学习算法可以利用重要性采样的方法在计算中避开状态转移函数T(s,a,s′)。In this embodiment, the user’s historical driving trajectory data known in the driving trajectory database 31 is used to perform relative entropy depth inverse reinforcement learning, restore the reward function R of various user personalities, and then simulate the corresponding driving strategy π . The relative entropy deep inverse reinforcement learning algorithm is a model-free algorithm. It does not need to know the state transition function T(s, a, s′) in the environment model. The relative entropy inverse reinforcement learning algorithm can use the method of importance sampling to calculate Avoid the state transition function T(s, a, s′) in it.
在本实施例中,汽车的自动驾驶决策过程是一个没有奖赏函数的马尔科夫决策过程MDP/R,可以表示为集合{状态空间S,动作空间A,环境定义的状态转移概率T(省略对环境转移概率T的要求)。汽车Agent的值函数(累计奖赏值)可以表示为而汽车Agent的状态动作值函数可以表示为Q(s,a)=Rθ(s,a)+γET(s,a,s′)[V(s′)]。为了解决更加复杂的真实驾驶问题,对奖赏函数的假设不再只是简单的线性组合,而是假设为一个深度神经网络R(s,a,θ)=g1(g2(…(gn(f(s,a),θn),…),θ2),θ1),其中f(s,a)表示(s,a)处的驾驶的道路特征信息,θi表示深度神经网络第i层的参数。In this embodiment, the automatic driving decision process of the car is a Markov decision process MDP/R without a reward function, which can be expressed as a set {state space S, action space A, state transition probability T defined by the environment (omitted for environment transition probability T requirements). The value function (cumulative reward value) of the car Agent can be expressed as The state-action-value function of the vehicle Agent can be expressed as Q(s,a)=R θ (s,a)+γE T(s,a,s′) [V(s′)]. In order to solve more complex real driving problems, the assumption of the reward function is no longer a simple linear combination, but a deep neural network R(s,a,θ)=g 1 (g 2 (…(g n ( f(s,a),θ n ),…),θ 2 ),θ 1 ), where f(s,a) represents the road feature information of driving at (s,a), θ i represents the deep neural network The parameters of layer i.
同时,为了满足更个性、更智能的真实驾驶场景,假设有多个奖赏函数R(目标)同时存在,代表用户驾驶员不同的驾驶习惯。假设存在G个奖赏函数,令这G个奖赏函数的先验概率分布为ρ1,…,ρG,奖赏权重为θ1,…,θG,令Θ=(ρ1,…,ρG,θ1,…,θG),表示这G个奖赏函数的参数集合。At the same time, in order to meet more personalized and intelligent real driving scenarios, it is assumed that there are multiple reward functions R (objectives) existing at the same time, representing the different driving habits of the user driver. Suppose there are G reward functions, let the prior probability distribution of these G reward functions be ρ 1 ,…,ρ G , the reward weights are θ 1 ,…,θ G , let Θ=(ρ 1 ,…,ρ G , θ 1 ,…,θ G ), which represent the set of parameters of these G reward functions.
请参见图2,在已知有假设奖赏函数(由初始化或经过迭代获得)的条件下,此时我们可以把问题描述为一个完全的马尔科夫决策过程MDP。此时在完全的马尔科夫决策过程MDP下,根据强化学习的知识,利用奖赏函数R(s,a,θ)=g1(g2(…(gn(f,θn),…),θ2),θ1),我们可以对V值以及Q值进行评估。对于强化学习的评估算法,采用一种新的软最大化方法(MellowMax)来估计V值的期望值。MellowMax的生成器定义为:MellowMax是一种更优化的算法,它能够保证对V值的估计能够收敛于唯一一点。同时,MellowMax又具备特质:科学的概率分配机制和期望估计方法。在本实施例中,结合了MellowMax的强化学习算法在自动驾驶过程中对环境的探索和利用方面将更加合理。这保证了在强化学习过程收敛时,自动驾驶系统对各种情景已经有了足够的学习并能对当前状态产生较科学的评估。Please refer to Figure 2, under the condition that there is a hypothetical reward function (obtained by initialization or iteration), we can describe the problem as a complete Markov decision process MDP at this time. At this time, under the complete Markov decision process MDP, according to the knowledge of reinforcement learning, use the reward function R(s,a,θ)=g 1 (g 2 (…(g n (f,θ n ),…) ,θ 2 ),θ 1 ), we can evaluate the V value and Q value. For the evaluation algorithm of reinforcement learning, a new soft maximization method (MellowMax) is adopted to estimate the expected value of V value. The generator for MellowMax is defined as: MellowMax is a more optimized algorithm, which can guarantee that the estimation of V value can converge to a unique point. At the same time, MellowMax has special features: scientific probability distribution mechanism and expectation estimation method. In this embodiment, the reinforcement learning algorithm combined with MellowMax will be more reasonable in the exploration and utilization of the environment during the automatic driving process. This ensures that when the reinforcement learning process converges, the autopilot system has learned enough about various scenarios and can produce a more scientific assessment of the current state.
在本实施例中,根据结合了一种软最大化算法MellowMax的强化学习,可以获得对状态的特征的期望值更科学的评价。利用MellowMax可以获得动作选取的概率分布为在该软最大化的动作选取的规则下,利用强化学习的迭代过程,可以获得在以当前深度神经网络的参数为θ构成的奖赏函数所能够获得特征的期望值μ。μ可以理解为特征的累计的期望。In this embodiment, according to reinforcement learning combined with a soft maximization algorithm MellowMax, a more scientific evaluation of the expected value of the feature of the state can be obtained. Using MellowMax, the probability distribution of action selection can be obtained as Under the rule of soft maximization action selection, using the iterative process of reinforcement learning, the expected value μ of the features that can be obtained in the reward function composed of the parameters of the current deep neural network as θ can be obtained. μ can be understood as the cumulative expectation of features.
在本实施例中,利用EM算法来求解上述带隐变量的多目标逆强化学习问题。EM算法按步骤可分为E步和M步,通过E步、M步的不断迭代,逼近似然估计的极大值。In this embodiment, the EM algorithm is used to solve the above multi-objective inverse reinforcement learning problem with hidden variables. The EM algorithm can be divided into E step and M step according to the steps. Through the continuous iteration of E step and M step, the maximum value of the likelihood estimate is approximated.
E步:首先计算其中Z为正则项。zij代表第i个驾驶轨迹属于驾驶习惯(奖赏函数)j的概率。Step E: first calculate where Z is a regular term. z ij represents the probability that the i-th driving trajectory belongs to driving habit (reward function) j.
令yi=j表示第i个驾驶轨迹属于驾驶习惯j,并用y=(y1,…,yN)的集合表示N个驾驶轨迹的从属集合。Let y i =j indicate that the i-th driving trajectory belongs to driving habit j, and use the set of y=(y 1 ,...,y N ) to represent the subordinate set of N driving trajectories.
计算似然估计值Q(Θ,Θt)=∑yL(Θ|D,y)Pr(y|D,Θt)(这里所指的Q函数Q(Θ,Θt)是EM算法的更新目标函数,注意与强化学习中的Q动作状态值函数相区别),经过推算获得似然估计值 Calculate the likelihood estimate Q(Θ,Θ t )=∑ y L(Θ|D,y)Pr(y|D,Θ t ) (the Q function Q(Θ,Θ t ) referred to here is the EM algorithm Update the objective function, pay attention to the difference from the Q action state value function in reinforcement learning), and obtain the likelihood estimate after calculation
M步:选取合适的多驾驶习惯参数集合Θ(ρl和θl)使得E步中的似然估计值Q(Θ,Θt)极大化。由于ρl和θl的相互独立性,可以分开求它们的极大化。可以得到后半部分 Step M: Select an appropriate multi-driving habit parameter set Θ(ρ l and θ l ) to maximize the likelihood estimate Q(Θ,Θ t ) in step E. Due to the mutual independence of ρ l and θ l , their maximization can be obtained separately. can get second half
对于极大化Q(Θ,Θt)后半部分的更新目标:可以理解为 是关于在第l簇目标的参数为θl的条件下得到观察的轨迹集合所能够获得最大似然方程。我们可以利用相对熵深度逆强化学习的知识来求解这个最大似然方程。相对熵的求解公式,在符合最大似然更新目标的同时,可以很自然应用到深度神经网络参数的反向传播更新。令深度神经网络的最大化目标函数为L(θ)=logP(D,θ|r),根据联合似然函数的分解公式,可以获得L(θ)=logP(D,θ|r)=logP(D|r)+logP(θ)。对该联合似然目标函数求偏导可以获得对于该偏导的前半部分,可以进一步做分解,表示为 For the update objective that maximizes the second half of Q(Θ,Θ t ): can be understood as is the set of observed trajectories under the condition that the parameter of the l-th cluster target is θ l The maximum likelihood equation can be obtained. We can use the knowledge of relative entropy deep inverse reinforcement learning to solve this maximum likelihood equation. The solution formula of relative entropy can be naturally applied to the backpropagation update of deep neural network parameters while meeting the maximum likelihood update objective. Let the maximization objective function of the deep neural network be L(θ)=logP(D,θ|r), and according to the decomposition formula of the joint likelihood function, L(θ)=logP(D,θ|r)=logP can be obtained (D|r)+logP(θ). The partial derivative of the joint likelihood objective function can be obtained For the first half of the partial derivative, it can be further decomposed, expressed as
其中根据相对熵逆强化学习的知识,可以得到求解结果为当前奖赏函数下的特征期望值与专家特征值的差值其中,利用重要性采样,其中,π是一种给定的策略,根据这种策略π采样获得个轨迹。其中其中τ=s1a1,…,sHaH。进一步的,其中表示为更新深度神经网络中隐藏层参数时通过反向传播算法计算的梯度。in According to the knowledge of relative entropy inverse reinforcement learning, the solution result can be obtained as the difference between the feature expectation value under the current reward function and the expert feature value Among them, using importance sampling, Among them, π is a given strategy, and according to this strategy π sampling is obtained track. in where τ=s 1 a 1 ,...,s H a H . further, in Represented as gradients computed by the backpropagation algorithm when updating hidden layer parameters in a deep neural network.
梯度更新完成标志着一次相对熵深度逆强化学习迭代更新的完成。利用更新完成了参数更新的新的深度网络奖赏函数产生新的策略π,进行新的迭代。The completion of the gradient update marks the completion of an iterative update of relative entropy deep inverse reinforcement learning. The new deep network reward function that has completed the parameter update is used to generate a new strategy π for a new iteration.
不断迭代进行E步和M步的计算,直至似然估计值Q(Θ,Θt)收敛至极大值。此时获得的参数集合Θ=(ρ1,…,ρG,θ1,…,θG),就是我们想要求解的代表多驾驶习惯的奖赏函数的先验分布和权重。E-step and M-step calculations are iteratively performed until the likelihood estimate Q(Θ,Θ t ) converges to the maximum value. The parameter set Θ=(ρ 1 ,...,ρ G ,θ 1 ,...,θ G ) obtained at this time is the prior distribution and weight of the reward function representing multiple driving habits that we want to solve.
在本实施例中,根据这个参数集合Θ,经过强化学习RL的计算,获得每个驾驶习惯R的驾驶策略π。输出多驾驶策略,并保存在云端的驾驶策略库中。用户便可在客户端中选择个性、智能的驾驶策略。In this embodiment, according to this parameter set Θ, the driving strategy π for each driving habit R is obtained through reinforcement learning RL calculation. Output multiple driving strategies and save them in the driving strategy library in the cloud. The user can choose a personalized and intelligent driving strategy in the client.
本发明还提供了一种基于相对熵深度逆强化学习的自动驾驶的方法,所述方法包括如下步骤:The present invention also provides a method for automatic driving based on relative entropy deep inverse reinforcement learning, said method comprising the following steps:
S1:采集道路信息并将所述道路信息传输给客户端及存储模块;S1: collect road information and transmit the road information to the client and the storage module;
S2:所述存储模块接收所述道路信息并根据所述道路信息分析计算并模拟多种驾驶策略,并将所述驾驶策略传递给所述客户端;S2: The storage module receives the road information, analyzes and calculates and simulates various driving strategies according to the road information, and transmits the driving strategies to the client;
S3:所述客户端接收所述道路信息及驾驶策略,并根据用户选择的个性驾驶策略及道路信息实施自动驾驶。S3: The client receives the road information and driving strategy, and implements automatic driving according to the personalized driving strategy and road information selected by the user.
综上所述:通过在系统中设置驾驶基础数据采集子系统2,实时采集道路信息,并将道路信息传递给存储模块3及客户端1,存储模块3接收道路信息后根据历史驾驶轨迹模拟驾驶策略,实现个性、智能的自动驾驶。To sum up: by setting the driving basic data acquisition subsystem 2 in the system, road information is collected in real time, and the road information is transmitted to the storage module 3 and the client 1. After receiving the road information, the storage module 3 simulates driving according to the historical driving trajectory strategies to realize personalized and intelligent autonomous driving.
基于本方法的自动驾驶中,驾驶策略都在云端3中实现计算,而不是在客户端1中运行计算过程。当用户在需要进行自动驾驶的时候,所有驾驶策略都已经在云端3完成。用户只需要选择下载自己所需的驾驶策略,车体就可以根据用户所选择的驾驶策略和实时道路信息进行实时的自动驾驶。同时,在完成任何一次驾驶之后,大量的道路信息上传到云端3被存储为历史驾驶轨迹。利用存储的历史驾驶轨迹大数据,再实现对驾驶策略库的更新。利用轨迹信息大数据,本系统将实现更加贴近用户需求的自动驾驶。In the automatic driving based on this method, the driving strategy is calculated in the cloud 3 instead of running the calculation process in the client 1 . When the user needs to drive automatically, all driving strategies have been completed on the cloud 3. Users only need to choose to download the driving strategy they need, and the car body can perform real-time automatic driving according to the driving strategy selected by the user and real-time road information. Simultaneously, after completing any driving, a large amount of road information is uploaded to the cloud 3 and stored as historical driving tracks. Utilize the big data of stored historical driving trajectories to update the driving strategy library. Using the big data of trajectory information, this system will realize automatic driving that is closer to the needs of users.
以上所述实施例的各技术特征可以进行任意的组合,为使描述简洁,未对上述实施例中的各个技术特征所有可能的组合都进行描述,然而,只要这些技术特征的组合不存在矛盾,都应当认为是本说明书记载的范围。The various technical features of the above-mentioned embodiments can be combined arbitrarily. To make the description concise, all possible combinations of the various technical features in the above-mentioned embodiments are not described. However, as long as there is no contradiction in the combination of these technical features, should be considered as within the scope of this specification.
以上所述实施例仅表达了本发明的几种实施方式,其描述较为具体和详细,但并不能因此而理解为对发明专利范围的限制。应当指出的是,对于本领域的普通技术人员来说,在不脱离本发明构思的前提下,还可以做出若干变形和改进,这些都属于本发明的保护范围。因此,本发明专利的保护范围应以所附权利要求为准。The above-mentioned embodiments only express several implementation modes of the present invention, and the descriptions thereof are relatively specific and detailed, but should not be construed as limiting the patent scope of the invention. It should be noted that those skilled in the art can make several modifications and improvements without departing from the concept of the present invention, and these all belong to the protection scope of the present invention. Therefore, the protection scope of the patent for the present invention should be based on the appended claims.
Claims (9)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710940590.XA CN107544516A (en) | 2017-10-11 | 2017-10-11 | Automated driving system and method based on relative entropy depth against intensified learning |
PCT/CN2018/078740 WO2019071909A1 (en) | 2017-10-11 | 2018-03-12 | Automatic driving system and method based on relative-entropy deep inverse reinforcement learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710940590.XA CN107544516A (en) | 2017-10-11 | 2017-10-11 | Automated driving system and method based on relative entropy depth against intensified learning |
Publications (1)
Publication Number | Publication Date |
---|---|
CN107544516A true CN107544516A (en) | 2018-01-05 |
Family
ID=60967749
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710940590.XA Pending CN107544516A (en) | 2017-10-11 | 2017-10-11 | Automated driving system and method based on relative entropy depth against intensified learning |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN107544516A (en) |
WO (1) | WO2019071909A1 (en) |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109460015A (en) * | 2017-09-06 | 2019-03-12 | 通用汽车环球科技运作有限责任公司 | Unsupervised learning agency for autonomous driving application |
CN109636432A (en) * | 2018-09-28 | 2019-04-16 | 阿里巴巴集团控股有限公司 | The project selection method and device that computer executes |
WO2019071909A1 (en) * | 2017-10-11 | 2019-04-18 | 苏州大学张家港工业技术研究院 | Automatic driving system and method based on relative-entropy deep inverse reinforcement learning |
CN110238855A (en) * | 2019-06-24 | 2019-09-17 | 浙江大学 | A robot out-of-sequence workpiece grasping method based on deep inverse reinforcement learning |
CN110321811A (en) * | 2019-06-17 | 2019-10-11 | 中国工程物理研究院电子工程研究所 | Depth is against the object detection method in the unmanned plane video of intensified learning |
WO2019237474A1 (en) * | 2018-06-11 | 2019-12-19 | 苏州大学 | Partially-observable automatic driving decision-making method and system based on constraint online planning |
WO2020000192A1 (en) * | 2018-06-26 | 2020-01-02 | Psa Automobiles Sa | Method for providing vehicle trajectory prediction |
CN110654372A (en) * | 2018-06-29 | 2020-01-07 | 比亚迪股份有限公司 | Vehicle driving control method and device, vehicle and storage medium |
CN110837258A (en) * | 2019-11-29 | 2020-02-25 | 商汤集团有限公司 | Automatic driving control method and device, system, electronic device and storage medium |
CN110850861A (en) * | 2018-07-27 | 2020-02-28 | 通用汽车环球科技运作有限责任公司 | Attention-based hierarchical lane change depth reinforcement learning |
CN110955239A (en) * | 2019-11-12 | 2020-04-03 | 中国地质大学(武汉) | A multi-target trajectory planning method and system for unmanned ships based on inverse reinforcement learning |
CN111026127A (en) * | 2019-12-27 | 2020-04-17 | 南京大学 | Automatic driving decision method and system based on partially observable transfer reinforcement learning |
CN111159832A (en) * | 2018-10-19 | 2020-05-15 | 百度在线网络技术(北京)有限公司 | Construction method and device of traffic information flow |
CN114194211A (en) * | 2021-11-30 | 2022-03-18 | 浪潮(北京)电子信息产业有限公司 | An automatic driving method, device, electronic device and storage medium |
CN114510031A (en) * | 2021-12-31 | 2022-05-17 | 中原动力智能机器人有限公司 | Robot vision navigation method, device, robot and storage medium |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110673602B (en) * | 2019-10-24 | 2022-11-25 | 驭势科技(北京)有限公司 | Reinforced learning model, vehicle automatic driving decision method and vehicle-mounted equipment |
TWI737437B (en) * | 2020-08-07 | 2021-08-21 | 財團法人車輛研究測試中心 | Trajectory determination method |
US20230143937A1 (en) * | 2021-11-10 | 2023-05-11 | International Business Machines Corporation | Reinforcement learning with inductive logic programming |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140278052A1 (en) * | 2013-03-15 | 2014-09-18 | Caliper Corporation | Lane-level vehicle navigation for vehicle routing and traffic management |
CN106842925A (en) * | 2017-01-20 | 2017-06-13 | 清华大学 | A kind of locomotive smart steering method and system based on deeply study |
CN107074178A (en) * | 2014-09-16 | 2017-08-18 | 本田技研工业株式会社 | Drive assistance device |
CN107084735A (en) * | 2017-04-26 | 2017-08-22 | 电子科技大学 | Navigation path framework for reducing redundant navigation |
CN107169567A (en) * | 2017-03-30 | 2017-09-15 | 深圳先进技术研究院 | The generation method and device of a kind of decision networks model for Vehicular automatic driving |
CN107200017A (en) * | 2017-05-22 | 2017-09-26 | 北京联合大学 | A kind of automatic driving vehicle control system based on deep learning |
CN107229973A (en) * | 2017-05-12 | 2017-10-03 | 中国科学院深圳先进技术研究院 | The generation method and device of a kind of tactful network model for Vehicular automatic driving |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103699717A (en) * | 2013-12-03 | 2014-04-02 | 重庆交通大学 | Complex road automobile traveling track predication method based on foresight cross section point selection |
CN105718750B (en) * | 2016-01-29 | 2018-08-17 | 长沙理工大学 | A kind of prediction technique and system of vehicle driving trace |
CN107544516A (en) * | 2017-10-11 | 2018-01-05 | 苏州大学 | Automated driving system and method based on relative entropy depth against intensified learning |
-
2017
- 2017-10-11 CN CN201710940590.XA patent/CN107544516A/en active Pending
-
2018
- 2018-03-12 WO PCT/CN2018/078740 patent/WO2019071909A1/en active Application Filing
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140278052A1 (en) * | 2013-03-15 | 2014-09-18 | Caliper Corporation | Lane-level vehicle navigation for vehicle routing and traffic management |
CN107074178A (en) * | 2014-09-16 | 2017-08-18 | 本田技研工业株式会社 | Drive assistance device |
CN106842925A (en) * | 2017-01-20 | 2017-06-13 | 清华大学 | A kind of locomotive smart steering method and system based on deeply study |
CN107169567A (en) * | 2017-03-30 | 2017-09-15 | 深圳先进技术研究院 | The generation method and device of a kind of decision networks model for Vehicular automatic driving |
CN107084735A (en) * | 2017-04-26 | 2017-08-22 | 电子科技大学 | Navigation path framework for reducing redundant navigation |
CN107229973A (en) * | 2017-05-12 | 2017-10-03 | 中国科学院深圳先进技术研究院 | The generation method and device of a kind of tactful network model for Vehicular automatic driving |
CN107200017A (en) * | 2017-05-22 | 2017-09-26 | 北京联合大学 | A kind of automatic driving vehicle control system based on deep learning |
Cited By (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109460015A (en) * | 2017-09-06 | 2019-03-12 | 通用汽车环球科技运作有限责任公司 | Unsupervised learning agency for autonomous driving application |
CN109460015B (en) * | 2017-09-06 | 2022-04-15 | 通用汽车环球科技运作有限责任公司 | Unsupervised learning agent for autonomous driving applications |
WO2019071909A1 (en) * | 2017-10-11 | 2019-04-18 | 苏州大学张家港工业技术研究院 | Automatic driving system and method based on relative-entropy deep inverse reinforcement learning |
WO2019237474A1 (en) * | 2018-06-11 | 2019-12-19 | 苏州大学 | Partially-observable automatic driving decision-making method and system based on constraint online planning |
WO2020000192A1 (en) * | 2018-06-26 | 2020-01-02 | Psa Automobiles Sa | Method for providing vehicle trajectory prediction |
CN110654372A (en) * | 2018-06-29 | 2020-01-07 | 比亚迪股份有限公司 | Vehicle driving control method and device, vehicle and storage medium |
CN110654372B (en) * | 2018-06-29 | 2021-09-03 | 比亚迪股份有限公司 | Vehicle driving control method and device, vehicle and storage medium |
CN110850861B (en) * | 2018-07-27 | 2023-05-23 | 通用汽车环球科技运作有限责任公司 | Attention-based hierarchical lane-changing depth reinforcement learning |
CN110850861A (en) * | 2018-07-27 | 2020-02-28 | 通用汽车环球科技运作有限责任公司 | Attention-based hierarchical lane change depth reinforcement learning |
CN109636432A (en) * | 2018-09-28 | 2019-04-16 | 阿里巴巴集团控股有限公司 | The project selection method and device that computer executes |
CN109636432B (en) * | 2018-09-28 | 2023-05-30 | 创新先进技术有限公司 | Computer-implemented item selection method and apparatus |
CN111159832A (en) * | 2018-10-19 | 2020-05-15 | 百度在线网络技术(北京)有限公司 | Construction method and device of traffic information flow |
CN111159832B (en) * | 2018-10-19 | 2024-04-02 | 百度在线网络技术(北京)有限公司 | Traffic information stream construction method and device |
CN110321811B (en) * | 2019-06-17 | 2023-05-02 | 中国工程物理研究院电子工程研究所 | Object Detection Method in UAV Aerial Video with Deep Inverse Reinforcement Learning |
CN110321811A (en) * | 2019-06-17 | 2019-10-11 | 中国工程物理研究院电子工程研究所 | Depth is against the object detection method in the unmanned plane video of intensified learning |
CN110238855A (en) * | 2019-06-24 | 2019-09-17 | 浙江大学 | A robot out-of-sequence workpiece grasping method based on deep inverse reinforcement learning |
CN110955239A (en) * | 2019-11-12 | 2020-04-03 | 中国地质大学(武汉) | A multi-target trajectory planning method and system for unmanned ships based on inverse reinforcement learning |
CN110837258A (en) * | 2019-11-29 | 2020-02-25 | 商汤集团有限公司 | Automatic driving control method and device, system, electronic device and storage medium |
CN110837258B (en) * | 2019-11-29 | 2024-03-08 | 商汤集团有限公司 | Automatic driving control method, device, system, electronic equipment and storage medium |
CN111026127A (en) * | 2019-12-27 | 2020-04-17 | 南京大学 | Automatic driving decision method and system based on partially observable transfer reinforcement learning |
CN111026127B (en) * | 2019-12-27 | 2021-09-28 | 南京大学 | Automatic driving decision method and system based on partially observable transfer reinforcement learning |
CN114194211A (en) * | 2021-11-30 | 2022-03-18 | 浪潮(北京)电子信息产业有限公司 | An automatic driving method, device, electronic device and storage medium |
CN114194211B (en) * | 2021-11-30 | 2023-04-25 | 浪潮(北京)电子信息产业有限公司 | An automatic driving method, device, electronic equipment, and storage medium |
CN114510031A (en) * | 2021-12-31 | 2022-05-17 | 中原动力智能机器人有限公司 | Robot vision navigation method, device, robot and storage medium |
Also Published As
Publication number | Publication date |
---|---|
WO2019071909A1 (en) | 2019-04-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107544516A (en) | Automated driving system and method based on relative entropy depth against intensified learning | |
CN112162555B (en) | Vehicle control method based on reinforcement learning control strategy in mixed fleet | |
EP3719603B1 (en) | Action control method and apparatus | |
CN111158401B (en) | Distributed unmanned aerial vehicle path planning system and method for encouraging space-time data exploration | |
CN113299085A (en) | Traffic signal lamp control method, equipment and storage medium | |
CN112101684B (en) | Plug-in hybrid electric vehicle real-time energy management method and system | |
CN108229730B (en) | A Trajectory Generation Method for Unmanned Vehicles Based on Fuzzy Rewards | |
CN112633591B (en) | Space searching method and device based on deep reinforcement learning | |
CN113238970B (en) | Training method, evaluation method, control method and device of automatic driving model | |
CN110488842A (en) | A kind of track of vehicle prediction technique based on two-way kernel ridge regression | |
CN111752304B (en) | UAV data collection method and related equipment | |
CN118535939B (en) | Vehicle trajectory prediction method, device, medium and equipment | |
CN110456799A (en) | A kind of online incremental learning method of automatic driving vehicle Controlling model | |
CN116895158B (en) | A traffic signal control method for urban road networks based on multi-agent Actor-Critic and GRU | |
CN119088074A (en) | A reinforcement learning method for dynamic battery management of unmanned aerial vehicles with adaptive path planning | |
CN113554680B (en) | Target tracking method, device, drone and storage medium | |
US20220107628A1 (en) | Systems and methods for distributed hierarchical control in multi-agent adversarial environments | |
CN118928464A (en) | Method and device for generating automatic driving decision based on hybrid expert model | |
CN116662815B (en) | Training method of time prediction model and related equipment | |
US12085399B2 (en) | Modular machine-learning based system for predicting vehicle energy consumption during a trip | |
CN118560529B (en) | Multimodal scenario driving behavior modeling method and system based on inverse reinforcement learning | |
CN119149978B (en) | Driving training teaching method and system based on imitation learning | |
CN119597020B (en) | An optimal path planning method and system for unmanned aerial vehicle swarm | |
CN119322530B (en) | A UAV control method and system with accuracy maintenance and enhanced generalization capability | |
CN110286677B (en) | Unmanned vehicle control method and system for data acquisition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
TA01 | Transfer of patent application right |
Effective date of registration: 20201228 Address after: 210034 building C4, Hongfeng Science Park, Nanjing Economic and Technological Development Zone, Jiangsu Province Applicant after: NANQI XIANCE (NANJING) TECHNOLOGY Co.,Ltd. Address before: 215006 No.8, Jixue Road, Xiangcheng District, Suzhou City, Jiangsu Province Applicant before: Suzhou University |
|
TA01 | Transfer of patent application right | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20180105 |
|
RJ01 | Rejection of invention patent application after publication |