[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN109709956A - A kind of automatic driving vehicle speed control multiple-objection optimization with algorithm of speeding - Google Patents

A kind of automatic driving vehicle speed control multiple-objection optimization with algorithm of speeding Download PDF

Info

Publication number
CN109709956A
CN109709956A CN201811600366.7A CN201811600366A CN109709956A CN 109709956 A CN109709956 A CN 109709956A CN 201811600366 A CN201811600366 A CN 201811600366A CN 109709956 A CN109709956 A CN 109709956A
Authority
CN
China
Prior art keywords
speeding
data
ttc
headway
ngsim
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201811600366.7A
Other languages
Chinese (zh)
Other versions
CN109709956B (en
Inventor
王雪松
朱美新
孙平
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tongji University
Original Assignee
Tongji University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tongji University filed Critical Tongji University
Priority to CN201811600366.7A priority Critical patent/CN109709956B/en
Publication of CN109709956A publication Critical patent/CN109709956A/en
Application granted granted Critical
Publication of CN109709956B publication Critical patent/CN109709956B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Traffic Control Systems (AREA)
  • Control Of Driving Devices And Active Controlling Of Vehicle (AREA)
  • Feedback Control In General (AREA)

Abstract

The present invention develop a kind of automatic driving vehicle speed control multiple-objection optimization with algorithm of speeding.The algorithm proposes a kind of model for automobile with speed control of speeding based on deeply study, which not only imitates mankind's driving, but directly optimizes drive safety, efficiency and comfort.In conjunction with collision time, the distribution of time headway experience, acceleration, construct reflection drive safety, the reward function of efficiency and comfort, (Next Generation Simulation is emulated using the next generation, NGSIM) practical driving data training pattern in project, and modeling is compared with speeding on the behavior observed in NGSIM empirical data, intensified learning intelligent body by test in simulated environment and trial and error, learnt in a manner of maximizing progressive award safety, it is comfortable, efficiently control car speed.The result shows that proposition shows better safe and efficient and comfortable driving ability with speed control algorithm of speeding compared with human driver in the real world.

Description

A kind of automatic driving vehicle speed control multiple-objection optimization with algorithm of speeding
Technical field
The present invention relates to automatic Pilots with control field of speeding, in particular to a kind of automatic driving vehicle speed control multiple target Optimization with algorithm of speeding.
Background technique
It is the important component of automatic Pilot intelligent decision with control of speeding, including the speed selection under freely driving, vehicle With the braking under the holding and emergency of spacing at any time.In the case where automatic Pilot and the mankind drive and coexist, automatically It drives vehicle and makes the comfort level and letter that will improve passenger with control decision of speeding similar to human driver's (referred to as personalizing) Ren Du, while other traffic participants also being facilitated to more fully understand and predict the behavior of automatic driving vehicle, it is driven automatically with realizing Sail the secure interactive between mankind's driving.However traditional following-speed model is when being applied to automatically with speeding to control that there are many limitations Property, such as the flexibility and accuracy of limited model, it is difficult to which the Driving Scene being generalized to other than nominal data and driver are applied to Driving style and the Driving Scene etc. of the practical driver of vehicle cannot be reacted when automatic Pilot.
Deeply learn (Deep Reinforcement Learning, DRL) be widely used in industry manufacture, Analogue simulation, robot control, optimization and scheduling and the fields such as game play, basic thought be by maximize intelligent body from The accumulative reward value obtained in environment, to learn to the optimal policy for completing target.DRL method more lays particular emphasis on study and solves to ask The strategy of topic, and non-logarithmic evidence is fitted, therefore its generalization ability is stronger, provides ginseng with control of speeding for automatic driving vehicle It examines.
Summary of the invention
The purpose of the present invention is a kind of: automatic driving vehicle speed control multiple-objection optimizations with algorithm of speeding.The algorithm mentions A kind of model for automobile with speed control of speeding is gone out, which directly optimizes drive safety, efficiency and comfort.In conjunction with Collision time TTC, the distribution of time headway experience, acceleration (Jerk), construct reflection drive safety, efficiency and comfort Reward function, using practical driving data training pattern in next generation's emulation (NGSIM) project, and by modeling with speeding The behavior observed in behavior and NGSIM empirical data is compared, and intensified learning intelligent body passes through the test in simulated environment And trial and error, learnt in a manner of maximizing progressive award safety, it is comfortable, efficiently control car speed.The result shows that with reality Human driver in the world compares, and proposition shows better safe and efficient and comfortable driving with speed control algorithm of speeding Ability.
The technical scheme adopted by the invention is that:
A kind of automatic driving vehicle speed control multiple-objection optimization with algorithm of speeding, steps are as follows:
Step 1: obtaining data.Using the data in NGSIM project, rested on same lane based on front truck and rear car and Vehicle follows the criterion such as length > 15 second of event to extract with the event of speeding, based on extraction with the event of speeding, by a part as training Data, another part is as test data.
Step 2: building reward function.It is proposed the feature of reflection automobile model- following control related objective (safety, comfortable, efficiency) Amount.
Step 2.1: safety is reflected using collision time (TTC).TTC indicates the remaining time before two cars collision Amount, formula areIt is wherein Sn-1, n (t) vehicle headway, △ Vn-1, n (t) are relative velocities.According to NGSIM empirical data determines that secure threshold is 7 seconds, and carries out TTC feature construction:If TTC less than 7 seconds, Then TTC characteristic index be negative value, with TTC approach zero, TTC feature will close to bear it is infinite, for close to collision the case where show Most severe punishment.
Step 2.2: driving efficiency is measured using time headway (headway).By analyzing, logarithm normal distribution is adapted to obtain The distribution of the training data taken, probability density function arex>0.According to being mentioned The data taken can estimate that the average value mu and logarithm standard deviation σ of distribution variable x is respectively 0.4226 and 0.4365.By time headway Feature construction is the probability density value of the time headway logarithm normal distribution of estimation: Fheadway=flognormal (headway | μ=0.4226, σ=0.4365).According to the headstock temporal characteristics, about 1.3 seconds time headways correspond to high characteristic value, headstock When away from it is too long or it is too short correspond to low characteristic value, therefore this feature value estimation high flow capacity spacing keeps behavior, while punishing dangerous Or spacing too far keeps behavior.
Step 2.3: driver comfort, feature construction are measured using the change rate Jerk of acceleration are as follows:
Step 2.4: establishing comprehensive reward function.R=w1FTTC+w2Fheadway+ is established according to above step W3Fjerk., wherein w1, w2, w3 be feature coefficient, be all set to 1.
Step 3: training pattern.Every time when training, sequence emulate in data with the event of speeding, training is repeated as many times, and is selected The model of maximum average reward is obtained in test data as final mask.
Step 4: evaluation model.NGSIM data and DDPG model are evaluated using Indexes Comparisons such as TTC, headway and jerk Simulate obtain with speed on for.
The invention has the advantages that
1. the automatic driving vehicle developed can be applied to automatic driving vehicle exploitation with control logic of speeding;
2. the algorithm model does not imitate mankind's driving, but directly optimizes drive safety, efficiency and comfort.
Detailed description of the invention
Fig. 1 is flow chart of the invention.
Fig. 2 NGSIM data are compared with DDPG model drive safety.
Fig. 3 NGSIM data are compared with driver comfort between DDPG model.
Specific embodiment
The algorithm proposes a kind of model for automobile with speed control of speeding based on deeply study, the model not mould Apery class drives, but directly optimizes drive safety, efficiency and comfort.In conjunction with collision time TTC, time headway experience point Cloth, acceleration (Jerk), construct reflection drive safety, and the reward function of efficiency and comfort is emulated using the next generation (NGSIM) practical driving data training pattern in project, and by modeling with speed on for observed in NGSIM empirical data To behavior be compared, intensified learning intelligent body is by test in simulated environment and trial and error, to maximize progressive award Mode learn safety, it is comfortable, efficiently control car speed.The result shows that being mentioned compared with human driver in the real world Out show better safe and efficient and comfortable driving ability with speed control algorithm of speeding.The result shows that with real world Human driver compare, proposition shows better safe and efficient and comfortable driving ability with speed control algorithm of speeding.
The present invention is described in detail with specific embodiment below in conjunction with the accompanying drawings, and steps are as follows:
Step 1: obtaining data.Using the data in next-generation emulation (NGSIM) project, rested on based on front truck and rear car On same lane and vehicle follow the criterion such as length > 15 second of event extract with the event of speeding, based on extraction with the event of speeding, by one Part is used as training data, and another part is as test data.
Step 2: building reward function.It is proposed the feature of reflection automobile model- following control related objective (safety, comfortable, efficiency) Amount.
Step 2.1: safety is reflected using collision time (TTC).TTC indicates the remaining time before two cars collision Amount, formula areIt is wherein Sn-1, n (t) vehicle headway, △ Vn-1, n (t) are relative velocities.According to NGSIM empirical data determines that secure threshold is 7 seconds, and carries out TTC feature construction:If TTC is less than 7 Second, then TTC characteristic index be negative value, with TTC approach zero, TTC feature will close to bear it is infinite, for close to collision the case where table Now most severe punishment.
Step 2.2: driving efficiency is measured using time headway (headway).By analyzing, logarithm normal distribution is adapted to obtain The distribution of the training data taken, probability density function arex>0.According to being mentioned The data taken can estimate that the average value mu and logarithm standard deviation σ of distribution variable x is respectively 0.4226 and 0.4365.By time headway Feature construction is the probability density value of the time headway logarithm normal distribution of estimation: Fheadway=flognormal (headway | μ=0.4226, σ=0.4365).According to the headstock temporal characteristics, about 1.3 seconds time headways correspond to high characteristic value, headstock When away from it is too long or it is too short correspond to low characteristic value, therefore this feature value estimation high flow capacity spacing keeps behavior, while punishing dangerous Or spacing too far keeps behavior.
Step 2.3: driver comfort, feature construction are measured using the change rate Jerk of acceleration are as follows:
Step 2.4: establishing comprehensive reward function.R=w1FTTC is established according to above step 2.1, step 2.2, step 2.3 + w2Fheadway+w3Fjerk., wherein w1, w2, w3 are the coefficients of feature, are all set to 1.
Step 3: training pattern.Every time when training, sequence emulate in data with the event of speeding, training is repeated as many times, and is selected The model of maximum average reward is obtained in test data as final mask.
Step 4: evaluation model.NGSIM data and DDPG model are evaluated using Indexes Comparisons such as TTC, headway and jerk Simulate obtain with speed on for.
Embodiment
By comparing experience NGSIM data and DDPG modeling obtain with speed on for, test the model can safely, Efficiently, front truck is comfortably followed.
Obtain data.Using the data in NGSIM project, rested on same lane based on front truck and rear car and vehicle with The criterion such as length > 15 second with event are extracted with the event of speeding.
In terms of drive safety, one has been randomly choosed from NGSIM data set with the event of speeding.Fig. 2, which is shown, to be observed Speed, spacing and acceleration, and the corresponding index value generated by DDPG model.Driver in NGSIM data was at 10 seconds Afterwards with the driving of very small following distance, and DDPG model remains that about 10 meters follow gap.
In terms of driver comfort, one has been randomly choosed in NGSIM data set with the event of speeding.Fig. 3, which is shown, to be observed Speed, spacing, acceleration and Jerk value, and the correspondence index value generated by DDPG model.Driver in NGSIM data Frequent acceleration change and big Jerk value are produced in driving procedure, and DDPG model can be kept close to constant acceleration It spends and generates low Jerk value.
Based on the above, what is proposed shows preferably with speed control algorithm of speeding compared with human driver in NGSIM Safe and efficient and comfortable driving ability.

Claims (1)

1. a kind of automatic driving vehicle speed control multiple-objection optimization with algorithm of speeding, which is characterized in that steps are as follows:
Step 1: obtaining data;Using the data in NGSIM project, rested on same lane based on front truck and rear car and vehicle Follow the criterion such as length > 15 second of event to extract with the event of speeding, based on extraction with the event of speeding, by a part as training number According to another part is as test data;
Step 2: building reward function;It is proposed the characteristic quantity of reflection automobile model- following control related objective (safety, comfortable, efficiency);
Step 2.1: safety is reflected using collision time (TTC);TTC indicates remaining time quantum before two cars collision, Formula isIt is wherein Sn-1, n (t) vehicle headway, △ Vn-1, n (t) are relative velocities;According to NGSIM Empirical data determines that secure threshold is 7 seconds, and carries out TTC feature construction:If TTC was less than 7 seconds, TTC Characteristic index is negative value, with TTC approach zero, TTC feature will close to bear it is infinite, for close to collision the case where performance it is most severe Punishment;
Step 2.2: driving efficiency is measured using time headway (headway);By analyzing, logarithm normal distribution is adapted to acquisition The distribution of training data, probability density function arex>0;According to extracted Data can estimate that the average value mu and logarithm standard deviation σ of distribution variable x is respectively 0.4226 and 0.4365;By time headway feature Be configured to estimation time headway logarithm normal distribution probability density value: Fheadway=flognormal (headway | μ= 0.4226, σ=0.4365);According to the headstock temporal characteristics, about 1.3 seconds time headways correspond to high characteristic value, time headway It is too long or it is too short correspond to low characteristic value, therefore this feature value estimation high flow capacity spacing keeps behavior, while punishing dangerous or mistake Remote spacing keeps behavior;
Step 2.3: driver comfort, feature construction are measured using the change rate Jerk of acceleration are as follows:
Step 2.4: establishing comprehensive reward function;R=w1FTTC+w2Fheadway+w3Fjerk. is established according to above step, Middle w1, w2, w3 are the coefficients of feature, are all set to 1;
Step 3: training pattern;Every time when training, sequence emulate in data with the event of speeding, training is repeated as many times, and selection is being surveyed The model for trying to obtain maximum average reward in data is as final mask;
Step 4: evaluation model;NGSIM data and DDPG modeling are evaluated using Indexes Comparisons such as TTC, headway and jerk Obtain with speed on for;
Using the Indexes Comparisons such as TTC, headway and jerk evaluate that NGSIM data and DDPG modeling obtain with speed on for.
CN201811600366.7A 2018-12-26 2018-12-26 Multi-objective optimized following algorithm for controlling speed of automatic driving vehicle Active CN109709956B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811600366.7A CN109709956B (en) 2018-12-26 2018-12-26 Multi-objective optimized following algorithm for controlling speed of automatic driving vehicle

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811600366.7A CN109709956B (en) 2018-12-26 2018-12-26 Multi-objective optimized following algorithm for controlling speed of automatic driving vehicle

Publications (2)

Publication Number Publication Date
CN109709956A true CN109709956A (en) 2019-05-03
CN109709956B CN109709956B (en) 2021-06-08

Family

ID=66258357

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811600366.7A Active CN109709956B (en) 2018-12-26 2018-12-26 Multi-objective optimized following algorithm for controlling speed of automatic driving vehicle

Country Status (1)

Country Link
CN (1) CN109709956B (en)

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110321605A (en) * 2019-06-19 2019-10-11 中汽研(天津)汽车工程研究院有限公司 A kind of human-computer interaction coordination control strategy based on Multiple Velocity Model PREDICTIVE CONTROL
CN110347043A (en) * 2019-07-15 2019-10-18 武汉天喻信息产业股份有限公司 A kind of intelligent driving control method and device
CN110488802A (en) * 2019-08-21 2019-11-22 清华大学 A kind of automatic driving vehicle dynamic behaviour decision-making technique netted under connection environment
CN110716562A (en) * 2019-09-25 2020-01-21 南京航空航天大学 Decision-making method for multi-lane driving of unmanned vehicle based on reinforcement learning
CN110843746A (en) * 2019-11-28 2020-02-28 的卢技术有限公司 Anti-lock brake control method and system based on reinforcement learning
CN112201069A (en) * 2020-09-25 2021-01-08 厦门大学 Deep reinforcement learning-based method for constructing longitudinal following behavior model of driver
CN112614344A (en) * 2020-12-14 2021-04-06 中汽研汽车试验场股份有限公司 Hybrid traffic system efficiency evaluation method for automatic driving automobile participation
CN112677984A (en) * 2019-10-18 2021-04-20 丰田自动车株式会社 Method for generating vehicle control data, vehicle control device, and vehicle control system
WO2021073523A1 (en) * 2019-10-15 2021-04-22 同济大学 Method for estimating road capacity and connected automatic driving vehicle equivalent coefficient
CN112698578A (en) * 2019-10-22 2021-04-23 北京车和家信息技术有限公司 Automatic driving model training method and related equipment
WO2021165113A1 (en) * 2020-02-17 2021-08-26 Psa Automobiles Sa Method for training at least one algorithm for a control device of a motor vehicle, method for optimising traffic flow in a region, computer program product, and motor vehicle
CN113353102A (en) * 2021-07-08 2021-09-07 重庆大学 Unprotected left-turn driving control method based on deep reinforcement learning
TWI745120B (en) * 2019-10-18 2021-11-01 日商豐田自動車股份有限公司 Vehicle control system, vehicle control device, and control method for a vehicle
CN113901718A (en) * 2021-10-11 2022-01-07 长安大学 Deep reinforcement learning-based driving collision avoidance optimization method in following state
CN113954865A (en) * 2021-09-22 2022-01-21 吉林大学 Following control method for automatically driven vehicle in ice and snow environment
CN113954874A (en) * 2021-11-03 2022-01-21 同济大学 Automatic driving control method based on improved intelligent driver model
CN114056332A (en) * 2022-01-14 2022-02-18 清华大学 Intelligent automobile following decision and control method based on cognitive risk balance
CN115123159A (en) * 2022-06-27 2022-09-30 重庆邮电大学 AEB control method and system based on DDPG deep reinforcement learning

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101606794A (en) * 2009-07-17 2009-12-23 梁秀芬 A kind of dynamic cinema seat equipment
US20100185369A1 (en) * 2009-01-19 2010-07-22 Jung-Woong Choi Automatic transmission
CN102955884A (en) * 2012-11-23 2013-03-06 同济大学 Safety distance calibration method in full-speed areas during following operation of high-speed train
CN103101559A (en) * 2013-02-16 2013-05-15 同济大学 Full-speed field train interval real-time control method based on car-following behavior quality evaluation
CN103248545A (en) * 2013-05-28 2013-08-14 北京和利时电机技术有限公司 Ethernetcommunication method and system for special effect broadcast system of dynamic cinema
CN105654779A (en) * 2016-02-03 2016-06-08 北京工业大学 Expressway construction area traffic flow coordination control method based on vehicle-road and vehicle-vehicle communication
CN106926844A (en) * 2017-03-27 2017-07-07 西南交通大学 A kind of dynamic auto driving lane-change method for planning track based on real time environment information
CN108313054A (en) * 2018-01-05 2018-07-24 北京智行者科技有限公司 The autonomous lane-change decision-making technique of automatic Pilot and device and automatic driving vehicle
CN108387242A (en) * 2018-02-07 2018-08-10 西南交通大学 Automatic Pilot lane-change prepares and executes integrated method for planning track
CN108492398A (en) * 2018-02-08 2018-09-04 同济大学 The method for early warning that drive automatically behavior based on accelerometer actively acquires
CN108932840A (en) * 2018-07-17 2018-12-04 北京理工大学 Automatic driving vehicle urban intersection passing method based on intensified learning

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100185369A1 (en) * 2009-01-19 2010-07-22 Jung-Woong Choi Automatic transmission
CN101606794A (en) * 2009-07-17 2009-12-23 梁秀芬 A kind of dynamic cinema seat equipment
CN102955884A (en) * 2012-11-23 2013-03-06 同济大学 Safety distance calibration method in full-speed areas during following operation of high-speed train
CN103101559A (en) * 2013-02-16 2013-05-15 同济大学 Full-speed field train interval real-time control method based on car-following behavior quality evaluation
CN103248545A (en) * 2013-05-28 2013-08-14 北京和利时电机技术有限公司 Ethernetcommunication method and system for special effect broadcast system of dynamic cinema
CN105654779A (en) * 2016-02-03 2016-06-08 北京工业大学 Expressway construction area traffic flow coordination control method based on vehicle-road and vehicle-vehicle communication
CN106926844A (en) * 2017-03-27 2017-07-07 西南交通大学 A kind of dynamic auto driving lane-change method for planning track based on real time environment information
CN108313054A (en) * 2018-01-05 2018-07-24 北京智行者科技有限公司 The autonomous lane-change decision-making technique of automatic Pilot and device and automatic driving vehicle
CN108387242A (en) * 2018-02-07 2018-08-10 西南交通大学 Automatic Pilot lane-change prepares and executes integrated method for planning track
CN108492398A (en) * 2018-02-08 2018-09-04 同济大学 The method for early warning that drive automatically behavior based on accelerometer actively acquires
CN108932840A (en) * 2018-07-17 2018-12-04 北京理工大学 Automatic driving vehicle urban intersection passing method based on intensified learning

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
MEIXIN ZHU,等: "Modeling car-following behavior on urban expressways in Shanghai: A naturalistic driving study", 《TRANSPORTATION RESEARCH PART C》 *
XUESONG WANG,等: "Drivers’ rear end collision avoidance behaviors under different levels of situational urgency", 《TRANSPORTATION RESEARCH PART C》 *
王雪松,朱美新,邢祎伦: "基于自然驾驶数据的避撞预警对跟车行为影响", 《同济大学学报(自然科学版)》 *
王雪松,朱美新,陈铭: "驾驶员前向避撞行为特征的降维及多元方差分析", 《同济大学学报(自然科学版)》 *
王雪松,等: "中美两国道路交通事故信息采集技术比较研究", 《中国安全科学学报》 *

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110321605A (en) * 2019-06-19 2019-10-11 中汽研(天津)汽车工程研究院有限公司 A kind of human-computer interaction coordination control strategy based on Multiple Velocity Model PREDICTIVE CONTROL
CN110347043A (en) * 2019-07-15 2019-10-18 武汉天喻信息产业股份有限公司 A kind of intelligent driving control method and device
CN110347043B (en) * 2019-07-15 2023-03-10 武汉天喻信息产业股份有限公司 Intelligent driving control method and device
CN110488802A (en) * 2019-08-21 2019-11-22 清华大学 A kind of automatic driving vehicle dynamic behaviour decision-making technique netted under connection environment
CN110716562A (en) * 2019-09-25 2020-01-21 南京航空航天大学 Decision-making method for multi-lane driving of unmanned vehicle based on reinforcement learning
WO2021073523A1 (en) * 2019-10-15 2021-04-22 同济大学 Method for estimating road capacity and connected automatic driving vehicle equivalent coefficient
TWI745120B (en) * 2019-10-18 2021-11-01 日商豐田自動車股份有限公司 Vehicle control system, vehicle control device, and control method for a vehicle
CN112677984A (en) * 2019-10-18 2021-04-20 丰田自动车株式会社 Method for generating vehicle control data, vehicle control device, and vehicle control system
EP3809340A1 (en) * 2019-10-18 2021-04-21 Toyota Jidosha Kabushiki Kaisha Method of generating vehicle control data, vehicle control device, and vehicle control system
US11673556B2 (en) 2019-10-18 2023-06-13 Toyota Jidosha Kabushiki Kaisha Method of generating vehicle control data, vehicle control device, and vehicle control system
AU2020256407B2 (en) * 2019-10-18 2022-03-03 Toyota Jidosha Kabushiki Kaisha Method of generating vehicle control data, vehicle control device, and vehicle control system
CN112677984B (en) * 2019-10-18 2024-08-13 丰田自动车株式会社 Method for generating control data for vehicle, control device for vehicle, and control system for vehicle
CN112698578A (en) * 2019-10-22 2021-04-23 北京车和家信息技术有限公司 Automatic driving model training method and related equipment
CN112698578B (en) * 2019-10-22 2023-11-14 北京车和家信息技术有限公司 Training method of automatic driving model and related equipment
CN110843746A (en) * 2019-11-28 2020-02-28 的卢技术有限公司 Anti-lock brake control method and system based on reinforcement learning
WO2021165113A1 (en) * 2020-02-17 2021-08-26 Psa Automobiles Sa Method for training at least one algorithm for a control device of a motor vehicle, method for optimising traffic flow in a region, computer program product, and motor vehicle
CN112201069A (en) * 2020-09-25 2021-01-08 厦门大学 Deep reinforcement learning-based method for constructing longitudinal following behavior model of driver
CN112201069B (en) * 2020-09-25 2021-10-29 厦门大学 Deep reinforcement learning-based method for constructing longitudinal following behavior model of driver
CN112614344B (en) * 2020-12-14 2022-03-29 中汽研汽车试验场股份有限公司 Hybrid traffic system efficiency evaluation method for automatic driving automobile participation
CN112614344A (en) * 2020-12-14 2021-04-06 中汽研汽车试验场股份有限公司 Hybrid traffic system efficiency evaluation method for automatic driving automobile participation
CN113353102A (en) * 2021-07-08 2021-09-07 重庆大学 Unprotected left-turn driving control method based on deep reinforcement learning
CN113353102B (en) * 2021-07-08 2022-11-25 重庆大学 Unprotected left-turn driving control method based on deep reinforcement learning
CN113954865A (en) * 2021-09-22 2022-01-21 吉林大学 Following control method for automatically driven vehicle in ice and snow environment
CN113954865B (en) * 2021-09-22 2023-11-10 吉林大学 Following control method for automatic driving vehicle in ice and snow environment
CN113901718A (en) * 2021-10-11 2022-01-07 长安大学 Deep reinforcement learning-based driving collision avoidance optimization method in following state
CN113954874A (en) * 2021-11-03 2022-01-21 同济大学 Automatic driving control method based on improved intelligent driver model
CN114056332B (en) * 2022-01-14 2022-04-12 清华大学 Intelligent automobile following decision and control method based on cognitive risk balance
CN114056332A (en) * 2022-01-14 2022-02-18 清华大学 Intelligent automobile following decision and control method based on cognitive risk balance
CN115123159A (en) * 2022-06-27 2022-09-30 重庆邮电大学 AEB control method and system based on DDPG deep reinforcement learning

Also Published As

Publication number Publication date
CN109709956B (en) 2021-06-08

Similar Documents

Publication Publication Date Title
CN109709956A (en) A kind of automatic driving vehicle speed control multiple-objection optimization with algorithm of speeding
CN109733415B (en) Anthropomorphic automatic driving and following model based on deep reinforcement learning
CN108595823B (en) Autonomous main vehicle lane changing strategy calculation method combining driving style and game theory
CN102109821B (en) System and method for controlling adaptive cruise of vehicles
CN109726804B (en) Intelligent vehicle driving behavior personification decision-making method based on driving prediction field and BP neural network
CN107168303A (en) A kind of automatic Pilot method and device of automobile
CN112906126B (en) Vehicle hardware in-loop simulation training system and method based on deep reinforcement learning
CN111409648B (en) Driving behavior analysis method and device
CN110956851B (en) Intelligent networking automobile cooperative scheduling lane changing method
CN115457782B (en) Automatic driving vehicle intersection conflict-free cooperation method based on deep reinforcement learning
Ni et al. A unified perspective on traffic flow theory, part III: validation and benchmarking
CN113901718A (en) Deep reinforcement learning-based driving collision avoidance optimization method in following state
CN114896869A (en) Automatic driving test scene generation method based on personalized driver model
CN115257789A (en) Decision-making method for side anti-collision driving of commercial vehicle in urban low-speed environment
CN117610681A (en) Automatic driving automobile decision-making method based on imitation learning and discrete reinforcement learning
CN114148349B (en) Vehicle personalized following control method based on generation of countermeasure imitation study
CN114492043B (en) Personalized driver following modeling method considering perception limited characteristics
CN114802306A (en) Intelligent vehicle integrated decision-making system based on man-machine co-driving concept
CN117877245A (en) Novel heterogeneous mixed traffic flow model grading evaluation and construction method
Guo et al. Modeling, learning and prediction of longitudinal behaviors of human-driven vehicles by incorporating internal human DecisionMaking process using inverse model predictive control
Teng et al. Car following model based on driving risk field for vehicle infrastructure cooperation
CN112801149B (en) Multi-vehicle-queue control method based on deep reinforcement learning
Liu et al. Personalized Automatic Driving System Based on Reinforcement Learning Technology
Wei et al. A learning-based autonomous driver: emulate human driver's intelligence in low-speed car following
CN116127853A (en) Unmanned driving overtaking decision method based on DDPG (distributed data base) with time sequence information fused

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant