CN109491377B - DP and QP based decision and planning for autonomous vehicles - Google Patents
DP and QP based decision and planning for autonomous vehicles Download PDFInfo
- Publication number
- CN109491377B CN109491377B CN201810745836.2A CN201810745836A CN109491377B CN 109491377 B CN109491377 B CN 109491377B CN 201810745836 A CN201810745836 A CN 201810745836A CN 109491377 B CN109491377 B CN 109491377B
- Authority
- CN
- China
- Prior art keywords
- path
- cost
- speed
- profile
- obstacle
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000005457 optimization Methods 0.000 claims abstract description 34
- 230000006870 function Effects 0.000 claims description 83
- 238000000034 method Methods 0.000 claims description 82
- 238000012545 processing Methods 0.000 claims description 55
- 230000001133 acceleration Effects 0.000 claims description 20
- 230000015654 memory Effects 0.000 claims description 18
- 230000036461 convulsion Effects 0.000 claims description 6
- 230000008569 process Effects 0.000 description 55
- 238000010586 diagram Methods 0.000 description 27
- 230000008447 perception Effects 0.000 description 23
- 230000008859 change Effects 0.000 description 16
- 238000004891 communication Methods 0.000 description 13
- 230000033001 locomotion Effects 0.000 description 12
- 238000004422 calculation algorithm Methods 0.000 description 9
- 230000002085 persistent effect Effects 0.000 description 7
- 238000010801 machine learning Methods 0.000 description 5
- 238000007405 data analysis Methods 0.000 description 4
- 238000009499 grossing Methods 0.000 description 4
- 230000003287 optical effect Effects 0.000 description 4
- 230000001413 cellular effect Effects 0.000 description 3
- 238000013459 approach Methods 0.000 description 2
- 230000005540 biological transmission Effects 0.000 description 2
- 238000010276 construction Methods 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000012886 linear function Methods 0.000 description 2
- 230000005291 magnetic effect Effects 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 230000004931 aggregating effect Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000004888 barrier function Effects 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000010267 cellular communication Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 235000019800 disodium phosphate Nutrition 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000009429 electrical wiring Methods 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 229910000078 germane Inorganic materials 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012887 quadratic function Methods 0.000 description 1
- 230000010076 replication Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000004043 responsiveness Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000001953 sensory effect Effects 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 238000010897 surface acoustic wave method Methods 0.000 description 1
- 238000010408 sweeping Methods 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0212—Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W30/00—Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
- B60W30/08—Active safety systems predicting or avoiding probable or impending collision or attempting to minimise its consequences
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W30/00—Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
- B60W30/10—Path keeping
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W30/00—Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
- B60W30/14—Adaptive cruise control
- B60W30/16—Control of distance between vehicles, e.g. keeping a distance to preceding vehicle
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W50/0098—Details of control systems ensuring comfort, safety or stability not otherwise provided for
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W60/00—Drive control systems specially adapted for autonomous road vehicles
- B60W60/001—Planning or execution of driving tasks
- B60W60/0011—Planning or execution of driving tasks involving control alternatives for a single driving scenario, e.g. planning several paths to avoid obstacles
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/0055—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots with safety arrangements
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/0088—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots characterized by the autonomous decision making process, e.g. artificial intelligence, predefined behaviours
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0212—Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory
- G05D1/0217—Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory in accordance with energy consumption, time reduction or distance reduction criteria
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0231—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0231—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
- G05D1/0238—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using obstacle or wall sensors
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0231—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
- G05D1/0246—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using a video camera in combination with image processing means
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0257—Control of position or course in two dimensions specially adapted to land vehicles using a radar
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0268—Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means
- G05D1/027—Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means comprising intertial navigation means, e.g. azimuth detector
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0268—Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means
- G05D1/0274—Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means using mapping information stored in a memory device
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0276—Control of position or course in two dimensions specially adapted to land vehicles using signals provided by a source external to the vehicle
- G05D1/0278—Control of position or course in two dimensions specially adapted to land vehicles using signals provided by a source external to the vehicle using satellite positioning signals, e.g. GPS
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/01—Detecting movement of traffic to be counted or controlled
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W2050/0001—Details of the control system
- B60W2050/0002—Automatic control, details of type of controller or control system architecture
- B60W2050/0013—Optimal controllers
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2554/00—Input parameters relating to objects
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2556/00—Input parameters relating to data
- B60W2556/45—External transmission of data to or from the vehicle
- B60W2556/50—External transmission of data to or from the vehicle of positioning data, e.g. GPS [Global Positioning System] data
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2720/00—Output or target parameters relating to overall vehicle dynamics
- B60W2720/10—Longitudinal speed
- B60W2720/103—Speed profile
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/16—Anti-collision systems
Landscapes
- Engineering & Computer Science (AREA)
- Automation & Control Theory (AREA)
- Physics & Mathematics (AREA)
- Radar, Positioning & Navigation (AREA)
- General Physics & Mathematics (AREA)
- Remote Sensing (AREA)
- Aviation & Aerospace Engineering (AREA)
- Transportation (AREA)
- Mechanical Engineering (AREA)
- Electromagnetism (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- Theoretical Computer Science (AREA)
- Business, Economics & Management (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Game Theory and Decision Science (AREA)
- Medical Informatics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Traffic Control Systems (AREA)
- Control Of Driving Devices And Active Controlling Of Vehicle (AREA)
Abstract
According to some embodiments, the system calculates the first trajectory based on the map and the route information. The system generates a path profile based on the first trajectory, the traffic rules, and the obstacle information describing the one or more obstacles perceived by the ADV. The system generates a speed profile based on the path profile, wherein the speed profile includes, for each obstacle of the plurality of obstacles, a decision to yield or exceed the obstacle. The system performs quadratic programming optimization on the path profile and the speed profile to identify the best path with the best speed. The system generates a second trajectory based on the optimal path and the optimal speed to autonomously control the ADV according to the second trajectory.
Description
Technical Field
Embodiments of the present disclosure generally relate to operating an autonomous vehicle. More particularly, embodiments of the present disclosure relate to Dynamic Programming (DP) and Quadratic Programming (QP) based decision and planning for autonomous vehicles (ADV).
Background
Vehicles operating in an autonomous driving mode (e.g., unmanned) may relieve passengers, particularly the driver, from some driving-related responsibilities. When operating in an autonomous driving mode, the vehicle may be navigated to various locations using onboard sensors, allowing the vehicle to travel with minimal human interaction or in some cases without any passengers.
Motion planning and control are key operations in autonomous driving. However, motion planning and control may be open and may be difficult to optimize without some initial constraints. Furthermore, motion planning and control is applied in some cases to all types of vehicles that may be inaccurate and unstable.
Disclosure of Invention
A computer-implemented method of generating a driving trajectory for an autonomous vehicle is provided. The method comprises the following steps: calculating a first trajectory based on the map and the route information; generating a path profile based on the first trajectory, traffic rules, and obstacle information describing one or more obstacles perceived by the autonomous vehicle; generating a speed profile based on the path profile, wherein the speed profile includes, for each of the plurality of obstacles, a decision to make or exceed the obstacle; performing quadratic programming optimization on the path profile and the speed profile to identify an optimal path having an optimal speed; and generating a second trajectory based on the optimal path and the optimal speed to autonomously control the autonomous vehicle in accordance with the second trajectory.
According to an embodiment of the application, the path profile and the speed profile are iteratively generated using dynamic programming.
According to an embodiment of the application, the path profile comprises, for each obstacle decision encountered, a decision to let go, ignore or skim from the left or right of the encountered obstacle.
According to an embodiment of the present application, performing quadratic programming optimization on the path profile and the speed profile comprises: optimizing a first cost function using quadratic programming to generate a base point-side shift map based on the path profile; and optimizing a second cost function using quadratic programming to generate a base point-time map based on the speed profile.
According to an embodiment of the application, the base-side shift map is generated by forming one or more obstacles based on one or more obstacle decisions.
According to an embodiment of the application, the first cost function comprises a heading cost, a curvature cost and a distance cost.
According to an embodiment of the application, the second cost function comprises an acceleration cost, a jerk cost and a distance cost.
According to an embodiment of the present application, further comprising inserting a plurality of points, which are not present in the first track, into the second track based on the base point-side shift map and the base point-time map.
The present application also provides a non-transitory machine-readable medium having instructions stored thereon. The instructions, when executed by a processor, cause the processor to perform operations comprising: calculating a first trajectory based on the map and the route information; generating a path profile based on the first trajectory, traffic rules, and obstacle information describing one or more obstacles perceived by the autonomous vehicle; generating a speed profile based on the path profile, wherein the speed profile includes, for each of the plurality of obstacles, a decision to make or exceed the obstacle; performing quadratic programming optimization on the path profile and the speed profile to identify an optimal path having an optimal speed; and generating a second trajectory based on the optimal path and the optimal speed to autonomously control the autonomous vehicle in accordance with the second trajectory.
The present application also provides a data processing system, comprising: a processor; and a memory coupled to the processor to store instructions that, when executed by the processor, cause the processor to perform operations comprising: calculating a first trajectory based on the map and the route information; generating a path profile based on the first trajectory, traffic rules, and obstacle information describing one or more obstacles perceived by the autonomous vehicle; generating a speed profile based on the path profile, wherein the speed profile includes, for each of the plurality of obstacles, a decision to make or exceed the obstacle; performing quadratic programming optimization on the path profile and the speed profile to identify an optimal path having an optimal speed; and generating a second trajectory based on the optimal path and the optimal speed to autonomously control the autonomous vehicle in accordance with the second trajectory.
Drawings
Embodiments of the present disclosure are illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings in which like reference numerals refer to similar elements.
FIG. 1 is a block diagram illustrating a networked system according to one embodiment.
FIG. 2 is a block diagram illustrating an example of an autonomous vehicle according to one embodiment.
Fig. 3A-3B are block diagrams illustrating an example of a perception and planning system for use with an autonomous vehicle, according to one embodiment.
FIG. 4 is a block diagram illustrating an example of a decision-making process and planning process in accordance with one embodiment.
FIG. 5A is a block diagram illustrating an example of a decision module according to one embodiment.
FIG. 5B is a block diagram illustrating an example of a planning module, according to one embodiment.
Fig. 6 is a block diagram illustrating a base-side shift map according to one embodiment.
Fig. 7A and 7B are block diagrams illustrating a base point-time diagram according to some embodiments.
FIG. 8 is a flow chart illustrating a method according to one embodiment.
FIG. 9 is a block diagram illustrating an example of a planning module, according to one embodiment.
FIG. 10 is a flow chart illustrating a method according to one embodiment.
FIG. 11 is a block diagram illustrating an example of a planning module, according to one embodiment.
Fig. 12A-12B are block diagrams illustrating a path cost module and a speed cost module, respectively, according to one embodiment.
FIG. 13 is a flow chart illustrating a method according to one embodiment.
FIG. 14A is a flow chart illustrating a method according to one embodiment.
FIG. 14B is a flow chart illustrating a method according to one embodiment.
FIG. 15 is a block diagram illustrating a data processing system in accordance with one embodiment.
Detailed Description
Various embodiments and aspects of the disclosure will be described with reference to details discussed below, and the accompanying drawings will illustrate the various embodiments. The following description and drawings are illustrative of the disclosure and are not to be construed as limiting the disclosure. Numerous specific details are described to provide a thorough understanding of various embodiments of the present disclosure. However, in certain instances, well-known or conventional details are not described in order to provide a concise discussion of embodiments of the present disclosure.
Reference in the specification to "one embodiment" or "an embodiment" means that a particular feature, structure, or characteristic described in connection with the embodiment can be included in at least one embodiment of the disclosure. The phrase "in one embodiment" is not necessarily all described in various places throughout this specification to refer to the same embodiment.
According to some embodiments, the ADV includes a decision and planning system that autonomously controls the ADV. Based on the starting location and the ending location, the system queries a routing service for a reference route and calculates a reference line for the ADV to drive from the starting location to the ending location. Based on perceived obstacles and traffic conditions around the ADV, the system determines a path decision and a speed decision to ignore, pass, yield, or exceed the obstacle according to traffic rules and/or conditions of the ADV. The system optimizes one or more reference lines as trajectories based on path decisions and speed decisions to plan when and where the vehicle should be at a particular point in time.
According to one aspect, the system calculates a first trajectory based on the map and route information. The system generates a Path Profile (Path Profile) based on the first trajectory, the traffic rules, and the obstacle information describing the one or more obstacles perceived by the ADV. The system generates a speed Profile (Space Profile) based on the path Profile, wherein the speed Profile includes, for each obstacle, a decision to yield or exceed the obstacle. The system performs quadratic programming optimization on the path profile and the velocity profile to identify an optimal path having an optimal velocity, and generates a second trajectory based on the optimal path and the optimal velocity, such that the ADV may be autonomously controlled based on the second trajectory. The second trajectory represents the first trajectory optimized using quadratic programming optimization.
According to another aspect, the system calculates a first trajectory based on the map and the route information. The system generates a path profile based on the first trajectory, the traffic rules, and obstacle information describing one or more obstacles perceived by the ADV, wherein the path profile includes, for each obstacle, a decision to give way or sweep from the left or right side of the obstacle. The system generates a speed profile based on the path profile according to the traffic rules. The system performs a gradient descent optimization based on the path profile and the velocity profile to generate a second trajectory representing the optimized first trajectory, and controls the ADV according to the second trajectory.
According to yet another aspect, the system generates a plurality of feasible decisions for tracking the ADV from the first location to the second location according to a set of traffic rules based on perception information that perceives a driving environment (including one or more obstacles) surrounding the ADV. The system calculates a plurality of trajectories based on a combination of one or more feasible decisions. The system calculates a total cost for each trajectory using a plurality of cost functions and autonomously controls the ADV by selecting the trajectory having the smallest total cost as the driving trajectory. The cost functions include a path cost function, a speed cost function, and an obstacle cost function.
Fig. 1 is a block diagram illustrating an autonomous vehicle network configuration according to one embodiment of the present disclosure. Referring to fig. 1, a network configuration 100 includes an autonomous vehicle 101 that may be communicatively coupled to one or more servers 103-104 through a network 102. Although one autonomous vehicle is shown, multiple autonomous vehicles may be coupled to each other and/or to servers 103-104 through network 102. The network 102 may be any type of network, such as a wired or wireless Local Area Network (LAN), a Wide Area Network (WAN) such as the Internet, a cellular network, a satellite network, or a combination thereof. The servers 103-104 may be any type of server or cluster of servers, such as a network or cloud server, an application server, a backend server, or a combination thereof. The servers 103 to 104 may be data analysis servers, content servers, traffic information servers, map and point of interest (MPOI) servers, or location servers, etc.
Autonomous vehicles refer to vehicles that may be configured to be in an autonomous driving mode in which the vehicle navigates through the environment with little or no input from the driver. Such autonomous vehicles may include a sensor system having one or more sensors configured to detect information related to the operating environment of the vehicle. The vehicle and its associated controller use the detected information to navigate through the environment. Autonomous vehicle 101 may operate in a manual mode, in a fully autonomous mode, or in a partially autonomous mode.
In one embodiment, autonomous vehicle 101 includes, but is not limited to, a perception and planning system 110, a vehicle control system 111, a wireless communication system 112, a user interface system 113, and a sensor system 115. Autonomous vehicle 101 may also include certain common components included in a common vehicle, such as: engines, wheels, steering wheels, transmissions, etc., which may be controlled by the vehicle control system 111 and/or the sensory and programming system 110 using a variety of communication signals and/or commands, such as, for example, acceleration signals or commands, deceleration signals or commands, steering signals or commands, braking signals or commands, etc.
The components 110-115 may be communicatively coupled to each other via an interconnect, bus, network, or combination thereof. For example, the components 110-115 may be communicatively coupled to one another via a Controller Area Network (CAN) bus. The CAN bus is a vehicle bus standard designed to allow microcontrollers and devices to communicate with each other in applications without a host. It is a message-based protocol originally designed for multiplexed electrical wiring within automobiles, but is also used in many other environments.
Referring now to fig. 2, in one embodiment, the sensor system 115 includes, but is not limited to, one or more cameras 211, a Global Positioning System (GPS) unit 212, an Inertial Measurement Unit (IMU)213, a radar unit 214, and a light detection and ranging (LIDAR) unit 215. The GPS system 212 may include a transceiver operable to provide information regarding the location of the autonomous vehicle. The IMU unit 213 may sense position and orientation changes of the autonomous vehicle based on inertial acceleration. Radar unit 214 may represent a system that utilizes radio signals to sense objects within the local environment of an autonomous vehicle. In some embodiments, in addition to sensing an object, radar unit 214 may additionally sense a speed and/or heading of the object. The LIDAR unit 215 may use a laser to sense objects in the environment in which the autonomous vehicle is located. The LIDAR unit 215 may include one or more laser sources, laser scanners, and one or more detectors, among other system components. The camera 211 may include one or more devices used to capture images of the environment surrounding the autonomous vehicle. The camera 211 may be a still camera and/or a video camera. The camera may be mechanically movable, for example, by mounting the camera on a rotating and/or tilting platform.
The sensor system 115 may also include other sensors, such as: sonar sensors, infrared sensors, steering sensors, throttle sensors, brake sensors, and audio sensors (e.g., microphones). The audio sensor may be configured to collect sound from an environment surrounding the autonomous vehicle. The steering sensor may be configured to sense a steering angle of a steering wheel, wheels of a vehicle, or a combination thereof. The throttle sensor and the brake sensor sense a throttle position and a brake position of the vehicle, respectively. In some cases, the throttle sensor and the brake sensor may be integrated into an integrated throttle/brake sensor.
In one embodiment, the vehicle control system 111 includes, but is not limited to, a steering unit 201, a throttle unit 202 (also referred to as an acceleration unit), and a brake unit 203. The steering unit 201 is used to adjust the direction or forward direction of the vehicle. The throttle unit 202 is used to control the speed of the motor or engine, which in turn controls the speed and acceleration of the vehicle. The brake unit 203 decelerates the vehicle by providing friction to decelerate the wheels or tires of the vehicle. It should be noted that the components shown in fig. 2 may be implemented in hardware, software, or a combination thereof.
Returning to fig. 1, wireless communication system 112 allows communication between autonomous vehicle 101 and external systems such as devices, sensors, other vehicles, and the like. For example, the wireless communication system 112 may be in direct wireless communication with one or more devices, or in wireless communication via a communication network, such as with the servers 103-104 through the network 102. The wireless communication system 112 may use any cellular communication network or Wireless Local Area Network (WLAN), for example, using WiFi, to communicate with another component or system. The wireless communication system 112 may communicate directly with devices (e.g., passenger's mobile device, display device, speaker within the vehicle 101), for example, using infrared links, bluetooth, etc. The user interface system 113 may be part of a peripheral device implemented within the vehicle 101, including, for example, a keypad, a touch screen display device, a microphone, and speakers, among others.
Some or all of the functions of the autonomous vehicle 101 may be controlled or managed by the perception and planning system 110, particularly when operating in an autonomous mode. The awareness and planning system 110 includes the necessary hardware (e.g., processors, memory, storage devices) and software (e.g., operating systems, planning and routing programs) to receive information from the sensor system 115, the control system 111, the wireless communication system 112, and/or the user interface system 113, process the received information, plan a route or path from the origin to the destination, and then drive the vehicle 101 based on the planning and control information. Alternatively, the sensing and planning system 110 may be integrated with the vehicle control system 111.
For example, a user who is a passenger may specify a start location and a destination of a trip, e.g., via a user interface. The perception and planning system 110 obtains trip related data. For example, the sensing and planning system 110 may obtain location and route information from an MPOI server, which may be part of the servers 103-104. The location server provides location services and the MPOI server provides map services and POIs for certain locations. Alternatively, such location and MPOI information may be cached locally in persistent storage of the sensing and planning system 110.
The perception and planning system 110 may also obtain real-time traffic information from a traffic information system or server (TIS) as the autonomous vehicle 101 moves along the route. It should be noted that the servers 103 to 104 may be operated by third party entities. Alternatively, the functionality of the servers 103-104 may be integrated with the perception and planning system 110. Based on the real-time traffic information, MPOI information, and location information, as well as real-time local environmental data (e.g., obstacles, objects, nearby vehicles) detected or sensed by sensor system 115, perception and planning system 110 may plan an optimal route and drive vehicle 101, e.g., via control system 111, according to the planned route to safely and efficiently reach a designated destination.
Server 103 may be a data analysis system that performs data analysis services for a variety of clients. In one embodiment, data analysis system 103 includes a data collector 121 and a machine learning engine 122. The data collector 121 collects driving statistics 123 from various vehicles (autonomous vehicles or regular vehicles driven by human drivers). The driving statistics 123 include information indicative of driving commands (e.g., throttle commands, brake commands, steering commands) issued at different points in time and the vehicle's response (e.g., speed, acceleration, deceleration, direction) captured by the vehicle's sensors. The driving statistics 123 may also include information describing the driving environment at different points in time, such as, for example, a route (including a start location and a destination location), an MPOI, road conditions, weather conditions, and so forth.
Based on the driving statistics 123, the machine learning engine 122 generates or trains a rule set, algorithm, and/or predictive model 124 for a variety of purposes. For example, server 103 may include routing service 125 to provide routing services (e.g., route and map information) to ADV 101. ADV 101 may request a reference route (e.g., an ideal route without obstacle information or traffic conditions) from routing service 125 by indicating a start location and an end location. The routing service 125 then returns the requested route. In one embodiment, returning to the reference route may include returning one or more tables, such as a reference point table and a road segment/lane table. Alternatively, route and map information may be downloaded and cached in the vehicle, which information may be used in real time.
The server 103 may generate a reference route, for example, the machine learning engine 122 may generate the reference route from map information (information such as a section of a road, a lane of the section of the road, and a distance from the lane to a curb). For example, a road may be divided into { A, B and C } sections or segments to represent three road segments. The three lanes of road segment A may be listed as { A1, A2, and A3 }. The reference route is generated by generating reference points along the reference route. For example, for a lane, the machine learning engine 122 may connect the midpoints of two opposing curbs or edges of the lane provided by the map data. Based on machine learning data representing collected data points at which the vehicle was driving on the lane at different points in time and the midpoint, the engine 122 may calculate the reference point by selecting a subset of the collected data points within a predetermined proximity of the lane and applying a smoothing function to the midpoint based on the subset of collected data points.
Based on the reference point or the lane reference point, the ADV receiving the reference point may generate a reference line by interpolating the reference point such that the generated reference line is used as a reference line for controlling the ADV on the lane. In some implementations, reference point tables and road segment tables representing reference lines are downloaded to the ADV in real-time so that the ADV can generate a reference line based on the geographical location and driving direction of the ADV. For example, in one embodiment, the ADV may generate the reference line by requesting routing services for the path segment using a path segment identifier representing the upcoming road segment ahead and/or an ADV-based GPS location. Based on the path segment identifier, the routing service may return to an ADV reference point table that contains reference points for all lanes of the road segment of interest. The ADV may consult a reference point of the lane of the path segment to generate a reference line for controlling the ADV on the lane. It should be noted that the above process is performed offline by the analysis server 103, wherein the reference point of the route is determined based on the route and the map information. However, the same data may be dynamically determined in real time within each vehicle, as will be described in further detail below.
Fig. 3A and 3B are block diagrams illustrating an example of a perception and planning system for use with an autonomous vehicle, according to one embodiment. The system 300 may be implemented as part of the autonomous vehicle 101 of fig. 1, including but not limited to the perception and planning system 110, the control system 111, and the sensor system 115. Referring to fig. 3A-3B, the awareness and planning system 110 includes, but is not limited to, a positioning module 301, an awareness module 302, a prediction module 303, a decision module 304, a planning module 305, a control module 306, and a routing module 307.
Some or all of modules 301 through 307 may be implemented in software, hardware, or a combination thereof. For example, the modules may be installed in persistent storage 352, loaded into memory 351, and executed by one or more processors (not shown). It should be noted that some or all of these modules may be communicatively coupled to or integrated with some or all of the modules of the vehicle control system 111 of fig. 2. Some of modules 301 to 307 may be integrated together into an integrated module. For example, the decision module 304 and the planning module 305 may be integrated into a single module.
The location module 301 determines the current location of the autonomous vehicle 300 (e.g., using the GPS unit 212) and manages any data related to the user's trip or route. The positioning module 301 (also referred to as a map and route module) manages any data related to the user's journey or route. The user may, for example, log in via a user interface and specify a starting location and a destination for the trip. The positioning module 301 communicates with other components of the autonomous vehicle 300, such as map and route information 311, to obtain trip related data. For example, the location module 301 may obtain location and route information from a location server and a map and poi (mpoi) server. The location server provides location services and the MPOI server provides map services and POIs for certain locations and may thus be cached as part of the map and route information 311. The location module 301 may also obtain real-time traffic information from a traffic information system or server as the autonomous vehicle 300 moves along the route.
Based on the sensor data provided by sensor system 115 and the positioning information obtained by positioning module 301, perception module 302 determines a perception of the surrounding environment. The perception information may represent what an average driver would perceive around the vehicle the driver is driving. Perception may include, for example, lane configuration in the form of an object (e.g., a straight lane or a curved lane), a traffic light signal, a relative position of another vehicle, a pedestrian, a building, a crosswalk, or other traffic-related indicia (e.g., a stop sign, a yield sign), and so forth.
The perception module 302 may include a computer vision system or functionality of a computer vision system to process and analyze images captured by one or more cameras to identify objects and/or features in an autonomous vehicle environment. The objects may include traffic signals, road boundaries, other vehicles, pedestrians, and/or obstacles, etc. Computer vision systems may use object recognition algorithms, video tracking, and other computer vision techniques. In some embodiments, the computer vision system may map the environment, track objects, and estimate the speed of objects, among other things. The perception module 302 may also detect objects based on other sensor data provided by other sensors, such as radar and/or LIDAR.
For each object, the prediction module 303 predicts the behavior of the object under the circumstances. The prediction is performed according to a set of map/route information 311 and traffic rules 312 based on perception data of the driving environment perceived at a certain point in time. For example, if the object is a vehicle at the opposite direction and the current driving environment includes an intersection, the prediction module 303 will predict whether the vehicle is likely to go straight ahead or likely to turn. If the perception data indicates that the intersection has no traffic lights, the prediction module 303 may predict that the vehicle may have to come to a complete stop before entering the intersection. If the perception data indicates that the vehicle is currently in a left-turn-only lane or a right-turn-only lane, the prediction module 303 may predict that the vehicle is more likely to make a left turn or a right turn, respectively.
For each subject, the decision module 304 makes a decision on how to treat the subject. For example, for a particular object (e.g., another vehicle in a crossing route) and metadata describing the object (e.g., speed, direction, turn angle), the decision module 304 decides how to encounter the object (e.g., cut, yield, stop, pass). The decision module 304 may make such a decision based on a rule set, such as traffic rules or driving rules 312, which may be stored in persistent storage 352.
Based on the decisions for each of the perceived objects, the planning module 305 plans a path or route and driving parameters (e.g., distance, speed, and/or turn angle) for the autonomous vehicle. In other words, for a given object, the decision module 304 decides what to do with the object, and the planning module 305 determines how to do. For example, for a given subject, the decision module 304 may decide to pass through the subject, while the planning module 305 may determine whether to pass on the left or right side of the subject. Planning and control data is generated by the planning module 305, including information describing how the vehicle 300 will move in the next movement cycle (e.g., the next route/path segment). For example, the planning and control data may instruct the vehicle 300 to move 10 meters at a speed of 30 miles per hour (mph), and then change to the right lane at a speed of 25 mph.
Based on the planning and control data, the control module 306 controls and drives the autonomous vehicle by sending appropriate commands or signals to the vehicle control system 111 according to the route or path defined by the planning and control data. The planning and control data includes sufficient information to drive the vehicle from a first point to a second point of the route or path at different points in time along the route or route using appropriate vehicle settings or driving parameters (e.g., throttle, brake, and turn commands).
In one embodiment, the planning phase is performed in a plurality of planning periods (also referred to as command periods), such as, for example, at intervals of every 100 milliseconds (ms). For each planning or command cycle, one or more control commands will be issued based on the planning and control data. That is, for every 100ms, the planning module 305 plans the next route segment or path segment, e.g., including the target location and the time required for the ADV to reach the target location. Alternatively, the planning module 305 may also specify a particular speed, direction, and/or steering angle, etc. In one embodiment, the planning module 305 plans a route segment or a path segment for the next predetermined time period (such as 5 seconds). For each planning cycle, the planning module 305 plans the target location for the current cycle (e.g., the next 5 seconds) based on the planned target location in the previous cycle. The control module 306 then generates one or more control commands (e.g., throttle control commands, brake control commands, steering control commands) based on the current cycle of the schedule and the control data.
It should be noted that the decision module 304 and the planning module 305 may be integrated as an integrated module. The decision module 304/planning module 305 may include a navigation system or functionality of a navigation system to determine a driving path of an autonomous vehicle. For example, the navigation system may determine a range of speeds and heading directions for enabling the autonomous vehicle to move along the following path: the path substantially avoids perceived obstacles while advancing the autonomous vehicle along a roadway-based path to a final destination. The destination may be set based on user input via the user interface system 113. The navigation system may dynamically update the driving path while the autonomous vehicle is running. The navigation system may combine data from the GPS system and one or more maps to determine a driving path for the autonomous vehicle.
The decision module 304/planning module 305 may also include a collision avoidance system or the functionality of a collision avoidance system to identify, assess, and avoid or otherwise overcome potential obstacles in the environment of the autonomous vehicle. For example, a collision avoidance system may implement a change in navigation of an autonomous vehicle by: one or more subsystems in the control system 111 are operated to take a turning maneuver, a braking maneuver, etc. The collision avoidance system may automatically determine a feasible obstacle avoidance maneuver based on surrounding traffic patterns, road conditions, and the like. The collision avoidance system may be configured such that no turn-changing maneuvers are taken when other sensor systems detect vehicles, building obstacles, etc. located in adjacent areas into which the autonomous vehicle will change direction. The collision avoidance system may automatically select maneuvers that are both available and that maximize the safety of the occupants of the autonomous vehicle. The collision avoidance system may select an avoidance maneuver that is predicted to cause a minimum amount of acceleration to occur in a passenger compartment of the autonomous vehicle.
The routing module 307 may generate a reference route, for example, from map information (such as road segments, lanes of the road segments, and distance from the lanes to the curbs). For example, a road may be divided into { A, B and C } sections or segments to represent three road segments. The three lanes of road segment A may be listed as { A1, A2, and A3 }. The reference route is generated by generating reference points along the reference route. For example, for a lane, the routing module 307 may connect the midpoints of two opposing curbs or edges of the lane provided by the map data. Based on machine-learned data representing collected data points at which the vehicle was driving on the lane at different points in time and the midpoint, the routing module 307 may calculate the reference point by selecting a subset of the collected data points within a predetermined proximity of the lane and applying a smoothing function to the midpoint based on the subset of collected data points.
Based on the reference point or the lane reference point, the routing module 307 may generate a reference line by interpolating the reference point such that the generated reference line is used as a reference line for controlling the ADV on the lane. In some implementations, a table of reference points and road segments representing a reference line may be downloaded to the ADV in real-time, such that the ADV may generate the reference line based on the geographical location and driving direction of the ADV. For example, in one embodiment, the ADV may generate the reference line by requesting routing services for the path segment using a path segment identifier representing the upcoming road segment ahead and/or an ADV-based GPS location. Based on the path segment identifier, the routing service may return to an ADV reference point table that contains reference points for all lanes of the road segment of interest. The ADV may consult a reference point of the lane of the path segment to generate a reference line for controlling the ADV on the lane.
As described above, the route or routing module 307 manages any data related to the user's itinerary or route. The user of the ADV specifies a start location and a destination location to obtain the trip related data. The trip-related data comprises route segments and reference lines or points of reference for the route segments. For example, based on the route map information 311, the route module 307 generates a route or link table and a reference point table. The reference points relate to road segments and/or lanes in the road segment table. The reference points may be interpolated to form one or more reference lines for controlling the ADV. The reference point may be limited to only the road segment and/or a particular lane of the road segment.
For example, the road segment table may be name-value pairs to include previous and next lanes of the road segments a-D. For example, for road segments a-D with lane 1, the road segment table may be: { (A1, B1), (B1, C1), (C1, D1) }. The reference point table may include reference points in the form of x-y coordinates for road segment lanes, such as { (a1, (x1, y1)), (B1, (x2, y2)), (C1, (x3, y3)), (D1, (x4, y4)) }, where a1 to D1 represent lane 1 of road segments a to D, and (x1, y1) to (x4, y4) are corresponding real world coordinates. In one embodiment, the road segments and/or lanes are divided into road segments/lanes of a predetermined length, such as approximately 200 meters. In another embodiment, road segments and/or lanes are divided into variable length segments/lanes according to road conditions such as the curvature of the road. In some embodiments, each road segment and/or lane may include several reference points. In some embodiments, the reference point may be transformed into another coordinate system, for example, latitude-longitude.
In some embodiments, the reference point may be transformed into a relative coordinate system, such as a base-side Shift (SL) coordinate. The base-side shift coordinate system is a coordinate system that follows a reference line with reference to a fixed reference point. For example, the (S, L) ═ 1, 0) coordinate may represent 1 meter before a base point (i.e., reference point) on a reference line with a lateral offset of 0 meters. The (S, L) — (2, 1) reference point may represent 2 meters ahead of the fixed reference point along the reference line and offset laterally by 1 meter from the reference line, e.g., by 1 meter to the left.
In one embodiment, the decision module 304 generates a coarse path profile based on reference lines provided by the routing module 307 and based on ADV-perceived obstacles around the ADV. The coarse path profile may be part of path/speed profile 313, which may be stored in persistent storage 352. A coarse path profile is generated by selecting points from the reference lines. For each point, the decision module 304 moves the point to the left or right of the reference line (e.g., candidate movement) based on one or more obstacle decisions on how to encounter the object, with the remaining points remaining stable. The candidate movement is iteratively performed using dynamic programming on the path candidates, wherein the path candidate with the smallest path cost is found using a cost function that is part of the cost function 315 of FIG. 3A, generating a coarse path profile. Examples of cost functions include costs based on: the curvature of the route path, the distance from the ADV to the perceived obstacle, and the distance from the ADV to the reference line. In one embodiment, the generated coarse route profile includes a base-side shift map that may be stored in persistent storage 352 as part of SL map/ST graphic 314.
In one embodiment, decision module 304 generates a coarse speed profile (as part of path/speed profile 313) based on the generated coarse path profile. The coarse velocity profile indicates the optimal velocity for controlling the ADV at a particular point in time. Similar to the coarse path profile, dynamic programming is used to iterate candidate velocities at different points in time, where velocity candidates (e.g., acceleration or deceleration) with minimum velocity costs are found based on a cost function (as part of cost function 315 of fig. 3A) according to the obstacles perceived by the ADV. The coarse speed profile determines whether the ADV should exceed or avoid the obstacle and steer to the left or right of the obstacle. In one embodiment, the coarse speed profile includes a base-time (ST) graphic (as part of SL map/ST graphic 314). The base-time graph shows distance traveled versus time.
In one embodiment, the planning module 305 recalculates the coarse path profile based on obstacle decisions and/or human obstacles to prohibit the planning module 305 from searching the geometric space of the obstacle. For example, if the coarse velocity profile determines that an obstacle is being swept from the left, the planning module 305 may place an obstacle (in the form of an obstacle) to the right of the obstacle to prevent calculation of an ADV sweeping the obstacle from the right. In one embodiment, the coarse path profile is recalculated by optimizing the path cost function (as part of the cost function 315) using Quadratic Programming (QP). In one embodiment, the recalculated coarse route profile includes a base-side shift map (as part of SL map/ST graphic 314).
In one embodiment, planning module 305 recalculates the coarse speed profile using Quadratic Planning (QP) to optimize the speed cost function (as part of cost function 315). Similar speed barrier constraints may be added to prohibit the QP solver from searching for some prohibited speeds. In one embodiment, the recalculated coarse velocity profile includes a base-point-time map (as part of SL map/ST graphic 314).
It should be noted that some or all of the components as shown and described above may be implemented in software, hardware, or a combination thereof. For example, such components may be implemented as software installed and stored in a persistent storage device, which may be loaded into and executed by a processor (not shown) in memory to implement the processes or operations described throughout this application. Alternatively, such components may be implemented as executable code programmed or embedded into dedicated hardware, such as an integrated circuit (e.g., an application specific integrated circuit or ASIC), a Digital Signal Processor (DSP) or Field Programmable Gate Array (FPGA), which is accessible via a respective driver and/or operating system from an application. Further, such components may be implemented as specific hardware logic within a processor or processor core as part of an instruction set accessible by software components through one or more specific instructions.
FIG. 4 is a block diagram illustrating an example of a decision and planning process in accordance with one embodiment. FIG. 5A is a block diagram illustrating an example of a decision module according to one embodiment. FIG. 5B is a block diagram illustrating an example of a planning module, according to one embodiment. Referring to fig. 4, the decision and planning process 400 includes a routing module 307, positioning/awareness data 401, a path decision process 403, a speed decision process 405, a path planning process 407, a speed planning process 409, an aggregator 411, and a trajectory calculator 413.
The path decision process 403 and the speed decision process 405 may be performed by the path decision module 501 and the speed decision module 503, respectively, of the decision module 304 in fig. 5A. Referring to fig. 4 and 5A, the route decision process 403 or route decision module 501 includes a route state machine 505, route traffic rules 507, and a base-side shift map generator 509. The path decision process 403 or the path decision module 501 may use dynamic planning to generate a coarse path profile as an initial constraint for the path planning process 407 and the speed planning process 409. In one embodiment, path state machine 505 includes at least three states: cruise status, lane change status, and idle status. The path state machine 505 provides previous planning results and important information such as whether the ADV is cruising or lane changing. The route traffic rules 507, which are part of the driving/traffic rules 312 of FIG. 3A, include traffic rules that can affect the results of the route decision module. For example, the route traffic rules 507 may include traffic information such as construction traffic signs, ADV, which may thereby avoid lanes having such construction signs. With status, traffic rules, reference lines provided by the routing module 307, and obstacles perceived by the ADV, the route decision process 403 may decide how to handle the perceived obstacles (i.e., ignore, exceed, yield, stop, pass) as part of a coarse route profile.
For example, in one embodiment, the coarse path profile is generated by a cost function that includes: cost based on curvature of the path; and a cost based on the distance from the reference line and/or the reference point to the obstacle. A point on the reference line is selected and moved to the left or right of the reference line as a candidate movement representing a path candidate. Each candidate movement has an associated cost. The associated costs of candidate moves of one or more points on the reference line can be solved sequentially (one point at a time) using dynamic programming to obtain the optimal cost. In one embodiment, SL map generator 509 generates a base-side shift map as part of a coarse path profile. The base-side shift map is a two-dimensional geometric map (similar to an x-y coordinate plane) that includes obstacle information perceived by the ADV. From the SL map, the routing decision process 403 may route an ADV route that follows the obstacle decision. Dynamic programming (or dynamic optimization) is a mathematical optimization method that decomposes the problem to be solved into a series of cost functions, solving only one of these cost functions at a time and storing its answer. The next time the same cost function occurs, only the previously computed answer needs to be looked up instead of recomputing the answer, saving computation time.
The speed decision process 405 or speed decision module 503 includes a speed state machine 511, a speed traffic rule 513, and a basepoint-time map generator 515. The speed decision process 405 or the speed decision module 503 may use dynamic programming to generate a coarse speed profile as an initial constraint for the path planning process 407 and the speed planning process 409. In one embodiment, the speed state machine 511 includes at least two states: acceleration or deceleration. The speed traffic rules 513, which are part of the driving/traffic rules 312 of fig. 3A, include traffic rules that can affect the results of the speed decision module. For example, the speed traffic rules 513 may include traffic information such as a red/green traffic light, another vehicle in an intersection route, and so on. Speed decision process 405 may generate a coarse speed profile to control when the ADV accelerates and/or decelerates through the state of the speed state machine, speed traffic rules, coarse path profile/SL map generated by decision process 403, and perceived obstacles. The base point-time map generator 515 may generate the base point-time map as part of a coarse velocity profile.
Referring to fig. 4 and 5B, path planning process 407 or path planning module 521 includes a radix-side shift map 525, a geometry smoother 527, and a path cost module 529. The base-side shift map 525 may include the base-side shift map generated by the SL map generator 509 of the route decision process 403. The path planning process 407 or the path planning module 521 may use a coarse path profile (e.g., a radix-side-shift map) as an initial constraint to recalculate the optimal reference line using quadratic planning. Quadratic programming involves minimizing or maximizing an objective function (e.g., a quadratic function with several variables) under the constraints of boundaries, linear equality, and inequality constraints. One difference between dynamic planning and quadratic planning is that quadratic planning optimizes all candidate moves for all points on the reference line at once. The geometry smoother 527 may apply a smoothing algorithm (such as B-spline or regression) to the output base-point-side shift map. The path cost module 529 may recalculate the reference line using the path cost function (as part of the cost function 315 of fig. 3A) to optimize the total cost of the candidate movement of the reference point, for example, using QP optimization performed by the QP module 540. For example, in one embodiment, the total path cost function may be:
where the path cost is summed over all points on a reference line, the heading represents the difference in radial angle (e.g., direction) between the points relative to the reference line, the curvature represents the difference between the curvature of the curve formed by the points relative to the reference line for that point, and the distance represents the lateral (perpendicular to the reference line) distance from the point to the reference line. In some embodiments, distance is the distance from the point to the destination location or an intermediate point of the reference line. In another embodiment, the curve cost is the change between the curvature values of the curves formed at adjacent points. It should be noted that a point on the reference line may be selected as a point having an equal distance to an adjacent point. Based on the path cost, the path cost module 529 may recalculate the reference lines by minimizing the path cost using quadratic programming optimization (e.g., by the QP module 540).
The speed planning process 409 or speed planning module 523 includes a base point-time map 531, a timing smoother 533, and a speed cost module 535. The base point-time graph 531 may include a base point-time (ST) graph generated by the ST graph generator 515 of the speed decision process 405. The speed planning process 409 or the speed planning module 523 may use the coarse speed profile (e.g., the radix point-time map) and the results of the path planning process 407 as initial constraints to calculate an optimal radix point-time curve. The timing smoother 533 may apply a smoothing algorithm (such as B-spline or regression) to the time series of points. The velocity cost module 535 may recalculate the ST graph using the velocity cost function (as part of the cost function 315 of fig. 3A) to optimize the total cost of the motion candidates (e.g., acceleration/deceleration) at different points in time. For example, in one embodiment, the total speed cost function may be:
where the velocity cost is summed over all time series points, speed' represents the acceleration value or the cost of changing the velocity between two adjacent points, speed "represents the jerk value, or the derivative of the acceleration value, or the cost of changing the velocity change between two adjacent points, and distance represents the distance of the ST point to the destination location. Here, the speed cost module 535 computes the basepoint-time map by minimizing the speed cost using quadratic programming optimization (e.g., by the QP module 540).
The aggregator 411 performs a function of aggregating the path planning result and the speed planning result. For example, in one embodiment, the aggregator 411 may combine a two-dimensional ST graphic and a SL map into a three-dimensional SLT graphic. In another embodiment, the aggregator 411 may interpolate (or fill with additional points) based on 2 consecutive points on the SL reference line or ST curve. In another embodiment, the aggregator 411 may convert the reference point from (S, L) coordinates to (x, y) coordinates. The trajectory generator 413 may calculate a final trajectory for controlling the ADV. For example, based on the SLT graph provided by the aggregator 411, the trajectory generator 413 computes a list of (x, y, T) points that indicate when the ADV should pass a particular (x, y) coordinate.
Thus, referring back to fig. 4, the path decision process 403 and the speed decision process 405 will generate a coarse path profile and a coarse speed profile taking into account obstacles and/or traffic conditions. Based on all path decisions and speed decisions related to the obstacle, the path planning process 407 and speed planning process 409 will optimize the coarse path profile and coarse speed profile using QP planning according to the obstacle to generate the best trajectory with the smallest path cost and/or speed cost. In another embodiment, the path cost and the speed cost may be calculated by the path cost module 1110 and the speed cost module 1120 of fig. 12A-12B, described further below.
FIG. 6 is a block diagram illustrating a base-side shift diagram according to one embodiment. Referring to fig. 6, the map 600 has an S horizontal axis (or base point) and an L vertical axis (or side shift). As described above, the base-side shift coordinate is a relative geometric coordinate system that is referenced to a specific base point on the reference line and follows the reference line. For example, the (S, L) ═ 1, 0) coordinate may represent 1 meter before the base point (i.e., the reference point) on the reference line and have a lateral offset of 0 meters. The (S, L) — (2, 1) reference point may represent 2 meters ahead of the fixed reference point along the reference line and offset vertically laterally from the reference line by 1 meter, e.g., left offset.
Referring to fig. 6, a map 600 includes a reference line 601 and obstacles 603 to 609 perceived by an ADV. In one embodiment, the obstacles 603-609 may be perceived by the RADAR or LIDAR unit of the ADV in different coordinate systems and may be translated to the SL coordinate system. In another embodiment, the obstacles 603-609 may be artificially created obstacles as constraints, such that the decision module and planning module do not search in constrained geometric space. In this example, a route decision module, such as route decision module 501, may generate a decision for each of obstacles 603-609, such as a decision to avoid obstacles 603-608 and to sweep (closely approach) obstacle 609 (i.e., these obstacles may be other cars, buildings, and/or structures). A path planning module, such as path planning module 521, may then recalculate or optimize the reference line 601 using QP planning based on the path costs in view of the obstacles 603-609 to optimize the reference line 601 with the smallest overall cost as described above. In this example, ADV 101 sweeps over (or approaches very closely to) obstacle 609 from the left side of obstacle 609.
Fig. 7A and 7B are block diagrams illustrating a base point-time diagram according to some embodiments. Referring to fig. 7A, a graph 700 has a base (or S) vertical axis and a time (or T) horizontal axis. The graph 700 includes a curve 701 and obstacles 703 to 707. As described above, the curve 701 on the base-point-time graph indicates time and the distance of the ADV from the base point. For example, (T, S) ═ 10000, 150) can represent 10000 milliseconds, and the ADV will be 150 meters from the base point (i.e., the reference point). In this example, the obstacle 703 may be a building/structure to be avoided, and the obstacle 707 may be a human obstacle corresponding to a decision to pass the moving vehicle.
Referring to fig. 7B, in this case, an artifact 705 is added to the ST graph 710 as a constraint. The human obstruction may be, for example, a red light or a pedestrian in the path at a distance of approximately S2 from the cardinal point of reference as perceived by the ADV. Obstacle 705 corresponds to a decision to "stop" the ADV until the man-made obstacle is later cleared (i.e., the traffic light changes from red to green, or the pedestrian is no longer in the path).
FIG. 8 is a flow chart illustrating a method according to one embodiment. Process 800 may be performed by processing logic that may comprise software, hardware, or a combination thereof. For example, the process 800 may be performed by the perception and planning system 110 of an autonomous vehicle. Referring to FIG. 8, at block 801, processing logic calculates a first trajectory based on map and route information. At block 802, processing logic generates a path profile based on the first trajectory, the traffic rules, and obstacle information describing one or more obstacles perceived by the ADV. At block 803, processing logic generates a speed profile based on the path profile, where for each obstacle, the speed profile includes a decision to make or exceed the obstacle. At block 804, processing logic performs quadratic programming optimization on the path profile and the speed profile to identify the best path with the best speed. Quadratic programming optimization may be performed using respective path cost functions, speed cost functions, and/or obstacle cost functions to determine an optimal route having a minimum total cost. At block 805, processing logic generates a second trajectory based on the optimal path and the optimal speed, thereby autonomously controlling the ADV according to the second trajectory.
In one embodiment, the path profile and the velocity profile are iteratively generated using dynamic programming. In one embodiment, for each obstacle decision encountered, the path profile includes decisions to avoid, yield, ignore, or skim from the left or right of the encountered obstacle.
In one embodiment, performing quadratic programming optimization on the path profile and the speed profile comprises: optimizing a first cost function (e.g., one or more path cost functions) using quadratic programming to generate a base-point-side shift map based on a path profile; and optimizing a second cost function (e.g., one or more speed cost functions) using quadratic programming to generate a basepoint-time map based on the speed profile. In another embodiment, the cardinal-lateral shift map is generated by forming one or more obstacles based on one or more obstacle decisions. In another embodiment, the first cost function includes a heading cost, a curvature cost, and/or a distance cost. In another embodiment, the second cost function includes an acceleration cost, a jerk cost, and/or a distance cost. In another embodiment, the processing logic further inserts a plurality of points to the second track that are not present in the first track based on the base-point-side shift map and the base-point-time map.
FIG. 9 is a block diagram illustrating an example of a planning module, according to one embodiment. Planning module 901 is similar to planning module 305 of FIG. 5B. In addition, the planning module 901 includes a gradient descent module 903. The gradient descent module 903 may perform a gradient descent optimization method to optimize the cost function. For example, in one embodiment, the gradient descent module 903 performs a gradient descent optimization on a path cost function similar to the path cost function described above, instead of a quadratic programming optimization performed by the path cost module 529. In another embodiment, the gradient descent module 903 performs a gradient descent optimization on a velocity cost function similar to the velocity cost function described above, instead of a quadratic programming optimization performed by the velocity cost module 535. Gradient descent is a first order iterative optimization algorithm used to find the minimum of a function. To find the local minimum of the function using gradient descent, the algorithm takes a step size proportional to the negative value of the gradient of the function at the current point. The algorithm may compute the derivative (i.e., gradient) of the cost function at the current value and take a step size proportional to the derivative and repeat until the minimum point is reached.
FIG. 10 is a flow chart illustrating a method according to one embodiment. Process 1000 may be performed by processing logic that may comprise software, hardware, or a combination thereof. For example, the process 1000 may be performed by the perception and planning system 110 of an autonomous vehicle. Referring to fig. 10, at block 1001, processing logic calculates a first trajectory based on map and route information. At block 1002, processing logic generates a path profile based on the first trajectory, traffic rules, and obstacle information describing one or more obstacles perceived by the ADV, wherein for each obstacle, the path profile includes a decision to sweep or let go from the left or right side of the obstacle. At block 1003, processing logic generates a speed profile from the traffic rules based on the path profile. At block 1004, processing logic performs a gradient descent optimization based on the path profile and the velocity profile to generate a second trajectory that represents the optimized first trajectory. The gradient descent optimization may be performed using respective path cost functions, speed cost functions, and/or obstacle cost functions to determine an optimal route having a minimum total cost. At block 1005, processing logic controls the ADV according to the second trajectory.
In one embodiment, the path profile and the speed profile are generated iteratively using dynamic programming. In one embodiment, for each obstacle encountered in the obstacle information, the speed profile includes a decision to follow, exceed, yield, stop, or pass the encountered obstacle.
In one embodiment, performing gradient descent optimization based on the path profile and the velocity profile comprises: optimizing the first cost function and the second cost function using gradient descent optimization to generate a base point-side shift map and a base point-time map based on the path profile and the speed profile, respectively; and generating a second trajectory based on the base point-side shift map and the base point-time map, thereby controlling the ADV according to the second trajectory. In another embodiment, the cardinal-lateral shift map is generated by forming one or more obstacles based on one or more obstacle decisions. In another embodiment, the first cost function and the second cost function include a heading cost, a curvature cost, a distance cost, an acceleration cost, and a jerk cost. In another embodiment, the processing logic further inserts a plurality of points in the second track that are not present in the first track based on the base-point-side shift map and the base-point-time map.
In one embodiment, the planning module may determine all feasible and/or possible decisions to move the ADV from the starting location to the ending location based on the obstacles encountered by the ADV. Without initial constraints, the planning module 1100 optimizes each feasible trajectory according to feasible and/or possible obstacle decisions (i.e., avoid, exceed, pass, yield, sweep, stop, ignore) based on the reference lines provided by the routing module 307 and calculates a minimum total cost for each feasible trajectory. For example, for a red traffic light obstacle in a lane, a likely decision would be to stop the ADV, although other obstacle decisions are also possible. The planning module 1100 then selects the trajectory with the smallest total cost to control the ADV. FIG. 11 is a block diagram illustrating an example of a planning module, according to one embodiment. Fig. 12A and 12B are block diagrams illustrating a path cost module and a speed cost module, respectively, according to one embodiment.
Referring to fig. 11, planning module 1100 is similar to planning module 305 of fig. 5B. The planning module 1100 includes an obstacle planning module 1101, a path cost module 1110, and a speed cost module 1120. The obstacle planning module 1101 includes an obstacle cost calculator 1103. The obstacle planning module 1101 plans how to control ADV according to obstacles. The obstacle cost calculator 1103 may calculate an obstacle cost for each obstacle perceived by the ADV. The obstacle cost may represent a cost of avoiding a conflict between the obstacle and the particular trajectory that is calculated. For example, the cost of avoiding a collision between an obstacle and a trajectory may include a cost based on the distance between the obstacle and the closest point of the trajectory ("distance cost"), and a cost of estimating the speed of passage past the obstacle ("cost of speed of passage").
In one embodiment, the distance cost may be ignored when the distance between the trajectory and the obstacle is greater than a threshold (such as 2 meters). In one embodiment, the distance cost is an exponential function. For example, the distance cost may be: w is a1*exp(2-x)-1, wherein w1Is a weight factor and x is the distance between the trajectory and the obstacle. In one embodiment, the cost of passing the speed is a logarithmic function. For example, the cost of passing speed may be: w is a2Log (speed,4), wherein w2Is the weighting factor and speed is the relative speed of the ADV with respect to the passing obstacle. In one embodiment, the total obstacle cost is calculated based on the distance cost and the cost of the passing speed, e.g., the total obstacle cost is the product of the two costs: (w)1*exp(2-x)-1)*(w2*log(speed,4))。
The path cost module 1110 is similar to the path cost module 529 of fig. 5B. Referring to fig. 12A, the path cost module 1110 includes a curvature cost calculator 1201, a curvature delta cost calculator 1203, and a forward length cost calculator 1205. The curvature cost calculator 1201 calculates a curvature cost based on the curvature of each point along the trajectory. In one embodiment, the curvature cost is an exponential function. For example, the curvature cost may be w3*exp(100 x c) -1, wherein w3Is a weighting factor, and c is typically at [0, 0.2 ]]Curvature within a range. The curvature incremental cost calculator 1203 may calculate a curvature incremental cost based on a difference in curvature between two adjacent points of the calculated trajectory. In one embodiment, the curvature increment cost is an exponential function. For example, the incremental cost of curvature may be w4Exp (100 × c') -1, wherein w4Is the weight factor and c' is the change in curvature. The forward length cost calculator 1205 may calculate a forward length cost representing a cost of moving forward toward the reference line of the trajectory. In one embodiment, the forward length cost is a linear function. For example, the forward length cost may be w5(X-X), wherein, w5Is a weight factor, X is the distance to the destination of the path segment, and X is the distance traveled. In one embodiment, the path cost module 1110 may calculate the total path cost based on the curvature cost, the curvature delta cost, and/or the forward length cost for each of all points along the calculated trajectory. For example, the total path cost may be the sum of these three costs for each of all points along the computed trajectory. The path cost may be optimized by quadratic programming and/or gradient descent optimization as described above to obtain a minimum path cost.
In another embodiment, the speed increment cost and the speed increment change cost are constant when the speed increment and the speed increment change are below or above a predetermined threshold, respectively. In one embodiment, the speed cost module 1120 then calculates a total speed cost based on the speed cost, the speed increment cost, and/or the speed increment cost for each of all points along the calculated trajectory. For example, the total velocity cost may be the sum of these three costs for each point along all points of the calculated base point-time curve. The total speed cost may be optimized by quadratic programming and/or gradient descent optimization as described above, resulting in a minimum speed cost. The ADV may then be autonomously controlled by the planning module 1100 by selecting the trajectory with the overall least path cost and velocity cost.
FIG. 13 is a flow chart illustrating a method according to one embodiment. Process 1300 may be performed by processing logic that may comprise software, hardware, or a combination thereof. For example, the process 1300 may be performed by the perception and planning system 110 of an autonomous vehicle. Referring to fig. 13, at block 1301, processing logic generates a number of possible decisions for tracking an ADV from a first location to a second location according to a set of traffic rules based on perception information that perceives a driving environment (including one or more obstacles) around the ADV. At block 1302, processing logic calculates a plurality of trajectories based on a combination of one or more feasible decisions. At block 1303, processing logic calculates a total cost for each trace using a plurality of cost functions. At block 1304, processing logic selects the one trajectory with the smallest total cost as the driving trajectory for autonomously controlling the ADV.
FIG. 14A is a flow diagram illustrating a process of calculating a path cost according to one embodiment. Process 1400 may be performed as part of block 1303 of fig. 13. Process 1400 may be performed by processing logic that may comprise software, hardware, or a combination thereof. For example, the process 1400 may be performed by the route cost for autonomous vehicle module 1110. Referring to fig. 14A, at block 1401, processing logic calculates a curvature cost based on the curvature of each point along the trajectory. At block 1402, processing logic calculates a curvature delta cost based on a difference in curvature between two neighboring points. At block 1403, processing logic calculates a forward length cost representing the cost of moving forward toward the reference line of the trace. At block 1404, processing logic calculates a total path cost based on the curvature cost, the curvature delta cost, and the forward length cost.
FIG. 14B is a flow chart illustrating a process of calculating a speed cost according to one embodiment. Process 1410 may be performed as part of block 1303 of fig. 13. Process 1410 may be performed by processing logic that may include software, hardware, or a combination thereof. For example, process 1410 may be performed by autonomous vehicle speed cost module 1120. Referring to FIG. 14B, at block 1411, processing logic calculates a speed cost based on the speed according to the speed limit at each point. At block 1412, processing logic calculates a speed increment cost representing the cost of changing speed between two adjacent points. At block 1413, processing logic calculates a speed delta change cost representing a cost to change the speed change between two adjacent points. At block 1414, processing logic calculates a total speed cost based on the speed cost, the speed delta cost, and the speed delta change cost.
In one embodiment, for each track, the processing logic further calculates a path cost using a path cost function representing a cost of tracking the ADV from the first location to the second location according to the track. The processing logic also calculates a velocity cost function representing a cost of controlling the ADV at different velocities along the trajectory, wherein the total cost is calculated based on the path cost and the velocity cost.
In one embodiment, calculating the path cost using the path cost function comprises: calculating a curvature cost based on the curvature of each point along the trajectory; and calculating a curvature delta cost based on a difference in curvature between two adjacent points, wherein the path cost is calculated based on the curvature cost and the curvature delta cost for all points along the trajectory. In another embodiment, the processing logic further calculates a forward length cost representing a cost of moving forward towards the reference line of the trajectory, wherein the path cost is further calculated based on the forward length cost for each point.
In one embodiment, calculating the speed cost using the speed cost function includes: calculating an individual speed cost based on the speed from the speed limit at each point, and calculating a speed increment cost representing a cost of changing the speed between two adjacent points, wherein the speed cost is calculated based on the individual speed cost and the speed increment cost for all points along the trajectory. In another embodiment, the processing logic further calculates an acceleration cost based on the acceleration of each point of the trajectory, wherein the velocity cost is further calculated based on the acceleration cost of each point.
In one embodiment, the processing logic is further to calculate a perceived obstacle cost for each obstacle, the obstacle cost representing a cost of avoiding a conflict between the trajectory and the obstacle, wherein the total cost is further calculated based on the obstacle cost. In another embodiment, calculating the obstacle cost includes calculating a minimum distance between the trajectory and the obstacle, and calculating a passing speed of the estimated passing obstacle, wherein the obstacle cost is calculated based on the minimum distance and the passing speed.
FIG. 15 is a block diagram illustrating an example of a data processing system that may be used with one embodiment of the present disclosure. For example, system 1500 may represent any of the data processing systems described above that perform any of the processes or methods described above, such as, for example, any of sensing and planning systems 110 or servers 103-104 of fig. 1. System 1500 may include many different components. These components may be implemented as Integrated Circuits (ICs), portions of integrated circuits, discrete electronic devices or other modules adapted for a circuit board, such as a motherboard or add-in card of a computer system, or as components otherwise incorporated within a chassis of a computer system.
It should also be noted that system 1500 is intended to illustrate a high-level view of many components of a computer system. However, it is to be understood that some embodiments may have additional components and, further, other embodiments may have different arrangements of the components shown. System 1500 may represent a desktop computer, a laptop computer, a tablet computer, a server, a mobile phone, a media player, a Personal Digital Assistant (PDA), a smart watch, a personal communicator, a gaming device, a network router or hub, a wireless Access Point (AP) or repeater, a set-top box, or a combination thereof. Further, while only a single machine or system is illustrated, the term "machine" or "system" shall also be taken to include any collection of machines or systems that individually or jointly execute a set (or multiple sets) of instructions to perform any one or more of the methodologies discussed herein.
In one embodiment, the system 1500 includes a processor 1501, memory 1503, and devices 1505-1508 connected by a bus or interconnect 1510. Processor 1501 may represent a single processor or multiple processors including a single processor core or multiple processor cores. Processor 1501 may represent one or more general-purpose processors, such as a microprocessor, Central Processing Unit (CPU), or the like. More specifically, processor 1501 may be a Complex Instruction Set Computing (CISC) microprocessor, Reduced Instruction Set Computing (RISC) microprocessor, Very Long Instruction Word (VLIW) microprocessor, or a processor implementing other instruction sets, or processors implementing a combination of instruction sets. Processor 1501 may also be one or more special-purpose processors, such as an Application Specific Integrated Circuit (ASIC), a cellular or baseband processor, a Field Programmable Gate Array (FPGA), a Digital Signal Processor (DSP), a network processor, a graphics processor, a communications processor, a cryptographic processor, a coprocessor, an embedded processor, or any other type of logic capable of processing instructions.
Processor 1501 (which may be a low-power multi-core processor socket such as an ultra-low voltage processor) may serve as a main processing unit and central hub for communicating with the various components of the system. Such a processor may be implemented as a system on a chip (SoC). Processor 1501 is configured to execute instructions for performing the operations and steps discussed herein. The system 1500 may also include a graphics interface to communicate with an optional graphics subsystem 1504, which may include a display controller, a graphics processor, and/or a display device.
The input device 1506 may include a mouse, a touch pad, a touch-sensitive screen (which may be integrated with the display device 1504), a pointing device (such as a stylus) and/or a keyboard (e.g., a physical keyboard or a virtual keyboard displayed as part of the touch-sensitive screen). For example, the input device 1506 may include a touch screen controller coupled to a touch screen. Touch screens and touch screen controllers, for example, may detect contact and movement or discontinuities thereof using any of a variety of touch sensitive technologies, including but not limited to capacitive, resistive, infrared, and surface acoustic wave technologies, as well as other proximity sensor arrays or other elements for determining one or more points of contact with the touch screen.
To provide persistent storage for information such as data, applications, one or more operating systems, etc., a mass storage device (not shown) may also be coupled to processor 1501. In various embodiments, such mass storage devices may be implemented via Solid State Devices (SSDs) in order to achieve thinner and lighter system designs and improve system responsiveness. However, in other embodiments, the mass storage device may be implemented primarily using a Hard Disk Drive (HDD), with a smaller amount of the SSD storage device acting as an SSD cache to enable non-volatile storage of context state and other such information during a power down event, enabling fast power up upon a system activity restart. Additionally, a flash device may be coupled to processor 1501, for example, via a Serial Peripheral Interface (SPI). Such flash memory devices may provide non-volatile storage of system software, including the BIOS and other firmware of the system.
The computer-readable storage medium 1509 may also be used to permanently store some of the software functions described above. While the computer-readable storage medium 1509 is shown in an exemplary embodiment to be a single medium, the term "computer-readable storage medium" should be taken to include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions. The term "computer-readable storage medium" shall also be taken to include any medium that is capable of storing or encoding a set of instructions for execution by the machine and that cause the machine to perform any one or more of the methodologies of the present disclosure. The term "computer-readable storage medium" shall accordingly be taken to include, but not be limited to, solid-state memories, and optical and magnetic media, or any other non-transitory machine-readable medium.
The processing module/unit/logic 1528, components, and other features described herein may be implemented as discrete hardware components or integrated within the functionality of hardware components, such as ASICS, FPGAs, DSPs, or similar devices. Further, the processing module/unit/logic 1528 may be implemented as firmware or functional circuitry within a hardware device. Further, the processing module/unit/logic 1528 may be implemented in any combination of hardware devices and software components.
It should be noted that while system 1500 is illustrated with various components of a data processing system, it is not intended to represent any particular architecture or manner of interconnecting the components; as such details are not germane to embodiments of the present disclosure. It will also be appreciated that network computers, hand-held computers, mobile telephones, servers, and/or other data processing systems which have fewer components or perhaps more components may also be used with embodiments of the present disclosure.
Some portions of the foregoing detailed description have been presented in terms of algorithms and symbolic representations of operations on data bits within a computer memory. These algorithmic descriptions and representations are the means used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. An algorithm is here, and generally, considered to be a self-consistent sequence of operations leading to a desired result. The operations are those requiring physical manipulations of physical quantities.
It should be borne in mind, however, that all of these and similar terms are to be associated with the appropriate physical quantities and are merely convenient labels applied to these quantities. Unless specifically stated otherwise as apparent from the above discussion, it is appreciated that throughout the description, discussions utilizing terms such as those set forth in the appended claims, refer to the action and processes of a computer system, or similar electronic computing device, that manipulates and transforms data represented as physical (electronic) quantities within the computer system's registers and memories into other data similarly represented as physical quantities within the computer system memories or registers or other such information storage, transmission or display devices.
Embodiments of the present disclosure also relate to apparatuses for performing the operations herein. Such a computer program is stored in a non-transitory computer readable medium. A machine-readable medium includes any mechanism for storing information in a form readable by a machine (e.g., a computer). For example, a machine-readable (e.g., computer-readable) medium includes a machine (e.g., computer) readable storage medium (e.g., read only memory ("ROM"), random access memory ("RAM"), magnetic disk storage media, optical storage media, flash memory devices).
The processes or methods depicted in the foregoing figures may be performed by processing logic that comprises hardware (e.g., circuitry, dedicated logic, etc.), software (e.g., embodied on a non-transitory computer readable medium), or a combination of both. Although the processes or methods are described above in terms of some sequential operations, it should be appreciated that some of the operations may be performed in a different order. Further, some operations may be performed in parallel rather than sequentially.
Embodiments of the present disclosure are not described with reference to any particular programming language. It will be appreciated that a variety of programming languages may be used to implement the teachings of embodiments of the disclosure as described herein.
In the foregoing specification, embodiments of the invention have been described with reference to specific exemplary embodiments thereof. It will be evident that various modifications may be made thereto without departing from the broader spirit and scope of the invention as set forth in the following claims. The specification and drawings are, accordingly, to be regarded in an illustrative sense rather than a restrictive sense.
Claims (18)
1. A computer-implemented method of generating a driving trajectory for an autonomous vehicle, the method comprising:
calculating a first trajectory based on the map and the route information;
generating a path profile based on the first trajectory, traffic rules, and obstacle information describing one or more obstacles perceived by the autonomous vehicle;
generating a speed profile based on the path profile, wherein the speed profile includes, for each of the plurality of obstacles, a decision to make or exceed the obstacle;
performing a quadratic programming optimization on the path profile and the speed profile to identify an optimal path having an optimal speed, wherein the performing quadratic programming optimization comprises:
optimizing a path cost function using quadratic programming to recalculate the path profile, the recalculated path profile including a base-side shift map, an
Re-computing the speed profile using quadratic programming to optimize a speed cost function, the re-computed speed profile comprising a base point-time map; and
generating a second trajectory based on the optimal path and the optimal speed to autonomously control the autonomous vehicle in accordance with the second trajectory.
2. The method of claim 1, wherein the path profile and the speed profile are iteratively generated using dynamic programming.
3. The method of claim 1, wherein the path profile includes, for each obstacle decision encountered, a decision to let go, ignore, or skim from the left or right of the encountered obstacle.
4. The method of claim 1, wherein the cardinal-lateral shift map is generated by forming one or more obstacles based on one or more obstacle decisions.
5. The method of claim 1, wherein the path cost function includes a heading cost, a curvature cost, and a distance cost.
6. The method of claim 1, wherein the velocity cost function includes an acceleration cost, a jerk cost, and a distance cost.
7. The method of claim 1, further comprising: inserting a plurality of points, which are not present in the first track, into the second track based on the base point-side shift map and the base point-time map.
8. A non-transitory machine-readable medium having instructions stored thereon, which when executed by a processor, cause the processor to perform operations comprising:
calculating a first trajectory based on the map and the route information;
generating a path profile based on the first trajectory, traffic rules, and obstacle information describing one or more obstacles perceived by an autonomous vehicle;
generating a speed profile based on the path profile, wherein the speed profile includes, for each of the plurality of obstacles, a decision to make or exceed the obstacle;
performing a quadratic programming optimization on the path profile and the speed profile to identify an optimal path having an optimal speed, wherein the performing quadratic programming optimization comprises:
optimizing a path cost function using quadratic programming to recalculate the path profile, the recalculated path profile including a base-side shift map, an
Re-computing the speed profile using quadratic programming to optimize a speed cost function, the re-computed speed profile comprising a base point-time map; and
generating a second trajectory based on the optimal path and the optimal speed to autonomously control the autonomous vehicle in accordance with the second trajectory.
9. The non-transitory machine-readable medium of claim 8, wherein the path profile and the speed profile are iteratively generated using dynamic programming.
10. The non-transitory machine-readable medium of claim 8, wherein the path profile includes, for each obstacle decision encountered, a decision to let go, ignore, or skim left or right of the encountered obstacle.
11. The non-transitory machine-readable medium of claim 8, wherein the base-point-side shift map is generated by forming one or more obstacles based on one or more obstacle decisions.
12. The non-transitory machine-readable medium of claim 8, wherein the path cost function includes a heading cost, a curvature cost, and a distance cost.
13. The non-transitory machine-readable medium of claim 8, wherein the velocity cost function includes an acceleration cost, a jerk cost, and a distance cost.
14. The non-transitory machine-readable medium of claim 8, further comprising: inserting a plurality of points, which are not present in the first track, into the second track based on the base point-side shift map and the base point-time map.
15. A data processing system comprising:
a processor; and
a memory coupled to the processor to store instructions that, when executed by the processor, cause the processor to perform operations comprising:
calculating a first trajectory based on the map and the route information;
generating a path profile based on the first trajectory, traffic rules, and obstacle information describing one or more obstacles perceived by an autonomous vehicle;
generating a speed profile based on the path profile, wherein the speed profile includes, for each of the plurality of obstacles, a decision to make or exceed the obstacle;
performing a quadratic programming optimization on the path profile and the speed profile to identify an optimal path having an optimal speed, wherein the performing quadratic programming optimization comprises:
optimizing a path cost function using quadratic programming to recalculate the path profile, the recalculated path profile including a base-side shift map, an
Re-computing the speed profile using quadratic programming to optimize a speed cost function, the re-computed speed profile comprising a base point-time map; and
generating a second trajectory based on the optimal path and the optimal speed to autonomously control the autonomous vehicle in accordance with the second trajectory.
16. The system of claim 15, wherein the path profile and the speed profile are iteratively generated using dynamic programming.
17. The system of claim 15, wherein the path profile includes, for each obstacle decision encountered, a decision to let go, ignore, or skim left or right of the encountered obstacle.
18. The system of claim 15, wherein the cardinal-lateral shift map is generated by forming one or more obstacles based on one or more obstacle decisions.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/701,411 | 2017-09-11 | ||
US15/701,411 US10754339B2 (en) | 2017-09-11 | 2017-09-11 | Dynamic programming and quadratic programming based decision and planning for autonomous driving vehicles |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109491377A CN109491377A (en) | 2019-03-19 |
CN109491377B true CN109491377B (en) | 2022-07-08 |
Family
ID=65631051
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810745836.2A Active CN109491377B (en) | 2017-09-11 | 2018-07-09 | DP and QP based decision and planning for autonomous vehicles |
Country Status (2)
Country | Link |
---|---|
US (1) | US10754339B2 (en) |
CN (1) | CN109491377B (en) |
Families Citing this family (50)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10606277B2 (en) * | 2017-09-18 | 2020-03-31 | Baidu Usa Llc | Speed optimization based on constrained smoothing spline for autonomous driving vehicles |
US11262756B2 (en) * | 2018-01-15 | 2022-03-01 | Uatc, Llc | Discrete decision architecture for motion planning system of an autonomous vehicle |
US10732632B2 (en) * | 2018-01-31 | 2020-08-04 | Baidu Usa Llc | Method for generating a reference line by stitching multiple reference lines together using multiple threads |
US10782699B2 (en) * | 2018-03-10 | 2020-09-22 | Baidu Usa Llc | Real-time perception adjustment and driving adaption based on surrounding vehicles' behavior for autonomous driving vehicles |
US10816985B2 (en) * | 2018-04-17 | 2020-10-27 | Baidu Usa Llc | Method on moving obstacle representation for trajectory planning |
US10823575B2 (en) * | 2018-06-27 | 2020-11-03 | Baidu Usa Llc | Reference line smoothing method using piecewise spiral curves with weighted geometry costs |
CN109345589A (en) * | 2018-09-11 | 2019-02-15 | 百度在线网络技术(北京)有限公司 | Method for detecting position, device, equipment and medium based on automatic driving vehicle |
US11747825B2 (en) | 2018-10-12 | 2023-09-05 | Boston Dynamics, Inc. | Autonomous map traversal with waypoint matching |
US10990099B2 (en) * | 2018-11-16 | 2021-04-27 | Great Wall Motor Company Limited | Motion planning methods and systems for autonomous vehicle |
US11181919B2 (en) * | 2018-11-27 | 2021-11-23 | Wipro Limited | Method and system for determining an optimal trajectory for navigation of an autonomous vehicle |
CN111684379B (en) * | 2018-12-26 | 2023-12-22 | 百度时代网络技术(北京)有限公司 | Optimal planner switching method for three-point turning of automatic driving vehicle |
US10809736B2 (en) * | 2018-12-27 | 2020-10-20 | Baidu Usa Llc | ST-graph learning based decision for autonomous driving vehicle |
US11137766B2 (en) | 2019-03-07 | 2021-10-05 | Zoox, Inc. | State machine for traversing junctions |
CN109947101B (en) * | 2019-03-18 | 2022-11-29 | 北京智行者科技股份有限公司 | Path smoothing processing method and device |
CN110576950B (en) * | 2019-03-25 | 2021-07-09 | 宁波靖海智航科技有限公司 | Overwater equipment with foreign matter collision detection device |
US11106212B2 (en) * | 2019-03-26 | 2021-08-31 | Baidu Usa Llc | Path planning for complex scenes with self-adjusting path length for autonomous driving vehicles |
CN110057362B (en) * | 2019-04-26 | 2022-09-16 | 安徽理工大学 | Mobile robot path planning method of limited unit map |
US11409284B2 (en) * | 2019-05-15 | 2022-08-09 | Baidu Usa Llc | Relaxation optimization model to plan an open space trajectory for autonomous vehicles |
JP7194640B2 (en) * | 2019-05-16 | 2022-12-22 | 本田技研工業株式会社 | VEHICLE CONTROL DEVICE, VEHICLE CONTROL METHOD, AND PROGRAM |
WO2020249995A1 (en) * | 2019-06-14 | 2020-12-17 | 日産自動車株式会社 | Travel assistance method and travel assistance device |
US11480962B1 (en) | 2019-06-28 | 2022-10-25 | Zoox, Inc. | Dynamic lane expansion |
CN110364009A (en) * | 2019-07-16 | 2019-10-22 | 华人运通(上海)自动驾驶科技有限公司 | Traveling planing method, device, roadside device and storage medium based on roadside device |
US20210027629A1 (en) * | 2019-07-25 | 2021-01-28 | Baidu Usa Llc | Blind area processing for autonomous driving vehicles |
EP4010659A1 (en) | 2019-08-06 | 2022-06-15 | Boston Dynamics, Inc. | Intermediate waypoint generator |
US20210048825A1 (en) * | 2019-08-13 | 2021-02-18 | GM Global Technology Operations LLC | Predictive and reactive field-of-view-based planning for autonomous driving |
US11958183B2 (en) | 2019-09-19 | 2024-04-16 | The Research Foundation For The State University Of New York | Negotiation-based human-robot collaboration via augmented reality |
US11378967B2 (en) * | 2019-09-25 | 2022-07-05 | Baidu Usa Llc | Enumeration based failure prevention QP smoother for autonomous vehicles |
US11815891B2 (en) * | 2019-10-22 | 2023-11-14 | Baidu Usa Llc | End dynamics and constraints relaxation algorithm on optimizing an open space trajectory |
US11427191B2 (en) | 2019-10-31 | 2022-08-30 | Zoox, Inc. | Obstacle avoidance action |
JP2023500825A (en) * | 2019-10-31 | 2023-01-11 | ズークス インコーポレイテッド | Obstacle avoidance action |
US11532167B2 (en) | 2019-10-31 | 2022-12-20 | Zoox, Inc. | State machine for obstacle avoidance |
CN112818727A (en) * | 2019-11-18 | 2021-05-18 | 华为技术有限公司 | Road constraint determination method and device |
US11567506B2 (en) * | 2019-12-30 | 2023-01-31 | Baidu Usa Llc | Speed planning guidance line for mild slow down |
EP4081876B1 (en) * | 2020-01-28 | 2024-04-10 | Five AI Limited | Planning in mobile robots |
EP3885226B1 (en) | 2020-03-25 | 2024-08-14 | Aptiv Technologies AG | Method and system for planning the motion of a vehicle |
US11586209B2 (en) * | 2020-04-08 | 2023-02-21 | Baidu Usa Llc | Differential dynamic programming (DDP) based planning architecture for autonomous driving vehicles |
US11745760B2 (en) * | 2020-05-19 | 2023-09-05 | Baidu Usa Llc | Parallel obstacle buffer system for path planning |
US11628858B2 (en) * | 2020-09-15 | 2023-04-18 | Baidu Usa Llc | Hybrid planning system for autonomous vehicles |
CN112498366B (en) * | 2020-11-20 | 2022-04-05 | 苏州智加科技有限公司 | Autonomous vehicle, control method, device, equipment and storage medium |
CN112572443B (en) * | 2020-12-22 | 2021-12-07 | 北京理工大学 | Real-time collision-avoidance trajectory planning method and system for lane changing of vehicles on highway |
CN112711255B (en) * | 2020-12-24 | 2024-01-19 | 南方科技大学 | Mobile robot obstacle avoidance method, control equipment and storage medium |
US20240300541A1 (en) * | 2021-03-26 | 2024-09-12 | Embotech Ag | Method and system for controlling autonomous or semi-autonomous vehicle |
DE102021113170A1 (en) * | 2021-05-20 | 2022-11-24 | Bayerische Motoren Werke Aktiengesellschaft | Method and device for determining an acceleration specification for a vehicle |
RU2767826C1 (en) | 2021-05-24 | 2022-03-22 | Общество с ограниченной ответственностью «Яндекс Беспилотные Технологии» | Method and device for vehicle control |
KR102317430B1 (en) * | 2021-05-27 | 2021-10-26 | 주식회사 라이드플럭스 | Method, server and computer program for creating road network map to design a driving plan for automatic driving vehicle |
CN113085850B (en) * | 2021-06-08 | 2021-09-21 | 新石器慧通(北京)科技有限公司 | Vehicle obstacle avoidance method and device, electronic equipment and storage medium |
CN113815647B (en) * | 2021-10-28 | 2023-03-14 | 广州文远知行科技有限公司 | Vehicle speed planning method, device, equipment and medium |
CN114019971B (en) * | 2021-11-04 | 2024-03-26 | 北京三快在线科技有限公司 | Unmanned equipment control method and device, storage medium and electronic equipment |
DE102022102118A1 (en) | 2022-01-31 | 2023-08-03 | Dr. Ing. H.C. F. Porsche Aktiengesellschaft | Method and apparatus for generating trajectories on a given route for multiple vehicles with given performance specifications |
CN115042783A (en) * | 2022-06-28 | 2022-09-13 | 重庆长安汽车股份有限公司 | Vehicle speed planning method and device, electronic equipment and medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2012088635A1 (en) * | 2010-12-30 | 2012-07-05 | 中国科学院自动化研究所 | Adaptive cruise control system and method for vehicle |
CN104133473A (en) * | 2008-10-24 | 2014-11-05 | 格瑞股份公司 | Control method of autonomously driven vehicle |
CN105599764A (en) * | 2014-11-12 | 2016-05-25 | 现代自动车株式会社 | Driving path planning apparatus and method for autonomous vehicle |
CN106094812A (en) * | 2015-04-27 | 2016-11-09 | 通用汽车环球科技运作有限责任公司 | Reaction equation path planning for autonomous driving |
CN106926844A (en) * | 2017-03-27 | 2017-07-07 | 西南交通大学 | A kind of dynamic auto driving lane-change method for planning track based on real time environment information |
Family Cites Families (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101054479B1 (en) * | 2009-03-27 | 2011-08-05 | 국방과학연구소 | Regional Route Planning Apparatus and Method for Unmanned Vehicles Using Directional Speed Maps for Each Direction |
US8554343B2 (en) * | 2010-12-08 | 2013-10-08 | Mitsubishi Electric Research Laboratories, Inc. | Method for solving control problems |
US10270642B2 (en) * | 2012-12-05 | 2019-04-23 | Origin Wireless, Inc. | Method, apparatus, and system for object tracking and navigation |
EP2918467A4 (en) * | 2012-11-08 | 2016-06-08 | Toyota Motor Co Ltd | Drive assist device and method, collision prediction device and method, and alerting device and method |
CN103035121A (en) * | 2012-12-06 | 2013-04-10 | 南京航空航天大学 | Planning method of intelligent vehicle autonomous running dynamic trajectory and system of the same |
JP6120166B2 (en) * | 2013-06-04 | 2017-04-26 | パナソニックIpマネジメント株式会社 | Vehicle stop determination result providing method, vehicle stop determination device, and vehicle stop determination system |
US9174672B2 (en) * | 2013-10-28 | 2015-11-03 | GM Global Technology Operations LLC | Path planning for evasive steering maneuver in presence of target vehicle and surrounding objects |
KR101581286B1 (en) * | 2014-01-17 | 2015-12-31 | 전남대학교산학협력단 | System and method for path planning for autonomous navigation of driverless ground vehicle |
CN104596533B (en) * | 2015-01-07 | 2017-08-01 | 上海交通大学 | Automatic guided vehicle and its guidance method based on map match |
JP6180458B2 (en) * | 2015-04-17 | 2017-08-16 | 三菱電機株式会社 | Vehicle energy management system |
CN104933228B (en) * | 2015-05-27 | 2018-03-02 | 西安交通大学 | Unmanned vehicle real-time track planing method based on Speed Obstacles |
WO2017002590A1 (en) * | 2015-06-29 | 2017-01-05 | 株式会社日立製作所 | Travel command generation device |
US9630619B1 (en) * | 2015-11-04 | 2017-04-25 | Zoox, Inc. | Robotic vehicle active safety systems and methods |
US10210765B2 (en) * | 2016-01-11 | 2019-02-19 | The Boeing Company | Inference and interpolation of continuous 4D trajectories |
US10800416B2 (en) * | 2016-09-21 | 2020-10-13 | Apple Inc. | Vehicle control system |
US10146224B2 (en) * | 2016-11-09 | 2018-12-04 | GM Global Technology Operations LLC | Processor-implemented systems and methods for automated driving |
US10296012B2 (en) * | 2016-12-21 | 2019-05-21 | X Development Llc | Pre-computation of kinematically feasible roadmaps |
US10209715B2 (en) * | 2017-01-19 | 2019-02-19 | Robert Bosch Gmbh | System and method of using crowd-sourced driving path data in an autonomous or semi-autonomous driving system |
US10353390B2 (en) * | 2017-03-01 | 2019-07-16 | Zoox, Inc. | Trajectory generation and execution architecture |
US10365653B2 (en) * | 2017-06-12 | 2019-07-30 | GM Global Technology Operations LLC | Personalized autonomous vehicle ride characteristics |
US10725470B2 (en) * | 2017-06-13 | 2020-07-28 | GM Global Technology Operations LLC | Autonomous vehicle driving systems and methods for critical conditions |
-
2017
- 2017-09-11 US US15/701,411 patent/US10754339B2/en active Active
-
2018
- 2018-07-09 CN CN201810745836.2A patent/CN109491377B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104133473A (en) * | 2008-10-24 | 2014-11-05 | 格瑞股份公司 | Control method of autonomously driven vehicle |
WO2012088635A1 (en) * | 2010-12-30 | 2012-07-05 | 中国科学院自动化研究所 | Adaptive cruise control system and method for vehicle |
CN105599764A (en) * | 2014-11-12 | 2016-05-25 | 现代自动车株式会社 | Driving path planning apparatus and method for autonomous vehicle |
CN106094812A (en) * | 2015-04-27 | 2016-11-09 | 通用汽车环球科技运作有限责任公司 | Reaction equation path planning for autonomous driving |
CN106926844A (en) * | 2017-03-27 | 2017-07-07 | 西南交通大学 | A kind of dynamic auto driving lane-change method for planning track based on real time environment information |
Also Published As
Publication number | Publication date |
---|---|
CN109491377A (en) | 2019-03-19 |
US20190079523A1 (en) | 2019-03-14 |
US10754339B2 (en) | 2020-08-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109491377B (en) | DP and QP based decision and planning for autonomous vehicles | |
CN109491376B (en) | Dynamic programming and gradient descent based decision and programming for autonomous vehicles | |
CN110119140B (en) | System and method for accelerating curvilinear projection | |
CN109947090B (en) | Non-blocking boundary for autonomous vehicle planning | |
CN109521761B (en) | Constrained smooth spline-based speed optimization method, medium, and system for autonomous vehicles | |
CN110083149B (en) | Path and speed optimized feedback mechanism for autonomous vehicles | |
CN109521762B (en) | Computer-implemented method, medium, and system for generating optimal reference line for autonomous vehicle | |
CN109521763B (en) | Constrained smooth spline-based path optimization for autonomous vehicles | |
US10515321B2 (en) | Cost based path planning for autonomous driving vehicles | |
CN108981730B (en) | Method and system for generating a reference path for operating an autonomous vehicle | |
CN109955853B (en) | Method, system and storage medium for operating an autonomous vehicle | |
US10775801B2 (en) | Determining speeds along a path for autonomous driving vehicles | |
US10816985B2 (en) | Method on moving obstacle representation for trajectory planning | |
CN110096053B (en) | Driving trajectory generation method, system and machine-readable medium for autonomous vehicle | |
CN110728014A (en) | Reference line smoothing method using segmented spiral curves with weighted geometric cost | |
US20190315357A1 (en) | Novel method for speed adjustment of autonomous driving vehicles prior to lane change | |
CN112009487B (en) | Determining speed of an autonomous vehicle | |
CN111830958B (en) | Method, system and machine readable medium for operating an autonomous vehicle | |
CN111830959A (en) | Method, system, and machine-readable medium for operating an autonomous vehicle | |
CN112020686B (en) | QP spline path and spiral path based reference line smoothing method for autopilot | |
CN111615476A (en) | Spiral curve based vertical parking planning system for autonomous vehicles | |
CN111649751B (en) | Super-free stitching method for reference line smoothing |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |