You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
-[Paper Reading in 2019](https://towardsdatascience.com/the-200-deep-learning-papers-i-read-in-2019-7fb7034f05f7?source=friends_link&sk=7628c5be39f876b2c05e43c13d0b48a3)
36
36
37
-
## 2024-06 (7)
37
+
## 2024-06 (8)
38
38
-[LINGO-1: Exploring Natural Language for Autonomous Driving](https://wayve.ai/thinking/lingo-natural-language-autonomous-driving/)[[Notes](paper_notes/lingo1.md)][Wayve, open-loop world model]
39
39
-[LINGO-2: Driving with Natural Language](https://wayve.ai/thinking/lingo-2-driving-with-language/)[[Notes](paper_notes/lingo2.md)][Wayve, closed-loop world model]
40
40
-[OpenVLA: An Open-Source Vision-Language-Action Model](https://arxiv.org/abs/2406.09246)[open source RT-2]
@@ -49,6 +49,7 @@ I regularly update [my blog in Toward Data Science](https://medium.com/@patrickl
49
49
-[trajdata: A Unified Interface to Multiple Human Trajectory Datasets](https://arxiv.org/abs/2307.13924) <kbd>NeurIPS 2023</kbd> [Marco Pavone, Nvidia]
50
50
-[Optimal Vehicle Trajectory Planning for Static Obstacle Avoidance using Nonlinear Optimization](https://arxiv.org/abs/2307.09466)[Xpeng]
51
51
-[Jointly Learnable Behavior and Trajectory Planning for Self-Driving Vehicles](https://arxiv.org/abs/1910.04586)[[Notes](paper_notes/joint_learned_bptp.md)] <kbd>IROS 2019 Oral</kbd> [Uber ATG, behavioral planning, motion planning]
52
+
-[HiVT: Hierarchical Vector Transformer for Multi-Agent Motion Prediction]
52
53
-[Enhancing End-to-End Autonomous Driving with Latent World Model](https://arxiv.org/abs/2406.08481)
53
54
-[OccNeRF: Advancing 3D Occupancy Prediction in LiDAR-Free Environments](https://arxiv.org/abs/2312.09243)[Jiwen Lu]
54
55
-[RenderOcc: Vision-Centric 3D Occupancy Prediction with 2D Rendering Supervision](https://arxiv.org/abs/2309.09502) <kbd>ICRA 2024</kbd>
@@ -246,6 +247,8 @@ I regularly update [my blog in Toward Data Science](https://medium.com/@patrickl
246
247
-[BEVDepth: Acquisition of Reliable Depth for Multi-view 3D Object Detection](https://arxiv.org/abs/2206.10092)[[Notes](paper_notes/bevdepth.md)][BEVNet, NuScenes SOTA, Megvii]
Copy file name to clipboardExpand all lines: learning_pnc/pnc_notes.md
+7-5Lines changed: 7 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -2,7 +2,7 @@
2
2
3
3
## Introduction
4
4
- The notes were taken for the [Prediction, Decision and Planning for Autonomous driving](https://www.shenlanxueyuan.com/course/671) from Shenlan Xueyuan mooc course.
5
-
- The lecturer is [Wenchao Ding](website: https://wenchaoding.github.io/personal/index.html), former engineer at Huawei and not AP at Fudan University.
5
+
- The lecturer is [Wenchao Ding](website: https://wenchaoding.github.io/personal/index.html), former engineer at Huawei and now AP at Fudan University.
6
6
7
7
# Model-based Prediction
8
8
## Overview
@@ -463,11 +463,13 @@
463
463
464
464
- Continuous state with belief MDP
465
465
- Put complexity into state transition, and solve with ML.
466
-
- Normal solution: MPC, with limited lookup ahead (forward simulation).
467
-
- MCTS
466
+
- Normal solution: MPC-like , with limited lookup ahead (forward simulation).
467
+
- MCTS with forward simulation
468
+
- The complexity lies in multi-agent interaction roll out, and branching out
469
+
- MPC: receding horizon planning
468
470
- Plan in MDP
469
-
- Assuming the most likely belief is the real state. In the ULT case, assuming the most likely behavior of the other car to be reality, and act accordingly.
470
-
-Cannot actively collect information. This is actually the charm of POMDP’s intelligence. POMDP will lead to some action that actively collects information.
471
+
-Approximate POMDP as MDP. Assuming the most likely belief argmax(b) is the real state. In the ULT case, assuming the most likely behavior of the other car to be reality, and act accordingly.
472
+
-MDP cannot actively collect information. This is actually the charm of POMDP’s intelligence. POMDP will lead to some action that actively collects information.
Copy file name to clipboardExpand all lines: paper_notes/eudm.md
+4-1Lines changed: 4 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -11,6 +11,9 @@ In order to make POMDP more tractable it is essential to incorporate domain know
11
11
12
12
In EUDM, ego behavior is allowed to change, allowing more flexible decision making than MPDM. This allows EUDM can make a lane-change decision even before passing the blocking vehicle (accelerate, then lane change).
EUDM does guided branching in both action (of ego) and intention (of others).
15
18
16
19
EUDM couples prediction and planning module.
@@ -20,7 +23,7 @@ It is further improved by [MARC](marc.md) where it considers risk-aware continge
20
23
#### Key ideas
21
24
- DCP-Tree (domain specific closed-loop policy tree), ego-centric
22
25
- Guided branching in action space
23
-
- Each trace only contains ONE change of action (more flexible than MPDM but still manageable).
26
+
- Each trace only contains ONE change of action (more flexible than MPDM but still manageable). This is a tree with pruning mechanism built-in. [MCDM](mcdm.md) essentially has a much more aggressive pruning as only one type of action is allowed (KKK, RRR, LLL, etc)
24
27
- Each semantic action is 2s, 4 levels deep, so planning horizon of 8s.
25
28
- CFB (conditional focused branching), for other agents
Copy file name to clipboardExpand all lines: paper_notes/marc.md
+19-10Lines changed: 19 additions & 10 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -7,18 +7,25 @@ tl;dr: Generating safe and non-conservative behaviors in dense dynamic environme
7
7
#### Overall impression
8
8
This is a continuation of work in [MPDM](mpdm.md) and [EUDM](eudm.md). It introduces dynamic branching based on scene-level divergence, and risk-aware contingency planning based on user-defined risk tolerance.
9
9
10
-
POMDP provides a theoretically sounds framework to handle dynamic interaction, but it suffers from curse of dimensionality and making it infeasible to solve in realtime.
10
+
POMDP provides a theoretically sounds framework to handle dynamic interaction, but it suffers from curse of dimensionality and making it infeasible to solve in realtime.
11
11
12
-
*[MPDM](mpdm.md) prunes belief trees heavily and decomposes POMDP into a limited number of closed-loop policy evaluations. MPDM has only one ego policy over planning horizon (8s). Mainly BP.
13
-
* EUDM improves by having multiple (2) policy in planning horizon, and performs DCP-Tree and CFB (conditoned focused branching) to use domain specific knowledge to guide branching in both action and intention space. Mainly BP.
14
-
* MARC performs risk-aware contigency planning based on multiple scenarios. And it combines BP and MP.
15
-
* All previous MPDM-like methods consider the optimal policy and single trajectory generation over all scenarios, resulting in lack of gurantee of policy consistency and loss of multimodality info.
12
+
[MPDM](mpdm.md) and [EUDM](eudm.md) are mainly BP models, but [MARC](marc.md) combines BP and MP.
13
+
14
+
belief trees heavily and decomposes POMDP into a limited number of closed-loop policy evaluations.
15
+
16
+
For the policy tree (or policy-conditioned scenario tree) building, we can see how the tree got built with more and more careful pruning process with improvements from different works.
17
+
18
+
*[MPDM](mpdm.md) is the pioneering work prunes belief trees heavily and decomposes POMDP into a limited number of closed-loop policy evaluations. MPDM has only one ego policy over planning horizon (8s).
19
+
*[MPDM](mpdm.md) iterates over all ego policies, and uses the most likely one policy given road structure and pose of vehicle.
20
+
*[MPDM2](mpdm2.md) iterates over all ego policies, and iterate over (a set of) possible policies of other agents predicted by a motion prediction model.
21
+
*[EUDM](eudm.md) itrates all ego policies, and then iterate over all possible policies of other agents to identify **critical scenarios** (CFB, conditioned filtered branching). Guide branching in both action and intention space. [EPSILON](epsilon.md) used the same method.
22
+
*[MARC](marc.md) iterates all ego policies, iterates over a set of predicted policies of other agents, identifies **key agents** (and ignores other agents even in critical scenarios).
23
+
24
+
25
+
All previous MPDM-like methods consider the optimal policy and single trajectory generation over all scenarios, resulting in lack of gurantee of policy consistency and loss of multimodality info.
16
26
17
27
#### Key ideas
18
-
- Planning is hard from uncertainty and interaction (inherently multimodal intentions).
19
-
- For interactive decision making, MDP or POMDP are mathematically rigorous formulations for decision processes in stochastic environments.
20
-
- For static (non-interactive) decision making, the normal trioka of planninig (sampling, searching, optimization) would suffice.
21
-
-*Contigency planning* generates deterministic behavior for mulutiple future scenarios. In other words, it plans a short-term trajectory that ensures safety for all potential scenarios.
28
+
-*Contigency planning* generates deterministic behavior for mulutiple future scenarios. In other words, it plans a short-term trajectory that ensures safety for all potential scenarios. --> This is very similar to the idea of *backup plan* in [EPSILON](epsilon.md).
22
29
- Scenario tree construction
23
30
- generating policy-conditioned critical scenario sets via closed-loop forward simulation (similar to CFB in EUDM?).
24
31
- building scenario tree with scene-level divergence assessment. Determine the latest timestamp at which the scenario diverge. Delaying branching time as much as possble.
@@ -35,7 +42,9 @@ POMDP provides a theoretically sounds framework to handle dynamic interaction, b
35
42
- with better effiency (avg speed) and riding comfort (max decel/acc).
36
43
37
44
#### Technical details
38
-
- Summary of technical details, such as important training details, or bugs of previous benchmarks.
45
+
- Planning is hard from uncertainty and interaction (inherently multimodal intentions).
46
+
- For interactive decision making, MDP or POMDP are mathematically rigorous formulations for decision processes in stochastic environments.
47
+
- For static (non-interactive) decision making, the normal trioka of planninig (sampling, searching, optimization) would suffice.
39
48
40
49
#### Notes
41
50
- Questions and notes on how to improve/revise the current work
Copy file name to clipboardExpand all lines: paper_notes/mpdm.md
+1-2Lines changed: 1 addition & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -33,5 +33,4 @@ Despite simple design, MPDM is a pioneering work in decision making, and improve
33
33
- Summary of technical details, such as important training details, or bugs of previous benchmarks.
34
34
35
35
#### Notes
36
-
- Questions and notes on how to improve/revise the current work
37
-
36
+
- The white paper from [May Mobility](https://maymobility.com/resources/autonomy-at-scale-white-paper/) explains the idea with more plain language and examples.
Copy file name to clipboardExpand all lines: paper_notes/mpdm2.md
+2-11Lines changed: 2 additions & 11 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -5,18 +5,9 @@ _June 2024_
5
5
tl;dr: Improvement of MPDM in predicting the intention of other vehicles.
6
6
7
7
#### Overall impression
8
-
The majority is the same as the previous work [MPDM](mpdm.md).
9
-
10
-
For the policy tree (or policy-conditioned scenario tree) building, we can see how the tree got built with more and more careful pruning process with improvements from different works.
11
-
12
-
*[MPDM](mpdm.md) iterates over all ego policies, and uses the most likely one policy given road structure and pose of vehicle.
13
-
*[MPDM2](mpdm2.md) iterates over all ego policies, and iterate over (a set of) possible policies of other agents predicted by a motion prediction model.
14
-
*[EUDM](eudm.md) itrates all ego policies, and then iterate over all possible policies of other agents to identify **critical scenarios** (CFB, conditioned filtered branching). [EPSILON](epsilon.md) used the same method.
15
-
*[MARC](marc.md) iterates all ego policies, iterates over a set of predicted policies of other agents, identifies **key agents** (and ignores other agents even in critical scenarios).
The majority is the same as the previous work [MPDM](mpdm.md). There is a follow up article on this as well [MPDM3](https://link.springer.com/article/10.1007/s10514-017-9619-z) which expands [MPDM2](mpdm2.md) with more experiments, but with the same methodology.
19
9
10
+
So the main idea of MPDM is already covered in the original short paper [MPDM](mpdm.md).
20
11
21
12
#### Key ideas
22
13
- Motion prediction of other agents with a classical ML methods (Maximum likelihood estimation).
0 commit comments