This paper reports on an integrated inference and decision-making approach for autonomous driving that models vehicle behavior for both our vehicle and nearby vehicles as a discrete set of closed-loop policies. Each policy captures a distinct high-level behavior and intention, such as driving along a lane or turning at an intersection. We first employ Bayesian changepoint detection on the observed history of nearby cars to estimate the distribution over potential policies that each nearby car might be executing. We then sample policy assignments from these distributions to obtain high-likelihood actions for each participating vehicle, and perform closed-loop forward simulation to predict the outcome for each sampled policy assignment. After evaluating these predicted outcomes, we execute the policy with the maximum expected reward value. We validate behavioral prediction and decision-making using simulated and real-world experiments.
@article{galceran2017mpdm, AUTHOR = {Enric Galceran and Alexander G. Cunningham and Ryan M. Eustice and Edwin Olson}, TITLE = {Multipolicy decision-making for autonomous driving via changepoint-based behavior prediction: Theory and experiment}, JOURNAL = {Autonomous Robots}, YEAR = {2017}, MONTH = {August}, VOLUME = {41}, NUMBER = {6}, PAGES = {1367--1382}, }