Sunday, 2025/06/15

  • CS/AI
  • C
  • TITECH
  • Switch Language
    • ja日本語 (Japanese)
    • enEnglish

Shimosaka Research Group

pursuing MIUBIQ (machine intelligence in UbiComp Research)

  • Home
    • Members
    • Location
  • News
  • Projects
  • Publications
  • Awards
  • Archives
    • Codes
    • Datasets
Navigation
Projects Fast and stable IRL based driving behavior modeling in continuous state space

Fast and stable IRL based driving behavior modeling in continuous state space

2021/03/15 | Projects | 1701 views |

Fast and stable driving behavior prediction in the high-dimensional state space

Driving behavior prediction has recently gained popularity and been expected towards Advanced Driver Assistance System (ADAS). For these predictions, designing a reward function which assigns high values to ideal behaviors, and obtaining a sequence of control inputs by maximizing these rewards are common ways. In particular, iterative Linear Quadratic Regulator and Model Predictive Control are well-known methods but they cannot stably get the high reward paths because they only search a small part of space.
On the other hand, Rapidly exploring Random Tree (RRT) can conduct stable planning by creating a large tree that covers the whole space; however, covering high-dimensional state space to express car models requires a large computational cost. In addition, template-based RRT (Ma et al. ITSC2014) an extended method of RRT. While it can deal with nonholonomic dynamics with reducing the computational cost, there remains inefficiency of RRT planning and applying it to the high-dimensional space is difficult. For these reasons, we should resolve a trade-off between stability and speed of planning in order to predict driving behaviors.
To deal with the problem, we proposed fast and stable template-based RRT (FASTEMLERRT) which consists of pruning and parallel multiple trees generation processes. By pruning unnecessary edges, it can focus on the specific part of space, where the high reward is expected.

Also, parallel generation of multiple trees can reduce the computational cost and efficiently cover a large area.

The combination of these two processes realizes stable and fast prediction in high-dimensional space. The experimental results on artificial highways and intersections show that our approach achieves rapid and robust planning in comparison with existing methods.

Inverse reinforcement learning for driving behavior modeling in the high-dimensional state space

In the actual driving behavior prediction, reward functions tend to be very complicated because they depend on human preference; such as slow acceleration and comfortable steering. Therefore, it is difficult to design these functions and to directly apply the above planning methods.
To resolve this problem, inverse reinforcement learning (IRL) is known as a prominent approach because it can directly learn complicated behaviors from expert demonstrations. Driving data tend to have several optimal behaviors because of their dependency on drivers’ preferences. To capture these features, maximum entropy IRL has been getting attention because its probabilistic model can consider suboptimality.

However, maximum entropy IRL needs to calculate the partition function, which requires large computational costs. Thus, we cannot apply this model to a high-dimensional space for detailed car modeling. In addition, existing research attempts to reduce these costs by approximating maximum entropy IRL; however, a combination of efficient path planning and the proper parameter updating is required for an accurate approximation, and existing methods have not achieved them.
In this study, we leverage RRT motion planner, and efficiently sample multiple informative paths from the generated trees. Also, we propose novel RRT-based importance sampling for an accurate approximation. These two processes ensure a stable and fast IRL model in a large high-dimensional space.

Experimental results on artificial environments show that our approach improves stability and is faster than the existing IRL methods.

Publications

逐次枝刈り型RRTの並列化による高速かつロバストな運転行動予測
細馬 慎平, 須ヶ﨑 聖人, 竹中 一仁, 平野 大輔, 孫 理天, 下坂 正倫
第38回 日本ロボット学会学術講演会 予稿集, オンライン開催, 10 2020

RRT-based maximum entropy inverse reinforcement learning for robust and efficient driving behavior prediction.
Shinpei Hosoma, Masato Sugasaki, Hiroaki Arie, and Masamichi Shimosaka.
2022 IEEE Intelligent Vehicles Symposium (IV 2022), pp. 1353-1359, Aachen, Germany, 6 2022.

  • tweet

Comments are disabled for this post

Social Networks

  • twitter
  • rss

Recent News

  • Presenting our paper on Exploiting Periodic UWB CIRs for Robust Activity Recognition with Attention-aware Multi-level Wavelet at PerCom2025 2025/02/15
  • Presenting our paper on revealing Universities’ Atmosphere from Visitor Interests has been presented at IEEE BigData 2024 2024/12/16
  • Our paper on adaptive incremental-decremental BLE placement optimization for accurate indoor positioning has been presented at IPIN2024. 2024/10/23
  • Presenting two papers at SIGSPATIAL 2024 2024/10/23
  • Forecasting Crowded Events using Public Announcements with Large Language Models 2024/10/15
  • Forecasting Lifespan of Crowded Events Inspired by Acoustic Synthesis Technique 2024/07/04
  • Our paper on forecasting lifespan of crowded events has been published in IEEE Access 2024/07/04
  • Presenting our paper on Stable IRL from failed demonstrations at IV2024 2024/05/30
  • Presenting our demo on the application “CityScouter” at UbiComp 2023 2023/10/11
  • Presenting our paper on efficient Bluetooth beacon deployment for campus-scale crowd density monitoring application at UbiComp2023 2023/10/05

Search

Copyright 2015 · Shimosaka Research Group at TITECH