2020
DOI: 10.48550/arxiv.2005.10622
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Triple-GAIL: A Multi-Modal Imitation Learning Framework with Generative Adversarial Nets

Abstract: Generative adversarial imitation learning (GAIL) has shown promising results by taking advantage of generative adversarial nets, especially in the field of robot learning. However, the requirement of isolated single modal demonstrations limits the scalability of the approach to real world scenarios such as autonomous vehicles' demand for a proper understanding of human drivers' behavior. In this paper, we propose a novel multi-modal GAIL framework, named Triple-GAIL, that is able to learn skill selection and i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(5 citation statements)
references
References 9 publications
0
5
0
Order By: Relevance
“…The numbers of hidden units of the two fully connected layers are 300 and 600, respectively, and tanh is used as the activation function. The goal of the joint discriminator network is to maximize the function shown in (5).…”
Section: B Pre-training the Policy Network Via An Adversarial Processmentioning
confidence: 99%
See 1 more Smart Citation
“…The numbers of hidden units of the two fully connected layers are 300 and 600, respectively, and tanh is used as the activation function. The goal of the joint discriminator network is to maximize the function shown in (5).…”
Section: B Pre-training the Policy Network Via An Adversarial Processmentioning
confidence: 99%
“…Reinforcement learning methods, which can optimize sequential decisions, have shown excellent performance on the driving decisionmaking of a single car agent [1] [2]. Imitation learning methods, which directly learn the behavior of a driver from recorded driving experiences and generate the driving policies, provide alternative solutions to autonomous driving [3]- [5]. Therefore, multi-agent systems are expected to be constructed to achieve autonomous driving through multi-agent reinforcement learning [6] or imitation learning [7].…”
Section: Introductionmentioning
confidence: 99%
“…Li et al [20] proposed the InfoGAIL framework to infer potential modal variables by maximizing the mutual information between potential variables and state-action pairs. Fei et al [8] proposed Triple-GAIL, which allows co-learning from expert demonstrations adaptive skill selection and imitation, and generates experience by introducing auxiliary skill selectors. Bhattacharyya [21] proposed PS-GAIL, which extends Gail through a parameter sharing approach based on course learning.…”
Section: Related Workmentioning
confidence: 99%
“…In contrast, IRL focuses on finding the reward function that best explains the demonstrated behaviour, however, the function search is ill-adapted because the demonstrated behaviour may correspond to multiple reward functions, and IRL methods inevitably suffer from intensive computation, which limits their efficiency in high-dimensional scenarios. Generative adversarial imitation learning (GAIL) [7] has become a very popular model-free imitation learning framework that can extract policies directly from expert demonstrations, overcoming cascading errors caused by behavioural cloning (BC) and reducing the computational burden of inverse reinforcement learning (IRL) [8]. Work has been done to show that GAIL can perform well in complex large-scale problems such as autonomous driving [4], simulation, and real robot manipulation [9].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation