Research Article

Learning from Demonstrations and Human Evaluative Feedbacks: Handling Sparsity and Imperfection Using Inverse Reinforcement Learning Approach

Table 3

Learning the driving styles from a human teacher in different conditions.

Driving styleDemonstrationsNumber of feedbacksEV of EV of teacher policy
TypeTimeOptimality degreeAllNegative

Style 1Abundant120 sec70%1646216.82117.164
Sparse20 sec100%3028416.906
No demo40923916.684

Style 2Abundant120 sec70%1244814.89315.275
Sparse20 sec100%2185914.969
No demo30618214.877