Within this paper we present a new concept on deriving test cases from simulation data and outline challenging tasks when testing and validating fully automated driving functions and Advanced Driver Assistance Systems (ADAS). Open questions on topics like virtual simulation and identification of relevant situations for consistent testing of fully automated vehicles are given. Well known criticality metrics are assessed and discussed with regard to their potential to test fully automated vehicles and ADAS. Upon our knowledge most of them are not applicable to identify relevant traffic situations which are of importance for fully automated driving and ADAS. To overcome this limitation, we present a concept including filtering and rating of potentially relevant situations. Identified situations are described in a formal, abstract and human readable way. Finally, a situation catalogue is built up and linked to system requirements to derive test cases using a Domain Specific Language (DSL).
An immense test space is pushing the development and testing of automated driving functions from real to virtual environments. The virtual world is provided by interconnected simulation models representing sensors, vehicle dynamics, and both static and dynamic environment. For the virtual validation of automated driving, special attention must be paid to the simulation's credibility, which can be impaired by inappropriate or inaccurate simulation models and tools. Therefore, in this work a method is proposed to assess the credibility of simulation-based testing for automated driving. The approach allows a qualitative and relatively quantitative comparisons between scenarios as well as between different simulation setups. Therefore, several uni-and multivariate metrics are applied towards a scoring of similarity of the behavior between simulation and real test drive. This is achieved by using ground truth data in form of simulation scenarios from real world measurement data. In this way, the virtual automated vehicle encounters the same conditions and surroundings than its counterpart in the real world for evaluating their similarity. The practical applicability of the proposed credibility assessment approach is demonstrated in a case study, in which the credibility of an exemplary simulation-based test bench is inferred.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.