skip to main content
research-article
Open Access

A Human Factors Approach to Validating Driver Models for Interaction-aware Automated Vehicles

Published:08 September 2022Publication History

Skip Abstract Section

Abstract

A major challenge for autonomous vehicles is interacting with other traffic participants safely and smoothly. A promising approach to handle such traffic interactions is equipping autonomous vehicles with interaction-aware controllers (IACs). These controllers predict how surrounding human drivers will respond to the autonomous vehicle’s actions, based on a driver model. However, the predictive validity of driver models used in IACs is rarely validated, which can limit the interactive capabilities of IACs outside the simple simulated environments in which they are demonstrated. In this article, we argue that besides evaluating the interactive capabilities of IACs, their underlying driver models should be validated on natural human driving behavior. We propose a workflow for this validation that includes scenario-based data extraction and a two-stage (tactical/operational) evaluation procedure based on human factors literature. We demonstrate this workflow in a case study on an inverse-reinforcement-learning-based driver model replicated from an existing IAC. This model only showed the correct tactical behavior in 40% of the predictions. The model’s operational behavior was inconsistent with observed human behavior. The case study illustrates that a principled evaluation workflow is useful and needed. We believe that our workflow will support the development of appropriate driver models for future automated vehicles.

Skip 1INTRODUCTION Section

1 INTRODUCTION

One of the great technological and societal promises of the 21st century is the autonomous vehicle (AV) [4, 10, 30]. This technology has been under development in laboratories and under controlled conditions for decades and is now transitioning to the real-world. However, a major challenge for real-world implementation of AV technologies is enabling AVs to handle complex interactions with human road users. AV controllers have recently been proposed that aim at addressing this challenge through interaction-awarecontrollers (IACs) [5, 6, 7, 9, 13, 14, 19, 21, 23, 34, 36, 42, 45, 46, 47]. IACs incorporate a model of human driver behavior in the controller, to predict how another driver is likely to respond to the AV’s behavior. Based on this prediction and its own reward function (e.g., incorporating safety, comfort), the IAC finds the optimal action for the AV (Figure 1). However, up to now the interactive capabilities of these controllers have only been demonstrated in simplified simulated environments (e.g., top-down view computer simulations). Whether the state-of-the-art IACs are capable of predicting naturalistic driver behavior and interacting with humans in real traffic remains an open question.

Fig. 1.

Fig. 1. A high-level diagram of a typical IAC for AVs. Such a controller operates in situations where the states and actions of a human-driven vehicle (superscript h) and an AV (superscript av) influence each other, e.g., the merging situation depicted in the left panel. Future states and actions are denoted with subscript \( t+1 \) , all other states and actions are at time t. An IAC determines the optimal action a for the AV based on the current state s of both the AV and the human. To find this optimal action, IACs make use of at least two prediction models: a dynamic model to predict future states ( \( s^*_{t+1} \) ) based on current states ( \( s_t^* \) ) and actions ( \( a_t^* \) )(the superscript \( * \) denotes it can either be used for the AV or the human), and a human driver model to predict the actions surrounding human drivers will take in response to the AV’s action. Both the dynamic and human behavior predictions are evaluated to find the optimal action for the AV, this is usually done with a reward function that incorporates aspects like safety and comfort. The validation of the human driver model is the focus of this work.

Although demonstrating a proposed controller in a simulated traffic environment is a necessary first step to show its potential, it does not provide sufficient evidence on how well the controller will generalize to real-world environments. In this work, we take the position that before implementing an IAC in vehicles to validate its behavior in the real-world, its underlying driver model should be validated on natural human driving behavior. If the model fails to predict real-world behavior accurately, the controller will act on false predictions which can lead to annoying or even unsafe situations. Such driver model validation can therefore provide an early indication of the IAC validity without much of the cost associated with implementing and testing it in real traffic interactions. However, driver model validation is currently not a part of the mainstream approach to IAC validation (see e.g., [34, 36, 45]), and a principled framework for such validation is missing from the literature.

The contribution of our work lies in proposing and demonstrating a human-factors-based evaluation workflow, in order to help IAC designers in the process of selecting appropriate driver models. The proposed workflow validates driver models using empirical data obtained from naturalistic (real-world) traffic interactions, acknowledging two levels of driving behavior [24]: tactical choices and operational safety margins. Tactical behavior refers to which maneuvers are executed (e.g., a lane change or car following) and operational behavior describes how they are executed (e.g., in terms of safety margins). To demonstrate the potential of this workflow, we perform a case study that shows that an inverse-reinforcement-learning-based model, replicated from a model used in a previously developed IAC [34], does not generalize to real-world data. Even though we do not quantify the implications of these results for any specific IAC, they still underline the importance of using validated driver models in AV controllers.

Skip 2VALIDATING DRIVER MODELS FOR INTERACTION-AWARE CONTROLLERS Section

2 VALIDATING DRIVER MODELS FOR INTERACTION-AWARE CONTROLLERS

2.1 Why Validate?

Part of the reason why model validation is necessary is that the simulated environments in which IACs are evaluated are not sufficient to assume safe generalization to the real-world. A particular aspect of the evaluation is the human response to the AV’s actions. Two approaches to generate this response are used. Some studies [6, 7, 9, 14, 19, 21, 23, 36, 42, 45, 47] simulate human driver responses using driver models. However, many of the driver models used for this purpose are also not validated on natural human driving behavior, which could indicate a discrepancy between the simulation and natural behavior. Other studies [5, 13, 34, 46] use real-time responses of a human test subject in an abstract top-down view computer simulation, much like a video game. The gap between such abstract test environments and real-world driving is large, e.g., due to the absence of risk perception [31], motion cues, and visual looming [18]. So, again we can expect the participants’ responses to differ from driver responses in real-world traffic. This means that both approaches can only provide very limited evidence for the generalization of the demonstrated interactive capabilities of the IAC to the real-world.

To show that the IAC’s behavior does generalize to the real-world, one could propose to implement the IAC in a real vehicle and demonstrate its workings in a natural environment. However, deploying a proof-of-concept IAC in the real-world might result in unsafe situations even under highly controlled conditions. This raises ethical concerns about such real-world testing. Another possibility would be to use real-time human responses and minimize the mismatch between the simulation environment and the real-world, e.g., by using a high-fidelity driving simulator. However, such experiments are expensive and time-consuming, and human behavior even in realistic driving simulators can still differ from behavior in real traffic [8, 31]. For this reason, we advocate a complementary approach: validating the driver model on naturalistic traffic data before implementing it in an IAC. The combination of the model validation on real-world data and demonstrating the IAC’s interactive capabilities in a (simplified) simulated environment provides a firm ground for the further implementation and testing of the IAC in real vehicles.

To the best of our knowledge, validation on naturalistic driving data for use in IACs has not been performed for two of the most commonly used driver models proposed for IACs. These models are the intelligent driver model IDM [43] (used in [6, 13, 45] to predict driver behavior and in [5, 14, 47] to simulate other drivers’ responses) and the expected-utility-maximizing model (used e.g., in [34, 36] to predict other drivers’ behavior) that uses a reward function learned from human demonstrations with inverse reinforcement learning (IRL). Although the reward function in this model is learned from naturalistic driving data, none of the studies which proposed IACs based on an IRL-based model have validated the resulting model with respect to its ability to capture human behavior.

2.2 How to Validate?

We propose a three-step evaluation workflow (Figure 2) that incorporates important aspects of driver model validation: evaluation against naturalistic data on both the tactical and operational levels.

Fig. 2.

Fig. 2. The proposed driver model validation workflow for interaction-aware autonomous vehicle controllers. The workflow consists of three steps. In the first step, a suitable dataset is selected to perform the validation of the driver model on. From this selected dataset, specific situations are automatically extracted. The actual validation of the model takes place in the last two steps. A distinction is made based on the level of behavior. First, the tactical behavior is validated in step 2. This step reveals to what extent the driver model shows tactical behavior that is consistent with human behavior in the dataset. Behavior inconsistent with human data, e.g., collisions, is not regarded in the final step. The third step evaluates the operational behavior of the model based on human-factors literature. This is done for every tactical behavior separately. The final conclusion of the validation should be based on the combined results of steps 2 and 3.

Step 1: Select Naturalistic Data.

When validating a driver model for an IAC, we propose that the model is compared against human behavior data recorded in a natural environment, i.e., a naturalistic driving dataset. There are increasingly many naturalistic datasets available, but which dataset should one choose? And once the dataset is chosen, should all the data in the dataset be used uniformly for model validation?

When selecting a naturalistic dataset, one should be aware of whether the data recording was done with obtrusive or unobtrusive methods. Obtrusive methods are methods where the driver is aware their behavior is being recorded (e.g., the SHRP2 dataset [2]). As a result, the driver might have changed their behavior e.g., to conform to the expectations of the researchers. Other datasets are gathered without the drivers knowing that their behavior is being recorded, typically with drones and cameras (several open-access datasets are available e.g., [3, 17, 44]). Because of the possibility of adapted behavior in obtrusive naturalistic datasets, unobtrusive datasets are preferable for model validation.

When a suitable dataset is chosen, specific parts of the data need to be selected to perform the validation on. Data recorded in the real-world often contains many different scenarios, e.g., different locations, vehicle types, and maneuvers. Using all this data to validate a driver model would be intractable because humans behave differently in different scenarios. Instead, comparable scenarios can be selected from the dataset to be evaluated together. These scenarios should fit the intended environment of the IAC. At the same time, one should avoid hand-picking scenarios, or selecting them on low-level characteristics (e.g., only include vehicles that reach a certain velocity) because this will reduce the variability in the data and thus negate the purpose of the validation, to show that the model generalizes to real-world behaviors. Instead, scenarios should be selected on higher-level similarities, e.g., including all lane changes or all unprotected left turns. Open-source software is available that includes examples of how to extract such scenarios automatically e.g., [39].

Behavior Validation.

After selecting relevant scenarios, the model can be trained and validated. Validation of models of human behavior is often difficult because there are many aspects that determine if the model’s behavior resembles human behavior. In most cases, the difference cannot be captured by a single metric. For example: when validating a driver model in a lane-changing scenario, it could be tempting to use a distance-based error-metric to describe the goodness-of-fit. However, an event like a collision with a vehicle in an adjacent lane can, in some cases, be described by a small lateral distance error with respect to a human-driven trajectory. If only this distance error would be examined when validating the model, it would seem to perform well, but in reality, the model predicts that a human would collide with another vehicle. The collision is missed in the single-metric validation procedure, and the (wrong) conclusion would be that the model describes human behavior with only a small error margin.

This example illustrates that a distinction should be made between what behavior is executed (e.g., car following, crashing, or lane changing) and how it is executed (i.e., specific trajectories and safety margins with respect to lane boundaries and traffic participants). This bears resemblance to the common distinction in driving behavior [24] of tactical and operational behavior (note that strategic behavior, e.g., route selection, is not covered by the models in IACs). In this distinction, the maneuvers executed by the driver, like a lane change, are tactical behavior. The manner in which they are executed, e.g., expressed in accelerations or dynamics of the gaps with respect to other vehicles, is called operational behavior. Making this distinction in driver model validation is especially relevant for driver models used in IACs because these models are mostly designed to incorporate multiple tactical behaviors. This is in contrast to traditional driver models that were more often designed to only represent one specific tactical behavior.

For many tactical behaviors, the corresponding operational behavior has been studied in human-factors experiments (e.g., for car following [11, 15, 16, 25, 29, 35]). These studies provide the important metrics of human operational behavior, given a specific tactical behavior. Making the same distinction during the validation allows one to leverage the existing human-factors literature, enabling researchers without in-depth human-factors expertise to validate their models.

Determining what tactical behavior is executed by the model and if it matches human behavior is something that can be done without any expert knowledge. For instance, it is straightforward to specify if a lane change is made and to compare if the model performs a lane change in the same situation where a human does. Once the tactical behavior is determined, the metrics specifying the operational behavior can be defined based on the relevant human-factors literature. This will require obtaining some knowledge on the subject, but with a properly specified tactical behavior, a brief, non-exhaustive driver-behavior literature survey would be enough for a researcher to make a motivated choice of the metrics characterizing the corresponding operational behavior.

Because making a distinction between tactical and operational behavior is relevant for IACs and makes the validation process easier, we propose a sequential two-stage validation process. The first stage (step 2 in the workflow of Figure 2) is to validate the model’s behavior on a tactical level, providing a quick and straightforward distinction between behavior that clearly resembles or does not resemble the observed human driving behavior in the same circumstances. The second stage (step 3 in Figure 2) examines the tactical behaviors separately on the operational level.

Step 2: Tactical Validation.

The purpose of the tactical validation step is two-fold. First, it serves to determine which of the model’s responses are consistent with human behavior and which are not. A valid driver model does not predict tactical responses inconsistent with human behavior, therefore we will refer to such responses as undesirable tactical behavior. Desirable behaviors, on the other hand, are all tactical responses that can be observed in naturalistic human driving data. Second, this step will categorize the model’s responses so its desirable behaviors can be validated in the operational validation step according to the right criteria. Undesirable behavior can be disregarded during the operational validation step because it does not matter how the model performs a behavior that is undesirable in the first place.

To achieve this, a mutually exclusive set of possible tactical behaviors exhibited by the model should be defined. The distinction between these tactical behaviors should be based on simple rules (or inclusion and exclusion criteria) such that all exhibited model behavior falls in one and only one tactical category. Which and how many of these categories to include depends on the outcome of the literature survey discussed earlier. All behaviors in one category should be validated on the same operational characteristics, which should be taken into account when determining the categories.

Step 3: Operational Validation.

For the operational validation step, human-factors literature provides signals and metrics that best describe human behavior for specific tactical behavior. This operational validation step can compare individual trajectories or averaged metrics between human and model behavior as long as the metrics and signals are chosen appropriately and the tactical behaviors are regarded separately. Examples of such metrics are metrics that relate to the dynamics of the behavior, e.g., the gap between vehicles, or to the properties of the maneuver, e.g., the duration of a lane change. Human-factors literature can also provide methods on how to compare the signals and metrics. For example, in [35] figures are presented that relate phase diagrams in car following to responsive actions of human drivers, such plotting methods can also be used for model validation.

The Validation Conclusion.

The final conclusion of the validation procedure should be based on both the tactical and operational behavior displayed by the model. The model should display desirable tactical behavior in a way that resembles how humans perform the same behavior on an operational level. But because the eventual goal is to incorporate the driver model in an IAC, the controller’s ability to safely operate while using the model’s predictions can be seen as the most important factor in the final conclusion.

When a driver model shows behavior that deviates from human behavior to a large extent, but the controller that implements the model can still safely operate with these errors, it can still be concluded that the model is “good enough” for use in the IAC. To draw such a conclusion, the maximal acceptable difference between the model’s output and human behavior has to be defined. This should be done for every IAC separately due to differences in IACs, scenarios, and regarded tactical behaviors. The maximal acceptable difference can for example be based on an evaluation that shows that the controller can still reliably execute safe and acceptable interactive behavior when confronted with predictions that have this maximal deviation from future human behavior.

However, even if an IAC is robust to inaccurate predictions of the driver model, we argue that it is still important to validate the model and report the magnitude of the deviation from human behavior. This improves the re-usability of the proposed model for other IACs and provides a basis for a re-evaluation of the model when extending or improving the IAC.

Skip 3CASE STUDY: METHODS Section

3 CASE STUDY: METHODS

To demonstrate the proposed workflow we use it to validate an inverse reinforcement learning (IRL) based model replicated from a study that proposed one of the first IACs for autonomous vehicles [34]. The choice to validate an IRL-based model was made because this increasingly popular type of model describes dynamic human behavior in multiple scenarios and has not been validated previously. The two IACs with IRL-based driver models discussed earlier [34, 36] use similar implementations of such a model. However, only the work by Sadigh et al. [34] provides enough detail, in the form of mathematical description and open-source code, to replicate the used IRL-based model. For that reason, the model used by Sadigh et al. is used as a reference for this case study.

3.1 Model Implementation

IRL-based driver models assume that human behavior is “driven” by an underlying reward function. A parameterized reward function is assumed and inverse reinforcement learning is used to infer the parameters directly from human demonstrations (see [1, 28, 49]). This reward function with the learned parameters can be used in an agent to generate individual predictions of human behavior. Driver models based on IRL use a utility-maximizing rational agent for this purpose. Throughout this article, we refer to this method of generating predictions combined with a specific assumed reward function as the model. We refer to instances of the model with a specific set of parameters as an agent. In IRL-based driver models, the used reward function consists of a linear combination of features, each with its own weight: (1) \( \begin{equation} R^h(s, a) = \sum \theta _i^h \phi _i(s, a). \end{equation} \)

In this formula, \( R^h \) denotes the reward of a specific human, s is the state (at time t) and a is the action sequence the human will take. This action sequence is subject to a finite planning horizon. \( \phi _i \) denotes the ith feature and \( \theta _i \) represents the corresponding weight, which is learned by IRL from demonstrations produced by a human driver h. Note that the features \( \phi _i \) in Equation (1) are designed beforehand and do not vary over humans, demonstrations, or situations. The weights \( \theta _i \) are learned from the demonstrations and vary over humans. These weights are learned by maximizing the log-likelihood of an observed demonstration with respect to the weights, given the assumed features.

3.2 Assumed Reward Function

The reward function \( R^h \) used for the IRL-based model in this work was replicated from [34] and consists of four features for maintaining a preferred velocity, lane-keeping, staying on the road, and collision avoidance. The collision avoidance feature is modeled by a two-dimensional Gaussian function, based on distances between the centers of vehicles. Because the human demonstrations we use for the case study were recorded on highways, the heading angles of the vehicles take very low values and are therefore neglected for collision avoidance. They are assumed to be equal to the road heading (this is a deviation from the model used in [34]). The lane-keeping and road boundary features are both Gaussian functions of the lateral road axis, they are constant over the longitudinal axis of the road. The velocity feature is the squared error with respect to the desired velocity. Since the exact desired velocity is not known for the human drivers that provide the demonstrations, and the legal speed limits that could be used for this purpose are not always provided with the data, the maximum recorded velocity of a vehicle is taken as the driver’s desired velocity. The full reward function is given in Equation (2). (2) \( \begin{equation} R^h(x, y, v_x) = \theta _\text{vel}^h \phi _\text{vel}(v_x) + \theta _\text{lane}^h \phi _\text{lane}(y) + \theta _\text{bounds}^h \phi _\text{bounds}(y) + \theta _\text{collision}^h \phi _\text{collision}(x, y), \end{equation} \) where \( \begin{align*} \phi _\text{vel}(v_x) &= (v_x - v_d) ^ 2, \\ \phi _\text{lane}(y) &= e^{-c (y_{lc} - y) ^2},\\ \phi _\text{bounds}(y) &= e^{-c (y_{rb} - y) ^2}, \end{align*} \) \( \begin{equation*} \phi _\text{collision}(x, y) = \frac{1}{\sigma _x \sqrt {2\pi }} e^{-(1/2) ((x - x_o)^2/\sigma _x^2)} \frac{1}{\sigma _y \sqrt {2\pi }} e^{-(1/2) ((y - y_o)^2/\sigma _y^2).} \end{equation*} \)

In these formulae, x and y denote the longitudinal and lateral position as defined in Figure 4, lc and rb denote the lane center and road boundaries respectively, where the road boundaries are defined at half a lane width outside the outermost marking. v represents velocity and subscript o denotes the other vehicle. The constants c, \( \sigma _x, \) and \( \sigma _y \) are used to shape the features. A visual representation of the reward function, excluding the velocity feature, can be found in Figure 3.

Fig. 3.

Fig. 3. A heat map of the reward function (Equation (2)) is used for the IRL-based driver model where the black block indicates the ego vehicle. Warmer colors indicate low reward, cooler colors indicate high reward. The feature for velocity is not shown here because it does not depend on the position. The dimensions and positions of the features displayed here are assumed to be constant over different humans. The weights, represented here by the colors, differ between humans and are learned from demonstrations.

The constants that shape these features were determined with a grid search on the first 15 demonstrations of the used dataset. Initial guesses of the parameters were based on a visual comparison of the heat map to the road image. Variations around these initial guesses were estimated based on the dimensions of the lanes. (For example, \( min(\sigma _y) = 1.4~m \), thus \( 95.4\% \) of the lateral influence on collision prevention lies within a \( 2.8~m \) distance between vehicle centers. With a lane width of \( 4~m \) and a \( 2~m \) wide vehicle, this means the lane marking has to be crossed before the collision prevention starts contributing to the reward. Thus, the lower bounds of our parameter grid are close to the smallest plausible parameter values.) We used the following sets in the grid search: \( c=\lbrace {\bf 0.14}, 0.18, 0.22\rbrace , \sigma _x=\lbrace 5.0, 10.0,{\bf 15.0},20.0\rbrace , \sigma _y=\lbrace {\bf 1.4}, 1.8, 2.2\rbrace \), where the bold value the selected value. Each parameter combination in the grid was evaluated based on the resulting number of desired tactical behaviors by the agent (see Section 2.2, Step 2 for the definition of desired behavior). The parameter sets \( c, \sigma _x, \sigma _y = 0.14, 15.0, 1.4 \) and \( c, \sigma _x, \sigma _y = 0.14, 20.0, 1.4 \) had the maximum number of desired tactical behaviors in this grid search, we chose to select the combination containing our initial guess.

3.3 Using the Proposed Workflow

Here we will discuss the use of the proposed workflow (Figure 2) to validate the IRL-based model with the reward function as shown in Equation (2) step by step.

3.3.1 Step 1: Select Data.

The first step of the proposed workflow is to select a naturalistic dataset. Among multiple naturalistic driving datasets that are openly available, in this case study we considered three datasets: the NGSIM dataset [44], the pNEUMA dataset [3], and the HighD dataset [17]. Of these three the NGSIM data have larger uncertainties in the trajectories because it was recorded with fixed-base cameras instead of drones. The pNEUMA dataset was recorded in an urban environment, this does not match the environment of the regarded IAC [34] which focuses on multi-lane scenarios (e.g., a highway) with human behavior that mostly consists of actions to prevent collisions, like lane changing. The HighD dataset contains high-precision data recorded in a multi-lane environment. It also contains dynamic behavior such as lane changes to prevent collisions. For these reasons, we will use the HighD dataset. This dataset consists of 60 separate recordings, recorded in six different locations in Germany. All recordings were made on highways using drones equipped with cameras; from these recordings, trajectory data was automatically extracted [17]. Every recording is of a fixed stretch of highway, the average length of these recorded stretches is \( 416~m \), and the average duration of a single-vehicle track is \( 14.34~s \).

To visualize the data and the resulting agent behavior we used TraViA [39], an open-source visualization and annotation tool for trajectory datasets. TraViA can visualize all mentioned datasets and we extended it to train and visualize the IRL-based model. The extension code is available online [38]. An example frame of the HighD dataset visualization can be found in Figure 4.

Fig. 4.

Fig. 4. An example frame of the HighD dataset [17] as visualized using TraViA [39]. The frame includes a stretch of a highway in Germany, where vehicles drive on the right side of the road and where, in some of the cases, there are no legal speed limits. The orange shapes represent regular cars, the green shapes are trucks. All vehicles have a vehicle-ID shown in white. The white arrows display the coordinate frame and the yellow marking shows a visualization of the gap between two vehicles as used in the metrics for step 3 of the validation workflow.

From this dataset, we automatically select suitable scenarios for training and validating the model. These scenarios should fit the intended use of the IAC [34]: in our case study, we assume the goal of the IAC is to interact with human drivers who perform lane changes. This means we could consider two distinct behaviors in the HighD dataset: lane changing and merging. A merging lane is only present in 3 of the 60 HighD recordings. For this reason, we will use human lane-changing maneuvers for validation. For consistency, the three recordings with a merging lane were not considered.

As mentioned before, the features in the reward function consider collision avoidance, lane-keeping, staying on the road, and maintaining a preferred velocity. This means that not all lane changes can be explained with this model. Lane changes to the right are not covered because they are “driven” by a need to adhere to (socially acceptable) traffic rules that are not incorporated in the reward function (In Germany, it is obligatory to drive in the rightmost lane if it is free. So a lane change to the right is most often performed simply because that lane is free, not to avoid a collision. It can therefore not be explained by the used reward function). Therefore, only single lane changes to a left lane are considered for training and validation. The highD dataset includes the number of lane changes for every trajectory (based on lane crossings) and the current lane number at every frame. We automatically extracted all used trajectories based on these metrics.

3.3.2 Step 2: Tactical Validation.

The next step is to define a set of tactical behavior categories. There are only a limited number of possible tactical behaviors on a highway without an exit lane, we will consider four possibilities: car following, lane changing, colliding, and crossing the road boundaries. Lane-keeping is not regarded as a separate behavior since all vehicles on a highway essentially follow another vehicle. In this set car following and lane changing are regarded as desirable behaviors, and colliding and going off-road are considered undesirable.

Besides defining the behavior categories, we established a procedure to place the trajectories produced by the agent in one of these categories based on a hierarchy in tactical behaviors. First, if an agent collided with another vehicle, this is labeled as a “collision”. If the agent did not collide, a check is done to see if the center of the vehicle stayed within the outer road boundaries; if not, the tactical behavior is labeled “off-road”. Agents that did not fall in one of the two categories above are checked for lane changes; if there is one, the tactical behavior is labeled “lane change”. And finally, agents that showed none of these three behaviors are placed in the “car following” category. All of these checks are implemented in the software and are performed automatically for all agents by checking for overlap with other vehicles and evaluating the vehicle’s center position for every time step.

The used hierarchy is based on the idea that a predicted collision has the highest impact on IACs. If a model predicts a collision, an IAC will act to avoid this, independent of the fact that the model predicts a lane change first. Vehicles leaving the road will also have a big impact on IAC behavior because it reduces the number of vehicles to consider and thus changes the scene. However, the IAC will not take drastic actions to avoid this, therefore it comes second in the hierarchy. Only if none of these undesirable behaviors are executed by the model, lane changes are relevant. Finally, all other behaviors within a single lane are grouped as car following. A more fine-grained distinction could have been made here by including behaviors such as nudging or aborted lane changes. But before considering those more sophisticated behaviors, we chose to evaluate if and how the model displays car following in general.

To evaluate if the model’s tactical performance is adequate for use in an IAC, a maximum acceptable deviation from human behavior needs to be specified. Because no IAC implementation is used in this case study, we cannot specify such a threshold here.

3.3.3 Step 3: Operational Validation.

The last step is to determine how to evaluate the model’s operational behavior for the cases where the tactical behavior falls in one of the desirable categories. We have defined two desirable categories: lane changes and car following. Earlier studies investigated human car-following behavior and risk perception using inverse time-to-collision vs. time gap plots [16, 26], these metrics were also used to evaluate human lane changes before [33].

Time-to-collision is defined as the time it will take until a vehicle collides with the preceding vehicle given that they both continue at their current velocity, (3) \( \begin{equation} \text{TTC} = \frac{x_\text{gap}}{v_\text{rel}}. \end{equation} \)

The time gap is the time it will take a vehicle to close the current gap with the preceding vehicle, given its current velocity, (4) \( \begin{equation} t_\text{gap} = \frac{x_\text{gap}}{v_\text{agent}}. \end{equation} \)

In these equations, TTC is time-to-collision, \( v_\text{rel} \) is the relative velocity of the agent and the preceding vehicles, and \( x_\text{gap} \) is the distance gap between the vehicles. This distance gap is visualized in Figure 4. Finally, \( v_\text{agent} \) is the longitudinal velocity of the agent vehicle.

Both TTC and time gap are available in the HighD dataset for human behavior; for the agent behavior, the metrics are calculated using the Equations (3) and (4).

Again, quantifying an acceptable error margin can only be done for a specific controller. Because we do not demonstrate a controller, we can only show the difference between the model and human behavior, but in this case study, we cannot quantify if this is acceptable for any specific IAC.

3.4 Model Training

The optimization procedure to find the weights that fit a human demonstration best is the same as used by Sadigh et al. [34]. The negated log-likelihood function as proposed by Levine and Koltun [20] is minimized with respect to the weights. To keep this tractable, the human demonstration is divided into sections with the same number of frames as the control horizon used in the agent (\( N=5 \)). All data frames are used, so the time step is \( \frac{1}{25}~s \) and the planning horizon is \( \frac{1}{5}~s \). The log-likelihood functions of the parts of the demonstration are summed and the summed negated log-likelihood is minimized. We assume that every lane-change trajectory in the dataset comes from a different human, an agent is trained separately for every trajectory, this resulted in 3,279 trained agents. Demonstrations on which the optimization procedure fails (i.e., no minimum of the negated log-likelihood function could be found) were discarded (2,302 demonstrations, 41%. For a discussion on why this number is so high, see Section 4).

Because highway data is used, the velocities of the vehicles are high (mean = \( 29.7 m/s \)) and heading angles are small. The heading angles of the vehicles are ignored in the dataset. For this reason, the dynamics of the vehicles are modeled as point masses. Because the trajectories are extracted from videos, no direct acceleration data was recorded. Acceleration data is available from the HighD dataset, but this has been reconstructed from velocity data. For this reason, the humans in the demonstrations are assumed to have direct control over the longitudinal and lateral velocities. Making the state and action vectors both 2-dimensional containing respectively an \( x, y \)-position, and -velocity. This assumption is justified because the goal of the model is to learn the reward function, not the dynamics of human control.

3.5 Validation of Agent Behavior

To validate the agent’s behavior, we evaluate the response of every agent individually in the same scenario that was used to train the agent. A dedicated test-set is not required contrary to most machine-learning approaches because the log-likelihood optimization proposed by Levine and Koltun accounts for sub-optimal demonstrations by humans. This means that the learned reward function does not need to be fully optimized in the human-driven demonstration, but the agent will fully optimize the reward function. So the agent might display different behavior than the human in the same situation and thus this situation can be re-used for validation.

For evaluation, the agent will be placed in the same initial position and its behavior is recorded for the same duration as the demonstration trajectory. Because the agent learned its reward function from this exact situation, this is a best-case scenario for the model. This approach also has the advantage that we can directly compare the agent’s behavior to the human demonstration it was trained on.

As for the IRL training, heading angles are neglected and the dynamics of the vehicle are assumed to be point mass dynamics. To approximate the states and actions of real drivers, the agents are assumed to have direct control over the linear accelerations of the vehicle. This results in a 4-dimensional state vector per vehicle, containing both \( x, y \)-position and -velocity, and a 2-dimensional action vector containing the \( x, y \)- accelerations. The agent is a utility-maximizing rational agent, so it will select an action a in state s, that maximizes its summed reward function R over a time horizon \( N=5 \). Again, the times-step is equal to the frame rate of the HighD dataset (\( \frac{1}{25}~s \)). The agent has full knowledge about the future trajectories of all adjacent vehicles. As for the choice of situation, this can be regarded as a best-case scenario for the agent, since it has a perfect prediction system to predict other human behavior.

The direct control over lateral accelerations, combined with the point mass dynamics, can result in trajectories that are not subject to normal vehicle dynamic constraints. To approximate normal vehicle dynamics, the agent’s actions (\( x, y \)- accelerations) are constrained to the maximal values of these accelerations found in the HighD dataset. The x-acceleration is constrained between \( (-6.63, 20.06)~m/s^2 \) and the y-acceleration between \( (-1.63, 1.63)~m/s^2 \).

Skip 4CASE STUDY: RESULTS Section

4 CASE STUDY: RESULTS

From the first 57 recordings in the HighD dataset, all 5,581 single lane changes to the left lane were automatically detected. These lane changes served as human demonstrations for the IRL-based driver model. Out of these 5,581 demonstrations, 3,279 resulted in a set of weights after the inverse reinforcement learning procedure. For the other 2,302 demonstrations, the IRL procedure failed to converge.

In practice, the failure of the IRL procedure means that the likelihood function adopted from [20] becomes intractable. This function contains the logarithm of the determinant of the Hessian matrix (\( log|-{\bf H}| \)). When this determinant becomes negative, the optimization fails. We found that this can happen when the optimization algorithm assigns a positive value to \( \theta ^h_{vel} \) (i.e., when deviating from the desired velocity is rewarded instead of punished). Note that weights are not restricted to be either positive or negative by the IRL procedure. IRL learns if a feature represents a reward or penalty for the human demonstration.

To examine if this was the cause of the high rate of failures in our training procedure, we estimated the Jacobian used in the optimization procedure for the initial values of \( \theta \). For \( 97\% \) of the demonstrations where IRL failed, the Jacobian value for \( \theta ^h_{vel} \) was negative and had a magnitude at least 10 times larger than all other Jacobian values. This did not happen in demonstrations where IRL succeeded (0 out of 100 randomly selected cases). Which indicates that the optimization algorithm attempted to use positive weights for \( \theta ^h_{vel} \) as they have a high likelihood to explain the demonstration in the failed cases.

This could mean that features in the reward function that are based on the deviation from a maximum observed (or allowed) velocity are not suitable for use on real-world traffic conditions. On the other hand, the IRL procedure might not have failed in these cases if \( \theta ^h_{vel} \) was restricted to always be negative (or more generally, if weights are restricted to represent either rewards or penalties). Further investigation to answer these questions is left for future work. We discarded the demonstrations where training failed and continued the attempt to validate the IRL-based driver model using the training data for which the model converged.

The 3,279 agents that trained successfully were placed in the same scenario they were trained on to examine to what extent they show human-like behavior on a tactical and operational level. We would like to remind the reader that, combined with the fact that all agents had access to perfect predictions of all surrounding vehicles, this constituted a “best-case scenario” for the model.

Tactical Behavior.

On a tactical level, we have defined four possible behaviors to categorize the resulting agent behavior: car following, lane changing, colliding, and crossing the road boundaries. Only in \( 40.2\% \) of the cases, the model showed the same tactical behavior as the human demonstration, a lane change (Table 1). In more than \( 41\% \) of the cases, the model either collided or went on an off-road adventure. This behavior was not present in the chosen subset of the human data, so we conclude that model behavior is inconsistent with human behavior.

Table 1.
number of agentspercentage of agentspercentage of human demonstrations
Lane-change1,31840.2%100.0%
Collision87526.7%0.0%
Car following59318.1%0.0%
Off-road49315.0%0.0%
Total3,279100%100%

Table 1. Tactical Behavior as Shown by the IRL Agents and in the Human-driven Demonstrations the Agents were Trained on

Operational Behavior.

We then compared the operational behavior of the model to the operational behavior in the human demonstrations using the inverse time to collision vs. time gap plots (Figure 5). Trajectories with multiple preceding vehicles show jumps in these plots due to suddenly changing values, for that reason those trajectories were omitted. Agents and humans that perform a lane change when the preceding vehicle is out of sight are also omitted since no inverse TTC and time gap data can be calculated for them for the final frames. All car-following trajectories are cropped to the point where the preceding vehicle gets out of sight.

Fig. 5.

Fig. 5. Inverse TTC vs. time gap plots of human demonstrations (panels a and c) and IRL agent behavior (panels b and d) in lane changes and car following. Panel a shows human behavior in the used demonstrations. Since these demonstrations do not contain any car following, panel c shows 55 illustrative examples of car-following behavior selected from other trajectories in the dataset, three are highlighted for clarity. Black diamonds indicate the initial position, this is the first frame in which a vehicle appears in the HighD dataset. In panels a and b, orange dots indicate a lane change, corresponding to the frame in which the center of a vehicle crosses the center-line between lanes. From panels a and b we conclude that the model’s lane change behavior has human-like dynamics in general, however, the model makes lane changes at substantially higher inverse-TTC (lower TTC) compared to humans. From panels c and d we conclude that the model’s car-following behavior does not resemble human car-following behavior.

The plots on the left side of Figure 5 (a) and (c) show human operational driving behavior. In the case of lane changing Figure 5(a), the inverse TTC increases while the time gap decreases, until the point where the center lane-marking is crossed, depicted with an orange circle. In the case of car-following Figure 5(c), humans oscillate around a preferred equilibrium point.

The model’s behavior for the same maneuvers can be seen on the right side of Figure 5. The model’s lane-changing behavior (Figure 5(b)) has human-like dynamics in general (as in Figure 5(a)); however, the model makes lane changes at substantially higher inverse TTC (lower TTC) compared to humans. Also, the time gap at the moment of the lane change is on average smaller than for the human demonstrations. To further illustrate the differences in the lane change dynamics, we investigated the distributions of inverse TTC and time gap at the moment of lane change (Figure 6). This shows substantial differences between the estimated distributions. We performed a paired t-test to check for significant differences. Both the inverse TTC (\( t(1,075)=-7.61 \), \( p=6.1\mathrm{e}{-14}\lt 0.001 \), Cohen d = \( 0.302 \)) and time gap (\( t(1,075)=13.49 \), \( p=2.0\mathrm{e}{-38}\lt 0.001 \), Cohen d =\( 0.234 \)) values at the moment of lane change differ significantly between the model and human demonstrations. So for lane-changing behavior, we conclude that the IRL-based model does not resemble human behavior on an operational level.

Fig. 6.

Fig. 6. Estimated distributions of inverse time to collision and time gap at the moment of the lane change. The orange distributions represent the model’s behavior and the blue distributions represent the human demonstrations. The mean values for inverse TTC are \( 0.19~s^{-1} \) for human lane changes and \( 0.47~s^{-1} \) for the model. The mean values for time gap are \( 1.05~s \) for the human behavior and \( 0.85~s \) for the model behavior.

When comparing the agent’s car-following behavior (Figure 5(d)) with the human’s car-following behavior (Figure 5(c)), there are no oscillations around an equilibrium point for most agents. The general shape resembles that of a human lane-changing maneuver (Figure 5(a)) without crossing the center lane-marking. From this, we conclude that if the model shows car-following behavior, it does not do that in a way that resembles human oscillatory car-following behavior, but instead it tailgates the preceding vehicle.

Reason for the Agents’ Behavior.

Why do the IRL-based agents show behavior that is so different from human behavior, even though their reward function was learned from human demonstrations? We randomly selected several agent trajectories for manual examination using the TraViA [39] traffic visualization tool to answer this question. Examples of these trajectories can also be found as videos in the supplementary materials. From these manual evaluations, two main causes were identified that explain why the behavior of the agents does not represent human behavior: the model’s assumptions and the IRL fitting procedure.

To start with the cases where the model’s assumptions cannot explain the desired behavior. Consider a demonstration where a human is merging in a slow-moving and crowded left lane to overtake a truck farther ahead in the right lane. This might be beneficial in the long run because the truck can be overtaken, but such behavior is unlikely to be beneficial within the short planning horizon of the model, especially because the distance-based collision features promote staying away from other vehicles. This issue is similar to the previously identified problem that lane changes to a right lane cannot be explained by the currently assumed reward function. In both of these cases, no matter the learned weights, the assumed reward function will not lead to the desired behavior within the planning horizon.

In other cases, the approach of learning the weights from a demonstration using an assumed reward function can be identified as the cause of the problem. Many agents that collided learned their weights from a demonstration where the human moves into the area influenced by the collision feature (see Figure 7 for an example). Because the dimensions of this collision feature are fixed and only the weights are learned in the IRL procedure, the resulting collision weight will be low, i.e., a low collision weight is the only way to explain the human moving into this area. When this low collision weight is used in the agent to generate behavior, the agent will not perform a lane change, because moving into the collision-feature area will always decrease the reward. Instead, the agent will stay in its lane. When it approaches the preceding vehicle, it will collide, because of the low collision-prevention weight.

Fig. 7.

Fig. 7. An example of a demonstration where the assumed anti-collision feature does not describe the human’s demonstrated behavior. In this figure, the black shape represents the position of the human-driven demonstration vehicle during a lane change maneuver, the white arrow indicates its direction of motion. Only the collision feature is visualized with warmer colors indicating a higher cost. In this example, the demonstrating vehicle is moving from a low-cost area (right lane) to a high-cost area (left lane). The only way to explain this behavior with the assumed features is to assign a low weight to the collision feature. Apparently, the demonstrating human does not care so much about moving into the higher cost area in the left lane, other features must be more important. When these learned weights are then used in a utility-maximizing agent in the same situation, it will not make the demonstrated lane change. Instead, it will stay in the right lane with a lower penalty and finally collide with the preceding vehicle (969), because collision prevention has a low weight.

The underlying problem here is that the assumed reward function cannot describe the human’s demonstrated behavior properly. Suppose using such a flawed reward function with hand-picked weights. In that case, one would expect prediction errors on the operational level, because the timing of the lane change is determined by the distance-based collision feature. In this case, however, the IRL procedure exaggerates the effects of the flawed reward function by learning weights that result in more collisions. So even though the problem lies in the flawed reward function and not the IRL procedure itself, the combination of the IRL-procedure and rewards function might not only limit the performance of the model, it can actively make it worse.

Skip 5DISCUSSION Section

5 DISCUSSION

In this work, we have proposed a validation workflow for driver models in interaction-aware AV controllers. We illustrated its utility through a case study of validating an inverse reinforcement learning-based driver model replicated from the literature [34] using naturalistic highway driving data extracted from the HighD dataset [17]. Our validation workflow (Figure 2) incorporates the automatic extraction of comparable lane-change scenarios (5,581) on which the IRL model was trained (step 1). The validation of the model was then performed in two related stages. First, we examined the tactical behavior of the model (step 2). Even though no collisions or off-road driving were present in the training data, the model produced such behavior in more than \( 41\% \) of the cases (Table 1). Second, we analyzed the operational behavior of the model in the \( 59\% \) remaining trajectories (step 3). This analysis revealed that even though the dynamics of the model’s lane changes are similar to humans (Figure 5(a) and (b)), the model performed the lane changes with significantly smaller safety margins (Figure 6). Furthermore, the dynamics of the model’s car following behavior were largely inconsistent with human behavior (Figure 5(c) and (d)).

In conclusion, despite training the IRL-based model on data of real-world driving behavior, our 3-step evaluation workflow exposed how the model is not able to produce realistic behavior in the same scenarios. This case study illustrated that despite promising results in simple IAC demonstrations, the models used for human behavior prediction in IACs can deviate substantially from actual human behavior, which can have serious ramifications for generalization of IACs to real-world environments. Our results highlight the importance of validating the models used in interaction-aware motion planning for autonomous vehicles and suggest an easy-to-use framework to aid researchers in doing so.

Practical Applicability of the Validation Workflow.

The case study of validating an IRL-based driver model illustrated the practical applicability of the proposed validation workflow (Figure 2). In the first step of the workflow, the case study showed the feasibility of automatically extracting data from an open-access naturalistic dataset. Even after narrowing down the extracted data to select specific scenarios (in our case, lane changes), the data were sufficiently rich to serve as training data for the IRL model. Note that multiple other datasets were available for consideration (e.g., NGSim [44] and PNeuma [3]) to further enlarge the data and/or use scenarios other than lane changes.

The second and third steps of the workflow propose a two-stage evaluation approach, separated into tactical and operational driver behavior. The case study illustrated why this two-stage validation is useful and necessary. On the tactical level, the large number of collisions and off-road driving would have been hard to identify in a one-stage metric-based validation (e.g., mean square-root error in [37]). On the operational level, the evaluation illustrated that the differences between car following and lane changing in human behavior were not reflected in the model’s behavior. This would have been impossible to identify without first examining the tactical behavior.

The results of the case study also underline the importance of validating driver models for IACs in general. The discrepancy between the driver model and human behavior suggests that an IAC using this model might not safely generalize to real-world scenarios. The case study shows that models that do not actually capture human behavior are not just a hypothetical issue, but a practical concern for IACs developed for autonomous vehicles.

Implications for Interaction-aware Controllers.

The results of the IRL-based model validation have implications for IACs that would use this model to predict other drivers’ responses. Wrong predictions on the tactical level can lead to dangerous situations. If an AV decides to accelerate based on an inaccurate prediction that a vehicle in an adjacent lane will stay there, a dangerous situation might occur when the other vehicle moves in front of the AV. The same holds for inaccurate predictions on an operational level. For example, the model will close the gaps to a (very) high inverse TTC (low TTC) compared to human drivers. This can lead to over-conservative AV behavior because the controller over-estimates the aggressiveness of the human. The full extent of these implications needs to be further examined in future work.

Related Work and Generalizability.

To the best of our knowledge, our work is the first attempt to validate a driver model used in interaction-aware controllers on both the tactical and operational levels. The work on which this model was based [34] does not use naturalistic data and reports no validation attempt of the behavior model. Another related study [36] does use naturalistic data (NGSim) to train the IRL-based model, but also does not report any validation of the trained model. In the supplementary material (available at [37]) Schwarting et al. do report the mean squared errors of their model for merging scenarios. However, given the complexity of human behavior in traffic interactions, such one-dimensional averaged error-metrics provide only rudimentary information on how well the model captures human behavior.

Driver model validation using naturalistic data has been performed for other use cases than IACs. In [48] five car-following models are validated for use in microscopic traffic simulations on naturalistic data collected in Shanghai. Their validation method could also be useful when designing IACs, and our validation method could as well be used to validate models developed for applications other than IACs. However, we argue that because our method includes the tactical and operational validation steps, it is more suitable to validate models displaying multiple higher-level behaviors.

Besides the IAC literature, there have been other driver modeling attempts using IRL. However, IRL-based models can differ substantially from each other in terms of the used reward function. Naumann et al. studied the suitability of different cost functions for different driving scenarios [27] and showed that there are substantial differences. The other modeling attempts that use IRL differ from our work precisely in the sense that they target another scenario (e.g., [32] who regards curve negotiation) or use different reward function features (e.g., [12] who use velocity-based features for risk perception). That means that the results of those works should be regarded as validations of different models, despite the fact that they are also based on IRL. This observation leads to two conclusions. First, other models that use different features should be validated as new models even when they also use IRL. Second, the choice of features for the reward function impacts the performance of the model, which might provide an opportunity to improve models that underperform.

The reasons why the IRL-based models perform poorly in our case study will most likely generalize to other IRL-based models that use similar distance-based collision features in the reward function. The results show that such distance-based features do not capture the essence of human driving behavior. Only changing the shape or dimension of a position-based feature will not solve this. Instead, we advocate that the metrics used in the reward function features should be based on human factors literature for the targeted tactical behaviors, as was done in the operational validation of the model; e.g., the distance-based collision feature could be replaced with a TTC-based feature (similar to the previously mentioned model in [12]).

Other validation attempts of human driver models that do not specifically target IACs and do not use IRL also exist, one especially related to our work is [40]. In that work, Srinivasan et al. compare the trajectories generated with a deep-learning-based model to naturalistic driving data. As in our work, the comparison is based on an in-depth analysis of the resulting trajectories instead of one-dimensional metrics. They show that, also for deep-learning-based driver models, validation should be grounded on a low-level comparison of trajectories, not just high-level metrics. They do however not provide a generalized framework for performing such validations as we do with our proposed workflow.

Limitations and Recommendations.

This work has three main limitations. First, we used only a single demonstration of a lane change to train the IRL, which might explain part of the discrepancy between human and model behavior in the results. However, providing the system with more training data might only slightly improve the model’s performance. In the case study, we identified the causes of the observed problems to be the features of the reward function, not the weights. Adding more training data could result in weights that better fit a specific driver. But it will not negate the problem with the features used in the reward function.

Second, it should be noted that the planning horizon of the model is very short due to the combination of a low number of frames and a high frame rate (\( N=5 \) at 25 Hz). The number of frames within the planning horizon was chosen based on the previous work [34] and to keep the IRL procedure tractable. The frame rate was directly adopted from the HighD dataset for simplicity, both for reproduction purposes and to not introduce any extra assumptions when down-sampling the data. Increasing the planning horizon and examining the model’s behavior under those conditions is left for future work.

Finally, our case study only attempts to validate the model, it does not quantify the implications of the outcome for use of the model in an IAC. Therefore, we are unable to say which aspects of the model’s behavior would be tolerable when used in an IAC or which aspects have major consequences. Quantifying the implications of the mismatch between the model’s, and naturalistic human behavior is left for future work. Answering such a question is an interesting topic of research on its own, a perspective on how to approach such an evaluation can be found in [22].

Future work should also focus on validating more driver models for use in AV controllers, e.g., the Intelligent Driver Model [43] mentioned in the introduction is used in many simulations and demonstrations to model individual human behavior for IACs and should be validated for such use. Future work on IRL-based driver models could focus on redesigning the used reward function such that it better captures similarities between human drivers by using human-factors literature as a starting point. Besides that, the IRL-based model used here could be extended to take the uncertainty in human behavior into account. Either the uncertainty over the learned rewards could be targeted by learning multiple reward functions (as is done in [27, 41]) instead of only single parameters and selecting the best fit, or stochasticity could be added when selecting the actions to relax the assumption of humans being utility maximizers (also done in [41]). However, such changes to the model could complicate the implementation in an IAC.

Skip 6CONCLUSIONS Section

6 CONCLUSIONS

In this article, we argued for validation of the driver models used in interaction-aware controllers. We proposed an evaluation workflow for such validation, illustrated through a concrete case study. Based on the findings in our article, we conclude the following:

The proposed workflow allowed for a detailed evaluation of a driver model replicated from literature, based on an open-source dataset from which 3,279 human-driven lane changes in moderately heavy highway traffic could be extracted. After training the model on each lane change, it did not reproduce adequate behavior when exposed to the same conditions. It generated crashes and road departures in 41.7% of the cases (inadequate tactical behavior). For the remaining cases, unrealistic safety margins were observed (inadequate operational behavior). These unrealistic predictions show that models that do not capture realistic human behavior are a practical concern for implementing IACs in future autonomous vehicles.

During the case study, the proposed workflow proved to be practically applicable, providing a structured basis for model validation in two stages:

First, validating the tactical behavior illustrated to what extent high-level choices are correctly predicted (e.g., that a lane change occurs, rather than staying behind the lead vehicle see Table 1).

Second, correct tactical behaviors produced by a model should be validated in additional detail, by evaluating to what extent the behavior is executed in a way that resembles the timing and spatio-temporal safety margins acceptable to human drivers (see Figure 5).

In these two stages, different tactical behaviors should be evaluated based on different operational criteria because differences in human operational behavior were observed for different tactical behaviors (see Figure 5).

REFERENCES

  1. [1] Abbeel Pieter and Ng Andrew Y.. 2004. Apprenticeship learning via inverse reinforcement learning. Proceedings, 21st International Conference on Machine Learning. 18. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. [2] Antin Jonathan F., Lee Suzie, Perez Miguel A., Dingus Thomas A., Hankey Jonathan M., and Brach Ann. 2019. Second strategic highway research program naturalistic driving study methods. Safety Science 119 (2019), 210. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  3. [3] Barmpounakis Emmanouil and Geroliminis Nikolas. 2020. On the new era of urban traffic monitoring with massive drone data: The pNEUMA large-scale field experiment. Transportation Research Part C: Emerging Technologies 111, November 2019 (2020), 5071. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  4. [4] Clements Lewis M. and Kockelman Kara M.. 2017. Economic effects of automated vehicles. Transportation Research Record: Journal of the Transportation Research Board 2606, 1 (2017), 106114. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  5. [5] Coskun Serdar, Zhang Qingyu, and Langari Reza. 2019. Receding horizon markov game autonomous driving strategy. In Proceedings of the 2019 American Control Conference. IEEE, 13671374. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  6. [6] Evestedt Niclas, Ward Erik, Folkesson John, and Axehill Daniel. 2016. Interaction aware trajectory planning for merge scenarios in congested traffic situations. In Proceedings of the 2016 IEEE 19th International Conference on Intelligent Transportation Systems.465472. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. [7] Garzón Mario and Spalanzani Anne. 2019. Game theoretic decision making for autonomous vehicles’ merge manoeuvre in high traffic scenarios. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference34483453. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. [8] Greenberg Jeffry and Blommer Mike. 2011. Physical fidelity of driving simulators. In Proceedings of the Handbook of Driving Simulation for Engineering, Medicine, and Psychology. Number 2011. CRC Press, 7–1–7–24. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  9. [9] Hang Peng, Lv Chen, Xing Yang, Huang Chao, and Hu Zhongxu. 2021. Human-like decision making for autonomous driving: A noncooperative game theoretic approach. IEEE Transactions on Intelligent Transportation Systems 22, 4 (2021), 20762087. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  10. [10] Harper Corey D., Hendrickson Chris T., Mangones Sonia, and Samaras Constantine. 2016. Estimating potential increases in travel with autonomous vehicles for the non-driving, elderly and people with travel-restrictive medical conditions. Transportation Research Part C: Emerging Technologies 72 (2016), 19. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  11. [11] Hoogendoorn Serge, Ossen Saskia, and Schreuder M.. 2006. Empirics of multianticipative car-following behavior. Transportation Research Record: Journal of the Transportation Research Board 1965, 1965 (2006), 112120. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  12. [12] Huang Zhiyu, Wu Jingda, and Lv Chen. 2021. Driving behavior modeling using naturalistic human driving data with inverse reinforcement learning. IEEE Transactions on Intelligent Transportation Systems (2021), 113. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  13. [13] Hubmann Constantin, Schulz Jens, Xu Gavin, Althoff Daniel, and Stiller Christoph. 2018. A belief state planner for interactive merge maneuvers in congested traffic. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems. IEEE, 16171624. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. [14] Isele David. 2019. Interactive decision making for autonomous vehicles in dense traffic. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference. IEEE, 39813986. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. [15] Jiang Rui, Hu Mao Bin, Zhang H. M., Gao Zi You, Jia Bin, and Wu Qing Song. 2015. On some experimental features of car-following behavior and how to model them. Transportation Research Part B: Methodological 80 (2015), 338354. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  16. [16] Kondoh Takayuki, Yamamura Tomohiro, Kitazaki Satoshi, Kuge Nobuyuki, and Boer Erwin Roeland. 2008. Identification of visual cues and quantification of drivers’ perception of proximity risk to the lead vehicle in car-following situations. Journal of Mechanical Systems for Transportation and Logistics 1, 2 (2008), 170180. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  17. [17] Krajewski Robert, Bock Julian, Kloeker Laurent, and Eckstein Lutz. 2018. The highD dataset: A drone dataset of naturalistic vehicle trajectories on german highways for validation of highly automated driving systems. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems. IEEE, 21182125. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. [18] Lee David N.. 1976. A theory of visual control of braking based on information about time-to-collision. Perception 5, 4 (1976), 437459. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  19. [19] Lenz David, Kessler Tobias, and Knoll Alois. 2016. Tactical cooperative planning for autonomous highway driving using Monte-Carlo tree search. In Proceedings of the 2016 IEEE Intelligent Vehicles Symposium (IV). IEEE, 447453. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. [20] Levine Sergey and Koltun Vladlen. 2012. Continuous inverse optimal control with locally optimal examples. Proceedings of the 29th International Conference on Machine Learning 1 (2012), 4148.Google ScholarGoogle Scholar
  21. [21] Liu Wei, Kim Seong-Woo, Pendleton Scott, and Ang Marcelo H.. 2015. Situation-aware decision making for autonomous driving on urban road using online POMDP. In Proceedings of the 2015 IEEE Intelligent Vehicles Symposium (IV). 11261133. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  22. [22] Markkula Gustav and Dogar Mehmet. 2022. How accurate models of human behavior are needed for human-robot interaction? For automated driving?IEEE Robotics and Automation Magazine.Google ScholarGoogle Scholar
  23. [23] Meng Fanlin, Su Jinya, Liu Cunjia, and Chen Wen-Hua. 2016. Dynamic decision making in lane change: Game theory with receding horizon. In Proceedings of the 2016 UKACC 11th International Conference on Control. IEEE, 16. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  24. [24] Michon John A.. 1985. A critical view of driver behavior models: What do we know, what should we do? In Proceedings of the Human Behavior and Traffic Safety. Springer, Boston, MA, 485524. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  25. [25] Mulder Mark, Mulder Max, Paassen M. M. Van, and Abbink D. A.. 2005. Effects of lead vehicle speed and separation distance on driver car-following behavior. IEEE International Conference on Systems, Man and Cybernetics 1, 4 (2005), 399404. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  26. [26] Mulder Mark, Paassen Marinus M. Van, Mulder Max, Pauwelussen Jasper J. A., and Abbink David A.. 2009. Haptic car-following support with deceleration control. In Proceedings of the 2009 IEEE International Conference on Systems, Man and Cybernetics, 16861691. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  27. [27] Naumann Maximilian, Sun Liting, Zhan Wei, and Tomizuka Masayoshi. 2020. Analyzing the suitability of cost functions for explaining and imitating human driving behavior based on inverse reinforcement learning. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation. IEEE, 54815487. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  28. [28] Ng Andrew and Russell Stuart. 2000. Algorithms for inverse reinforcement learning. Proceedings of the 17th International Conference on Machine Learning 0 (2000), 663670. Retrieved from http://www-cs.stanford.edu/people/ang/papers/icml00-irl.pdf.Google ScholarGoogle Scholar
  29. [29] Ossen Saskia and Hoogendoorn Serge P.. 2011. Heterogeneity in car-following behavior: Theory and empirics. Transportation Research Part C: Emerging Technologies 19, 2 (2011), 182195. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  30. [30] Pettigrew Simone. 2017. Why public health should embrace the autonomous car. Australian and New Zealand Journal of Public Health 41, 1 (2017), 57. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  31. [31] Ranney Thomas A.. 2011. Psychological fidelity: Perception of risk. In Proceedings of the Handbook of Driving Simulation for Engineering, Medicine, and Psychology, Fisher D. L., Rizzo M., Caird J., and Lee J. D. (Eds.), Chapter 9, 9–1 –9–13.Google ScholarGoogle ScholarCross RefCross Ref
  32. [32] Rosbach Sascha, James Vinit, Grosjohann Simon, Homoceanu Silviu, and Roth Stefan. 2019. Driving with style: Inverse reinforcement learning in general-purpose planning for automated driving. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 26582665. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. [33] Dang Ruina, Zhang Fang, Wang Jianqiang, Yi Shichun, and Li Keqiang. 2013. Analysis of chinese driver’s lane change characteristic based on real vehicle tests in highway. In Proceedings of the 16th International IEEE Conference on Intelligent Transportation Systems. IEEE, 19171922. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  34. [34] Sadigh Dorsa, Landolfi Nick, Sastry Shankar S., Seshia Sanjit A., and Dragan Anca D.. 2018. Planning for cars that coordinate with people: Leveraging effects on human actions for planning and active information gathering over human internal state. Autonomous Robots 42, 7 (2018), 14051426. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  35. [35] Saifuzzaman Mohammad and Zheng Zuduo. 2014. Incorporating human-factors in car-following models: A review of recent developments and research needs. Transportation Research Part C: Emerging Technologies 48 (2014), 379403. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  36. [36] Schwarting Wilko, Pierson Alyssa, Alonso-Mora Javier, Karaman Sertac, and Rus Daniela. 2019. Social behavior for autonomous vehicles. Proceedings of the National Academy of Sciences 116, 50 (2019), 2497224978. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  37. [37] Schwarting Wilko, Pierson Alyssa, Alonso-Mora Javier, Karaman Sertac, and Rus Daniela. 2019. Social behavior for autonomous vehicles - Supporting Information. Retrieved 19 July, 2021 from https://www.pnas.org/content/suppl/2019/11/22/1820676116.DCSupplemental.Google ScholarGoogle Scholar
  38. [38] Siebinga O.. 2021. IRL Model Validation - TraViA extension code. Retrieved 21 March, 2022 from https://github.com/tud-hri/irlmodelvalidation.Google ScholarGoogle Scholar
  39. [39] Siebinga Olger. 2021. TraViA: A traffic data visualization and annotation tool in python. Journal of Open Source Software 6, 65 (2021), 3607. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  40. [40] Srinivasan Aravinda Ramakrishnan, Hasan Mohamed, Lin Yi-Shin, Leonetti Matteo, Billington Jac, Romano Richard, and Markkula Gustav. 2021. Comparing merging behaviors observed in naturalistic data with behaviors generated by a machine learned model. In IEEE International Intelligent Transportation Systems Conference (ITSC). 3787–3792. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  41. [41] Sun Liting, Wu Zheng, Ma Hengbo, and Tomizuka Masayoshi. 2020. Expressing diverse human driving behavior with probabilistic rewards and online inference. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 20202026. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  42. [42] Tian Ran, Li Sisi, Li Nan, Kolmanovsky Ilya, Girard Anouck, and Yildiz Yildiray. 2018. Adaptive game-theoretic decision making for autonomous vehicle control at roundabouts. In Proceedings of the 2018 IEEE Conference on Decision and Control. IEEE, 321326. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  43. [43] Treiber Martin, Hennecke Ansgar, and Helbing Dirk. 2000. Congested traffic states in empirical observations and microscopic simulations. Physical Review E - Statistical Physics, Plasmas, Fluids, and Related Interdisciplinary Topics 62, 2 (2000), 18051824. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  44. [44] Administration U.S. Department of Transportation Federal Highway. 2016. Next Generation Simulation (NGSIM) Vehicle Trajectories and Supporting Data. [Dataset]. Retrieved 19 July, 2021 from https://data.transportation.gov/Automobiles/Next-Generation-Simulation-NGSIM-Vehicle-Trajector/8ect-6jqj.Google ScholarGoogle Scholar
  45. [45] Ward Erik, Evestedt Niclas, Axehill Daniel, and Folkesson John. 2017. Probabilistic model for interaction aware planning in merge scenarios. IEEE Transactions on Intelligent Vehicles 2, 2 (2017), 11. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  46. [46] Yu Hongtao, Tseng H. Eric, and Langari Reza. 2018. A human-like game theory-based controller for automatic lane changing. Transportation Research Part C: Emerging Technologies 88, October 2017 (2018), 140158. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  47. [47] Zhang Qingyu, Filev Dimitar, Tseng H. E., Szwabowski Steve, and Langari Reza. 2018. Addressing mandatory lane change problem with game theoretic model predictive control and fuzzy markov chain. In Proceedings of the 2018 Annual American Control Conference. IEEE, 47644771. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  48. [48] Zhu Meixin, Wang Xuesong, Tarko Andrew, and Fang Shou’en. 2018. Modeling car-following behavior on urban expressways in Shanghai: A naturalistic driving study. Transportation Research Part C: Emerging Technologies 93, September 2017 (2018), 425445. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  49. [49] Ziebart Brian D., Maas Andrew, Bagnell J. Andrew, and Dey Anind K.. 2008. Maximum entropy inverse reinforcement learning. In Proceedings of the National Conference on Artificial Intelligence. 14331438.Google ScholarGoogle Scholar

Index Terms

  1. A Human Factors Approach to Validating Driver Models for Interaction-aware Automated Vehicles

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      HTML Format

      View this article in HTML Format .

      View HTML Format