Start
1. Introduction
1.1. Research question & approach
2. Background
2.1. Different approaches
2.2. Social driving behaviour
2.3. Primer on AV design
3. Method
3.1. Participants
3.2. Materials
3.3. Procedure
3.4. Data logging & analysis
4. Results
4.1. Defining and framing social driving behaviour
4.2. Process phase I: Requirements
4.3. Process phase II: Development
4.4. Process phase III: Evaluation
5. Discussion
5.1. Main takeaways
5.2. Limitations and recommendations
6. Conclusion
Appendix
Rederences

Come together: an exploration on social driving behaviour of automated vehicles

Abstract

Interaction between road users is a fundamental part of the traffic system. The advent of automated vehicles (AVs) has given rise to requirements for interactions between AVs and other road users, expressed in high-level terms like ‘demonstrate anticipatory behaviour’, ‘not confusing other road users’, and ‘being predictable and manageable for other road users’. Operationalizing these social driving behaviours requires social science knowledge on human interaction. However, translating social driving behaviour requirements unambiguously to the engineering domain necessitates that social scientists have a rudimentary understanding of the language of engineers (and vice versa). The present study seeks to accommodate interdisciplinary collaboration between social scientists and engineers by providing insight into current AV technological capabilities with regards to social driving behaviour and road safety, and their development in the near future. To this end, an exploratory interview study was performed with 7 engineers with backgrounds in industry, academia, research institutes, and/or vehicle authorities. The engineers provided several real-world examples of implications of AV algorithms on social driving behaviour. Thematic analysis of the interview transcripts resulted in clusters of themes relating to the product development process: requirements (i.e., societal, legal, customers), development (i.e., process, implementation), and evaluation (i.e., assessment, monitoring). Choices made in each of these phases appear to influence the final behaviour of automated vehicles in traffic. Knowledge on social driving behaviour and its impact on traffic safety can guide these choices to ensure safe operation of AVs within the social environment of traffic.

1. Introduction

The advent of automated vehicles (SAE level 3+) is predicted to change transportation on a global scale (Elvik, 2020; Hoogendoorn et al., 2014). The rapid improvements made to (semi-)automated vehicles and towards fully automated vehicles (AVs) are made possible due to technological and engineering solutions. These solutions could enable vehicles to travel from A to B autonomously. However, being a part of the traffic system requires more than the ability to control a vehicle; it also requires interaction with other road users, including vulnerable road users (VRUs) such as bicyclists and pedestrians, and human driven vehicles (Grahn et al., 2020; Markkula et al., 2020; Straub & Schaefer, 2019). These interactions between road users are based on formal regulations such as ‘yielding to traffic coming from the right’ (the rules of the road), but also on informal regulations based on societal norms and values, such as drivers voluntarily giving way to VRUs beyond their obligations. Drivers also use subtle cues to make clear what their intentions are to other road users such as timely and gradually slowing down when nearing a crosswalk to indicate to pedestrians they can cross the road safely. This social aspect of driving is a crucial element that is not always present in current automated driving solutions, whether in production vehicles or proposed in the literature (Brown & Laurier, 2017; Negash & Yang, 2023; Villadsen et al., 2023). The importance of the human and social aspects of driving for automated vehicles are being acknowledged by legislative bodies. For example, regulations on automated lane keeping systems (UNECE 2021) state that “the system shall aim to keep the vehicle in a stable lateral position inside the lane of travel to avoid confusing other road users” and regulations on advanced driver assistance steering systems (UNECE 2025) state that “A lane change procedure shall be predictable and manageable for other road users.”. However, which variations in lane position would actually confuse other road users and what constitutes a predictable and manageable lane change is still under discussion. The further development and specification of these requirements necessitates insight in human behaviours. Such insight is often the domain of the so-called ‘social sciences’ (i.e., psychology, sociology, linguistics, and other disciplines with a focus on human cognition, interaction and communication), but it is not always readily available or usable to other domains. Collaboration is therefore needed between social scientists and engineers to develop AVs that integrate seamlessly into a traffic system where social dynamics are crucial for managing and preventing conflicts. Yet, this collaboration comes with challenges (Jones & Jones, 2016; Lowe et al., 2013), and differences between the two fields create a so-called ‘epistemological gap’ both in terms of knowledge and approaches Hadfield-Hill et al. (2020). The aim of this study is to reduce this gap, by identifying the challenges engineers face when developing social aspects of AV driving behaviour and highlighting how social scientists can contribute.

1.1. Research question & approach

This study addresses the following research question: How can social scientists use their knowledge to improve social driving behaviour of automated vehicles?

Understanding the hurdles to implementing social behaviour in AVs is essential for integrating social science knowledge into the development of social driving behaviour. Therefore seven engineers working in the field of AV perception and decision making were interviewed about their insights in the development of algorithms and requirements for AVs regarding social driving behaviour. To convey these insights effectively to social scientists a strong focus was put on obtaining examples of related real-world behaviours. Such examples help bridge the gap between the obtained engineering insights and the perspective of social scientists, by creating a point of congruence (Hadfield-Hill et al., 2020). Furthermore the focus of this exploratory study was on getting a wide range of engineering insights, rather than interviewing a homogeneous group of participants. We first present the identified themes that arose during the interviews, each accompanied with the previously mentioned related real-world examples. We then use these results to explore how social scientists can apply their expertise to improve social driving behaviour of AVs.

2. Background

The perceived relevance of social sciences for automative engineering has been increasing, especially in research focusing on the vehicle interior. Multiple standards, e.g., ISO 9241-210:2019 (ISO, 2019), were developed to provide requirements and recommendations for human centred HMI designs. More recently there has been a growing interest in the interactions between road users, especially when AVs are part of the interaction (e.g., Brown & Laurier, 2017; Knoop et al., 2019; Madigan et al., 2019).

Previous work (e.g., Grahn et al., 2020) investigated how human drivers enable safe driving within the social environment on the road. Transferring these skills from human drivers to automated driving results in AVs that are more seamlessly integrated into the social traffic system, likely improving safety and efficiency as a result (Straub & Schaefer, 2019).

2.1. Different approaches

The approach to investigating the interactions between human drivers and AVs appears to differ substantially between engineers and social scientists. An engineering perspective requires well-defined descriptions of the behaviour that needs to be performed, and when to perform this behaviour or when not to. Engineering studies often aim to provide an approach to solving the described interaction, often in the form of an algorithm or set approach. Defining human behaviour into a more precise, often mathematical description results into what is also known as a behavioural specification (Bin-Nun et al., 2022). These behavioural specifications define intended behaviour in a way that AVs can implement, similar to how traffic laws guide human drivers. Traffic laws, written from a social science perspective, allow for driver judgment and adaptation based on the circumstances, often using broad terms like “when safe” or “when appropriate.” In contrast, behavioural specifications follow a precise, mathematical approach, reflecting the differing perspectives of social scientists and engineers.

A clear example of differing perspectives is an AV approaching a pedestrian crossing. Engineers require precise braking parameters to achieve socially desirable interactions (e.g., Lee et al., 2017). Social science research on pedestrian crossings offers valuable insights (e.g., Beggiato et al., 2018; Razmi Rad et al., 2020; Tian et al., 2023), but lacks the detail needed for direct implementation, creating a gap between knowledge and engineering application.

2.2. Social driving behaviour

In a broad context, Schmitt (1998) defines social behaviour as: “a person’s behaviour is social when its causes or effects include the behaviour of others”. It is important to note that this definition is ‘value free’, meaning that social behaviour as defined by Schmitt could manifest as pro-social and anti-social behaviour. An example of pro-social behaviour in a driving context is the granting of right-of-way for another driver even if not required to do so by law, while an anti-social example could be cutting in front of another driver. Some aspects of social behaviour can be considered as intended behaviours while there are also behaviours that influence others without the intention to do so. An example of intended social behaviour in a driving context is the use of vehicle indicators to communicate a desire to change lanes to other road users. An example of unintended behaviour that influences others could be the braking pattern of a vehicle. Even if the braking pattern is solely chosen to brake at a set position it still communicates information about intent and future position to other road users, whose state, behaviour and goals could be influenced by it. For example, braking early and gradually for a pedestrian crossing also communicates the intent to stop to any waiting pedestrians, while braking late and suddenly could be interpreted by waiting pedestrians as unwillingness to stop or a failure to notice them (Tian et al., 2023). Additionally, not accounting for unexpected braking patterns may result in rear end crashes (Favarò et al., 2017). Therefore, unintended influences of AV behaviour should be taken into account when developing AVs. Other examples of unintended consequences of AV behaviour are slow driving and the adoption of a large following distance causing frustration in other drivers resulting in dangerous manoeuvres (Knoop et al., 2019) and the miscommunication in merging scenarios due to unusually large gaps left by an AV (Brown & Laurier, 2017). Aspects of social behaviour can be proactive or reactive, with proactive behaviours occurring as anticipation to situation that has not manifested yet (e.g., moving to the right-most lane early on the highway in preparation for an upcoming offramp or slowing down further when in a school zone around the time schools end) and reactive behaviours only occurring when a situation presents itself (e.g., braking in response to another vehicle using indicator lights).

Based on the above, a preliminary value-free definition of social driving behaviour was prepared for the interview study that attempts to unite both engineering and social scientist perspectives. We define social driving behaviour as: driving behaviour that directly or indirectly influences and/or takes into account other road users, e.g. their state, behaviour and goals.

2.3. Primer on AV design

A condensed description of AV design follows to facilitate a basic understanding of the engineering perspective on social driving behaviour.

According to Pendleton et al. (2017), AVs interact with the driving environment through hardware, involving sensors (e.g., radar), actuators (e.g., a steering system), as well as communication with other vehicles or infrastructure (e.g., vehicle-to-Everything, V2X), see Figure 1. Within the software realm, Pendleton et al. (2017) discerns three subsystems: perception, planning, and control. Algorithms in the perception subsystem make use of sensor and V2X data to perceive and recognize objects in the environment as well as the ego vehicle’s state (e.g. the own speed). For example, a radar sensor may report distances to objects, as well as changes in these distances, while the algorithms translate this data into specific other road users or objects and their states and characteristics. One step further, these algorithms can also predict possible future states of surrounding objects. Consequently, a model of the vehicle position in relation to its environment is created. Next, the planning subsystem uses this information and plans target actions for the vehicle. The planning subsystem utilizes a number of factors to determine what target actions to take. This is done by a so called ‘cost function’: a mathematical formula that combines a number of different factors with weights to provide a ‘cost’ for each option in a set of candidate target actions. The higher the ‘cost’, the less desirable that specific candidate action becomes. Factors could relate to for example travel time; comfort of the driver; fuel efficiency; safety. By optimizing for the lowest cost (or highest benefit), these factors and their weights help a vehicle determine what decision to make and what to avoid. Different types of ‘planners’ are described by Pendleton et al. (2017), reminiscent of Michon’s (1985) hierarchical levels of driving: the mission planner determines, e.g., that the vehicle drives from A to B, whereas the behavioural planner determines, e.g., when to perform an overtaking manoeuvre, and the motion planner determines the exact desired motion path, including positions and velocities at each time step. Finally, execution of the motion path takes place in the control subsystem. The implementation of these functions may differ between vehicles and manufacturers.

Figure 1
Figure 1 Schematic overview of a typical autonomous vehicles systems

V2X = Vehicle-to-Everything. Figure adapted from Pendleton et al. (2017).

3. Method

A semi-structured interview study was conducted to gain insight into current AV technological capabilities with regards to social driving behaviour and road safety, and their development in the near future. A thematic analysis (Braun & Clarke, 2006) was performed on the interview results.

3.1. Participants

Prospective participants were identified through web search (and follow-up inquiries within the resulting companies and institutes) and connections within the network of the authors. They were subsequently selected based on their experience with the development of AVs (e.g., based on published papers). A total of 12 prospective participants were contacted directly through an e-mail, which detailed the purpose of the study and the request to participate in an interview. Seven of them (all males, age range 35-49 years, M = 41.8, SD = 4.8) replied positively and agreed to participate anonymously, covering backgrounds ranging from academia and research institutes to automotive industry and vehicle authorities (see Table 1). At the time of the interview, participants were working across three European countries (Germany, Netherlands, Sweden), with backgrounds within and outside Europe. Experience across the participants included object detection, object tracking, object path prediction, motion planning, decision making, safe driving policies, safety acceptance criteria, modelling human driving behaviour, and cooperation between AVs and infrastructure, pedestrians, and conventional vehicles. Ethical approval for this study was obtained from the SWOV ethics committee.

Table 1 Overview of participant (Pp) roles and affiliations
Pp Role Aca Ind Res Pp
1 Employee automotive supplier of AV software X 1
2 Assistant professor computer vision and perception (previously: robotaxi company) X (X) 2
3 Associate professor computer vision, perception and prediction (previously: automotive company) X (X) 3
4 Leader of the group for cooperative system automation and integration X 4
5 Chief inspector 5
6 Research scientist and manager AV behaviour and motion planning (X) X 6
7 Assistant professor / Senior research scientist AV safe and social interactions X X 7

Aca = Academia (e.g., university), Ind = Automotive industry, Res = Research institute, Aut = Vehicle authority. X-marks within parentheses correspond with previous affiliations.

3.2. Materials

The study consisted of two parts: an online survey and a semi-structured interview. The survey gathered background information (e.g., age, profession, experience with AV development) on participants and sensitized them to the topic of social driving behaviour for a more fruitful interview. Three scenarios were prepared for the survey: a right-angle conflict with a bicycle, merging on a highway, and approaching a school (see Appendix A). Each illustrated scenario consisted of three consecutive steps in time, zooming in from the overall traffic context to specific interactions between an ego vehicle and other road users. Each step was accompanied by a set of three questions asking which driver behaviour is required to avoid a conflict, which cues facilitate such behaviour, and if one’s answer would differ if the ego vehicle would have been an AV. The scenarios fostered reflection on interactions between AVs and other road users before ultimately raising the question on how social driving behaviour should be defined from the perspective of an AV.

A semi-structured interview was prepared to ensure the same topics and terminology were used for all participants. The research question was transformed into interview questions using the guidelines described in (Kumar, 2011). Four main points were addressed (see also Appendix B): 1) social aspects of driving behaviour (e.g., definition of social driving behaviour, example scenarios), 2) the development process (e.g., requirements, algorithms, limitations, updates and improvements), 3) the relation between social driving behaviour and traffic safety, and 4) challenges on realizing potential benefits of social driving behaviour. An interview template was created by interleaving the interview questions with answer fields and pointers for the interviewer to ask for additional details based on the answers provided by the interviewees. The understandability of the survey and interview questions, as well as the general flow of the interview were piloted with a human factors researcher with a focus on AV technology.

3.3. Procedure

A link to the online survey was sent in the week of the interview. To shorten the duration, for each participant two out of three scenarios were randomly selected and included in the survey. All surveys were filled out by the time the interviews were conducted. The interviews were held online using Microsoft Teams in November 2023, with a duration of approximately one hour. Participants were again informed about the purpose of the study, asked if they prefer to be anonymous, and asked how they should be referred to in subsequent reporting. With their consent, the interviews were recorded to facilitate analysis by the authors of the present paper. Next, open-ended semi-structured interview questions were posed to allow the participants to share their views in their own terms. When needed, follow-up questions were asked to ensure participants’ answers were interpreted correctly. When participants seemed to have difficulty providing examples of social driving behaviour, previous answers relating to the selected survey scenarios were used in the interviews to facilitate discourse. Assisted by the interview template, all interviews were conducted by author RJJ, while the authors DC and RdZ made notes. Any additional questions based on these notes were asked by the end of the interview. During interviews, participants frequently provided elaborate answers not only in relation to the original interview question, but also relating to an interview question not yet asked. In such cases the latter question was skipped.

3.4. Data logging & analysis

A thematic analysis was performed in line with the phases described in Braun and Clarke (2006). The main purpose of this analysis was to provide a ‘rich description’ of the data set, in which themes are identified at the ‘semantic level’. A rich description means an accurate reflection of the content of the entire data set where all statements are related to a theme (implying that themes can be based on a subset of the participants), and interpretation at the semantic level means that themes are identified within the explicit or surface meanings of the data (cf. Braun and Clarke (2006)).

Directly after the interviews both note-takers scanned through their notes for relevant statements that were made during the interviews (phase 1: familiarization). In case of unclear notes, the interview recording was consulted. All statements were gathered in a spreadsheet and categorized according to the original interview questions. Initial codes referring to interesting features of each statement were assigned bottom-up (phase 2: generating initial codes). Next, two of the authors individually identified themes by looking at statements with common codes (phase 3: searching for themes).

A joint session focused on the resulting set of themes (phase 4: reviewing the themes). Themes identified by both authors were selected straightaway, whereas themes identified by one author were first discussed with the other author and selected only if the theme did not overlap with previously selected themes. A thematic map was created to identify how themes relate to each other, to ensure there is not too much overlap between themes, and to identify sub-themes. An initial description of each (sub-)theme was formulated (phase 5: defining and naming themes). Finally, statements capturing the essence of a theme were selected, and a description of the theme was prepared (phase 6: producing the report). For each theme, concrete examples were provided through text boxes to make the theme more tangible for social scientists (as recommended by Hadfield-Hill et al. (2020)). Typically, summaries of the actual verbatim were created, in part to reduce the length of the (at times elaborate verbatim), and in part to improve readability, given the fact that examples were frequently interrupted by other examples or trains of thought. A draft version of the results section was presented to the participants, with the question to indicate whether the interpretation of their statements (including aforementioned summaries of the verbatim) corresponded with the original intention of the statements. All participants provided feedback, and the results section was revised accordingly.

4. Results

Statements made by specific participants are hereafter referred to by their participant number (e.g., ‘P1’, see Table 1 for more information on the participants).

4.1. Defining and framing social driving behaviour

In the questionnaire, participants were asked to provide their own definition of social driving behaviour. These definitions generally contained the following terminology: anticipation, social norms, social context, interactions, politeness, communication, social interaction, interpretation of behaviour and common understanding of behaviour in traffic. Most definitions were highlighting the positive aspects of social driving behaviour. At the start of the interview participants were shown our value-free definition of social aspects of driving behaviour: “driving behaviour that directly or indirectly influences and/or takes into account other road users, i.e. their state, behaviour or goals”. Where social driving behaviour can be anywhere on a continuous scale from anti-social to pro-social, i.e., from disrupting to facilitating other’s state, behaviour, and/or goals. All participants understood and agreed to adopt this definition for the remainder of the interview.

The remarks made by participants during the interview related to a wide range of themes throughout the development process, which eventually provided the basis for the thematic map shown in Figure 2. The process consists of three phases: requirements, development and evaluation. Within these phases several themes are represented by blocks. In the following paragraphs each theme is presented in more detail.

Figure 2
Figure 2 Thematic map of processes underlying social driving behaviour: ‘Requirements’, ‘Development’ (with sub-themes relating to ‘process’ and ‘implementation’) and ‘Evaluation’

The thematic map is the authors’ interpretation of themes and relations between themes (cf. thematic analysis in Braun and Clarke (2006)).

4.2. Process phase I: Requirements

Requirements for development follow from stakeholder input across several domains (legal, society, customers), as well as from industry itself.

Legislation was named multiple times as a way to create incentive for the automotive industry to implement or take into account social aspects of driving behaviour (P2, P4, P7). In this regard, also the requirement in the ALKS legislation (UNECE 2021) about minimizing risks to at least the level of a careful and competent human driver was mentioned. The main challenge here is how do you prove that you adhere to this requirement, i.e., how is it technically defined and measured (P5, P6, P7)?

Another requirement that was mentioned was the transparency of the algorithms (P2, P3). This can influence the choice of model approach, i.e., black or white box. Also, legislation regarding exceptional road users, such as funeral processions and blind people, was mentioned as difficult to deal with for AVs (P2).

Legislation was also mentioned as a means to support AVs, such as counteracting bullying of AVs (P2).

Examples of legal requirements or changes suggested:

  • Detection of special road users such as blind pedestrians and funeral processions. The current legal situation does not work. Blind people are allowed to cross anywhere and have priority, this might be unfeasible. Funeral processions might need rules to be easier to detect. (P2)

  • “There are people exploiting the conservative behaviour of AVs in all kinds of ways […] that clearly make AVs impossible. I can imagine in the future you may need some laws for that” (P2)

  • There needs to be a necessity to implement social behaviour, and this might need to be a legal necessity (P4)

4.2.2. Society

The role of society is also diverse. On the one hand, society plays a role in defining requirements, such as deciding what safety target should be met before AVs are allowed on the road (P1, P3), or what is considered a careful and competent driver (P4, P7). Also, these requirements and in general driving behaviour can differ between countries, which can result in differences in requirements for AVs in different countries (P1, P5, P6, P7).

On the other hand, society was also mentioned as a means to label data. Open datasets such as those of Waymo can, for example, be accessed by the public who can aid in labelling it and identifying troublesome events.

Furthermore, society does not always discriminate between AV technologies of different companies when judging these technologies (P6). Consequently, a crash with one AV model of one particular manufacturer may potentially cause a decline in trust in all other AV manufacturers. This makes the whole industry interdependent. Also, as the challenges for this industry are viewed as immense, participant P6 mentioned the need to work together, rather than as competitors. As mentioned before, more open access datasets should be made available (e.g., Department of Motor Vehicles (DMV), n.d.; National Highway Traffic Safety Administration (NHTSA), n.d.; Waymo, n.d.-a; Zenseact, n.d.), but also more data and knowledge sharing between industries and research institutes should occur.

Examples of societal remarks:

  • Different countries have different safety margins. (P1)

  • Different countries have different alignment of traffic and social rules. In some coun-tries it might be more expected to deviate from traffic rules to adhere to social rules (P6)

4.2.3. Customers

Customers in particular play a significant role in AV development. Often the customer benefits are reflected by choices in the algorithms cost function, such as passenger comfort or travel time (P2, P4). These benefits often weigh stronger than the benefits for other road users, who are unintended users of the technology (P4).

Customers can also aid with data acquisition. Their vehicles can log data during their trips which can be used by the manufacturer to improve their algorithms and detect edge cases (P1). Additionally, participant P1 mentioned work on specialized interfaces that allow customers to provide input on incidents during their trip.

Finally, several participants mentioned that there is a difference between companies (P2, P5, P7). Participant P2 mentioned their different purpose and different customer type. The robotaxi’s provide mobility as a service, while more conventional car manufacturers focus more on a personal vehicle. These customers can have different requirements, leading to different design choices. Another difference between these two manufacturer groups is that the robotaxi’s, such as Cruise and Waymo, originate from software related companies, while manufacturers of personal vehicles, such as BMW/Mercedes/Volvo, have a long-standing history in car manufacturing (P7). Their different backgrounds can also be of influence on the development of their AVs and the priorities they have. Finally, also differences in safety requirements were mentioned (P5), e.g., differences in safety requirements in general, but also during the monitoring stage where definitions of what is deemed an “incident” might differ between companies.

Examples of customer influence:

  • “In the end the customer tells the company, through feedback, what they want. You could even imagine that a [race car, red.] driver has a less pro-social behaviour than a family car driver. […] The product is being designed accordingly, more ag-gressive, more conservative, etc.” (P2) Some companies might prefer higher or lower acceleration limits. Progress towards the destination is the goal, safety a constraint (P2)

4.2.4. Program of requirements

The input from the previously mentioned stakeholders, as well as the industry itself, comes down to several requirements regarding social aspects of driving behaviour. First and foremost, AVs should imitate humans, or at least careful and competent humans (P2, P4, P5, P7). By imitating human driving behaviour, the AV should be as readable and predictable for other road users as any human driven vehicle is (P4). If they are not, the safe driving strategies humans adopt in traffic, such as anticipating events, will be jeopardized (P7).

However, humans should not always be imitated (P7). On one hand, unsafe behaviour of humans should not be imitated. On the other hand, some safe human behaviour might be compensating for their limitations. If the AV does not have such limitations, the compensating safe behaviour might also not be necessary.

The big open question thus still remains what should be imitated and what not. There is a need for a framework or rule book of social rules and norms that can be used to define them, implement them in AVs and be used for the assessment of these vehicles (P4, P6, P7).

Currently, requirements related to safety targets are clearly defined in advance (P1, P2, P3, P4, P5, P6, P7). Such requirements are defined as, for example, the number of collisions per driving hour (P5) or the maximal number of collisions of type X per km (P1). Requirements regarding social aspects of driving behaviour are not clearly defined at the start of the development process (P2, P4, P5, P7), but are an emergent property when trying to reach the safety target (P1, P2, P5) or quickly decided upon during development and are often very reliant on context, e.g., in situation X behaviour Y is expected (P2). They are generally not a goal, but did gain more interest in recent years (P4, P6). Participant P5 mentioned that some automotive companies assume that the safety targets set cannot be reached without implementing social driving behaviour. Other participants mentioned that there are some predefined requirements on interaction, such as interaction behaviour during merging (P1, P2). As requirements regarding social driving behaviour seem to mainly emerge during development, developers potentially play an important role in defining them.

Examples of human behaviour that should not be imitated:

  • Unsafe behaviour such as excessive speeding and extreme tail gaiting.

  • Behaviour that compensates for human limitations, such as changing lanes before a crowded merging section to avoid any stress of dealing with other vehicles merging in front. However, changing lanes in that scenario can also benefit other road users in that is leaves space for them to merge. In that case, the AV should change lanes. (P7)

4.3. Process phase II: Development

The requirements, synthesized from different stakeholder inputs, are generally the start of the development process. Participants mentioned that the global development cycle of the algorithms for automated driving includes setting up a model structure, gathering data, label it, fit model parameters to the data then test the models (P1, P3, P5, P6, P7). Accordingly, themes within the development process involve ‘data acquisition’, ‘metrics & data labelling’, and ‘communication between disciplines’. Additionally, the development process involves a cluster of themes relating to (software) implementation, including differences in ‘approaches’ and challenges relating to ‘generalization’. Following the system description of Figure 1, the perception subsystem is associated with the themes ‘recognition’ and ‘prediction’, and the planning subsystem is associated with the themes ‘decision-making’ and ‘target actions’ (i.e., as output from decision-making). The control subsystem in Figure 1 was not discussed in the interviews.

4.3.1. Data acquisition

The real-world data is gathered with probing vehicles, which do not necessarily have the automated driving functions active, and can just be customer vehicles (P1, P5). Alternatively, tests on the public road with specialized test vehicles and test drivers can be used in certain circumstances (P1, P5, P6). For some processes, such as updating models, specific parts of customer data from for example incidents or crashes can be used (P1, P4, P5). In other cases, all data from driving in a specific region is used to fit model parameters of general driving behaviour in that region (P5).

Example challenges for data acquisition:

  • “Tesla was really good to drive in [area X, red.] and really bad in [area Y, red.]. The reason is obvious: if you live in [area X, red.] you have money to buy a Tesla, so Tesla has many more owners driving in [area X and therefore, red.] lots of data to the map. The amount of Tesla’s driving in [area Y, red.] was much less.” (P7)

  • Data is often gathered per region to extract average behaviour in that region. Howev-er, sometimes it is difficult to acquire enough data on behaviour of certain road users, such as for motorcyclist on the Parisian Périphérique (a limited-access dual-carriageway ring road in Paris). In this case, data acquisition can be extended to re-gions outside of the operational design domain of the AV (the operating conditions under which a given driving automation system is specifically designed to function), such as on larger roadways inside the Périphérique (P5).

4.3.2. Metrics and data labelling

Before data can be used to fit model parameters, it has to be labelled (P1, P3, P4). This can be done manually by experts, automatically or a combination of the two. A shift towards automatic annotations was visible according to two participants (P1, P3). For interactions, crashes and incidents that are automatically detected, often a human interface is used to detect false positives (P1). Participant P1 also mentioned that his company aimed to have HMI’s so customers can give feedback on scenarios. Another participant mentioned that a robotaxi company also reaches out to the public in that they have an open dataset so everyone can see and help labelling incidents and accidents (P4).

For crashes and incidents, measures such as time-to-collision or even hard braking can be used to label part of the data as a safety critical situation (P5). GPS location is used to determine if data was obtained in a specific area (P5). Measures such as average speed, average time headway or how quickly lane changes are performed are used to describe general driving behaviour in geographic locations (P5). Participant P1 mentioned that in other cases “behaviour” is difficult to define in terms of measurable signals.

Example challenge for extracting metrics from data:

  • How does a child behave in a school area? Which measures would one use to de-scribe it so it can be used in an algorithm? (P1)

4.3.3. Communication between disciplines

Two participants mentioned the effect of differences in disciplinary background on the development of AVs (P4, P6). This is specifically true for engineers who are designing the systems and psychologists who investigate how the systems’ behaviour is perceived. For example, psychologists are knowledgeable on the various messages conveyed by different braking patterns in diverse situations. However, translating this knowledge into more quantitative engineering specifications for a braking algorithm proves to be challenging.

The importance of communication between disciplines is further highlighted by the fact that these participants (P4, P6) explicitly mentioned the effect of the interview itself on their view of social driving behaviour in traffic. Participant P4 mentioned the interview created more awareness of the complexity and importance of social driving behaviour in traffic, while participant P6 mentioned the interview highlighted the benefit of a handbook of social driving behaviour for engineers, which could be used to quantify and evaluate such behaviour.

Examples of communication between disciplines:

  • Communication between different disciplines is difficult due to differences in individ-ual background (P4).

  • Engineers delve too much in mathematical methods instead of looking at social norms and questions (P6).

4.3.4. Approaches

The participants mentioned that there are several approaches to defining the algorithms in the AV subsystems: rule-based, machine learning and hybrid approaches (P3, P5, P6, P7). In the rule-based approach, the algorithm is based on a set of rules designed by experts (P3, P7). Data is used to fit the corresponding model parameters (P3, P7). A main benefit of the rule-based approach is the transparency of the algorithms, as they are based on explicit rules (P2, P3, P6). Additionally, there are less parameters so less data is needed to fit these models (P3). In the machine learning approach, the algorithm is based on patterns found in the data during a learning phase (P1, P2, P3, P6). These models have many parameters and thus need large amounts of data to be fitted (P3). Additionally, especially for complex systems, this approach results in black box models that are hard to interpret (P2, P3).

For complex tasks such as automated driving, it can be nearly impossible to define all necessary rules for safe driving (P6), which is why more and more use is made of machine learning approached (P3, P6). These models severely lean on the quality and quantity of the data that is used, and with an ever-changing dynamic environment that is traffic, gathering and processing sufficient data is a huge challenge on its own (P1). In a hybrid approach both rule-based and machine learning approaches are combined, ideally leveraging the best of both approaches (P6).

Examples of a rule-based approach:

  • Determine future path via the factors head detection (where is someone looking), traffic light information and map information (this is the middle of the road which a par-ticipant is less likely to stand still on than on a sidewalk) (P3)

  • Machine learning approaches are for example used in perceptions algorithms, such as detection of lane markers (P1).

4.3.5. Generalizability

One of the major challenges in vehicle automation is the lack of data and the corresponding challenges with generalizability of the models (P1, P4, P7). Models can produce desired behaviour in one situation, but perform poorly in another. As traffic is a highly dynamic system, these changes can occur at any time anywhere. The importance of context in this regard was highlighted by several participants (P1, P3, P7). In other words, small changes to a scenario can result in a large difference in desired behaviour.

Examples on generalizability (taken from survey answers by P2, P3, P6, and P7 in response to the presented school scenario):

  • Scenario 1: Entering an area with a school (no official school zone). Kids are walking outside. The desired behaviour is cautious driving.

  • Scenario 1 + a ball rolls onto the street. After a while a kid appears on the side of the street. The desired behaviour is explicit communication with the children and further reduction of speed.

4.3.6. Recognition

Participants were divided on vastness of the perception capabilities of AVs. Participant P7 mentioned that there is virtually no limitation with respect to technology and object recognition. However, other participants gave multiple examples of current challenges that object recognition and intention recognition algorithms face.

A first step in the recognition process is detecting an object. A second step is recognizing this object and assigning characteristics to it. A third step can be to derive the intention of a road user based on, for example, their movement patterns and attitude. In the social science domain this roughly corresponds with the perception and comprehension levels of situation awareness (Endsley, 1995). Two participants mentioned that especially the latter, intention recognition, is a very challenging task for AVs (P1, P4). In general, the AVs are not capable yet of fully understanding the world around them (P7), i.e., assigning characteristics, importance and meaning to all that is detected.

Examples of recognition challenges:

  • Recognizing children as such and thus assigning their characteristic of being unpredictable (P1). Additionally, very small children might not be detected at all as they can be fully in the blind spot of the car (P2).

  • “There are multiple indicators for blind people, but we basically came to the conclusion that we will probably not be able to reliably detect if it is a blind person, so that means you need to always have conservative behaviour.” (P2)

  • Reliably detecting a ball as such (P2). AV companies do not classify balls as such and the object may be below a critical size that can realistically be captured (e.g. by li-dar at a certain range) (P2). Ball object could be considered anomaly or specifically detected as significant object (P3).

  • AVs currently do not use the vehicle indicator information to infer the intention of the driver. So, unless the vehicle is also swerving towards the other (weaved) lane, the AV would just be left to infer from prior statistics whether a car will weave or not (P2).

  • Recognizing school areas based on contextual cues without relying on pre-existing human-annotation map data (P2).

  • Depending on the road, the vehicle may not be able to detect the last vehicle on the right road. In any case, the detection may be late and therefore uncomfortable when it comes to planning (P4).

  • Detection of motorcyclists is difficult (P5).

  • AVs can currently not recognize emergency vehicles (P4).

  • Convoys and funeral processions are difficult to recognize (P2).

Examples of what can be perceived/recognized:

  • A sidewalk can be detected as such and assigned probabilities of a pedestrian stop-ping there rather than in the middle of the road (P3).

  • “According to Waymo, and I have no way to verify this, they can see 300m in 360 de-grees. They can even see, they claim and I have seen the animation so I believe it, they can see around corners.” (P7)

4.3.7. Prediction

All participants agreed that anticipating future scenarios makes driving safer as there will be less need for emergency manoeuvres, but that this at the same time is a big challenge for AVs (P1, P2, P3, P4, P5, P6, P7). As the future is never a 100% sure, algorithms, like humans, make multiple predictions, aiming to cover the relevant part of the prediction space (P3). Thus, hoping that the actual outcome is among those predictions. Selecting such a relevant subspace out of the potentially infinite number of possible predictions is a challenge on its own (P3).

Another challenge lies in the way these predictions are made (P3). In its simplest form, one can look at each road user independently and predict their future behaviour. However, this is a strong simplification of reality, where road users are influenced by behaviour of others. A more realistic approach is therefore to make a joint prediction, where a prediction is made of all road users collectively, including their interactions. This is a much more complex problem to solve algorithmically and simple upscaling of current methods would require significantly more computational power (P3, P6). What makes the problem even more complex is to make such joint prediction of reactions to your reaction to their behaviour, similar to what humans do (P1, P3, P6). And this prediction of reactions and all the options one has as how to react results in a large problem that at some point can no longer be solved in real time (P3). Therefore, choices must be made as to which subset of predictions are considered (P3).

Participants (P2, P4, P6) mentioned that vehicle-to-vehicle communication could aid predictions, as intentions of surrounding traffic can then be directly communicated, rather than observed from actions. However, as participant P4 mentioned, 1) this is only effective if a large part of the vehicles are using this technology which will not be for another 10 to 15 years, 2) it will not apply to human controlled ways of transport, and 3) the prediction horizon of such messages is only in the order of 5 seconds.

Example prediction challenges:

  • “I am happy when I know there is an obstacle and even more happy when I know it is a pedestrian. Getting a walking pattern is difficult, predicting it even more.” (P4)

  • Humans can detect a foreign number plate and assign characteristics of a tourist to it and with that anticipate unpredictable behaviour and increase their distance to this ve-hicle. AVs currently make no such inferences. (P1, P6)

4.3.8. Decision making (cost function)

Mainly in the planning part of the vehicle automation algorithm decisions are made on what behaviours to exhibit (P3). This can for example be decisions on which route to take (mission planning), which manoeuvres to execute (motion/path planner) or which speed or following distance to stick to (behavioural planner). These decisions are based on trade-offs of what is important (costs and benefits) and to which constraints to adhere to (P2, P3, P4, P6). Some participants mentioned that safety is mainly seen as a constraint (P2, P3), i.e., keep a minimum distance to other objects. Aspects to optimize for are trip duration and passenger comfort. Other participants mentioned that safety is, next to being a constraint, also an aspect to optimize on (P1, P5, P6), i.e., a slightly safer action can be chosen, even if it negatively influences trip duration.

Participants did not think social driving behaviour on its own was optimized for. Instead, social driving behaviour either emerged when optimizing for safety (P1, P2, P5) or was seen as a side project (P4). Participant P4 mentioned that, while not currently the case, comfort and safety of other road users should be considered in the decision-making process. For example, if certain behaviour would frustrate other drivers, this should be represented by a specific cost in the cost function. Several participants agreed that including some sort of cost for social aspects of driving behaviour in the cost function is desirable (P4, P6, P7), but that it is difficult to define this in a way that it can be implemented.

Examples of constraints and trade-offs in decision making

Constraint:

  • A hard constraint could be maintaining a minimum distance of x meters from an ob-ject (P3).
    Benefit vs constraint:

  • In a highway merging scenario, the vehicle may need to accelerate significantly to merge ahead of another car. Here, the objective (benefit) is to make progress, but the algorithm must achieve this within predefined acceleration limits (constraint) (P2)

Benefit for ego vehicle vs traffic system

  • “Creating a safe distance to the car in front will lead to less traffic jams. It’s counter-intuitive, you are leaving more space so it seems like it is inefficient, but because you are avoiding these kinds of traffic jams that can propagate in waves, you are overall more efficient. Sometimes the aspect is that it is not more efficient for you, more efficient for the hundred [road users, red.] behind you.” (P2)

Safety vs efficiency

  • Generally, giving way more often could increase safety. However, there is a trade off with making progress. The more social behaviour the safer it is, but for efficiency this might not be the case (P2).

  • There are some decisions that are different based on the type of road user that the AV interacts with. For instance, motorcyclists are vulnerable, so AVs might take this into account by braking slower than usual to try and prevent a collision. In this case, the safety of the other road user (or staying out of crash statistics) can outweigh efficien-cy (P5).

4.3.9. Target actions

Participant P7 mentioned that AVs are expected to excel in collision avoidance (i.e., reactive behaviour), such as stopping for a child that suddenly runs onto the road, but that avoiding such emergency situations in the first place (i.e., proactive behaviour), is more difficult for AVs. Accordingly, target actions can be classified as reactive or proactive. Proactive measures are focused on making the task of the AV simpler. However, this is often done without any regard for any secondary effects this might have on surrounding road users. Participant P4 noted that cautious driving behaviour might be only applicable to the early generation AVs. This behaviour might not be representative for next generation AVs. Proactive target actions can be further stratified in terms of Michon’s (1985) hierarchical model of driving: operational (associated with automatic action patterns), tactical (associated with controlled action patterns), and strategical control (associated with general plans). Most of the participants (P1, P2, P3, P4, P5, P6) mentioned the ability of AVs to proactively adjust their driving style to circumstances at the operational level. At the tactical level, some participants (P2, P5, P7) provided examples of proactive manoeuvres that AVs execute or proactively not execute to increase safety. Some participants (P2, P3, P5) mentioned strategical decisions to increase safety.

Examples of reactive target actions:

  • AVs currently do not use vehicle light information to infer the intention of the driver and thus do not directly react to such signal. (P2)

  • Braking behaviour is not only relevant for obstacle avoidance, but also for communication to other road users (P1, P4, P6, P7). P6 mentioned that their vehicle already brakes gradually to indicate cooperation and interaction.

Examples of proactive target actions:

  • Operational: Speed is adapted to the presence of children (P3), school areas and suburban areas (P1), if they can be detected as such.

  • Operational: Lane position is adapted to the lead vehicle (P4, P5) to make sure that they shift to the side similar to surrounding human driven traffic when an emergency vehicle is approaching, leaving a rescue lane. Another reason to stay behind the lead vehicle is to avoid wrong-way drivers, as they are hard to detect for an AV on its own (P5).

  • Tactical: Avoiding harsh braking and lane changes in the vicinity of motorcycles (P5). Touching a motorcycle will have a more disastrous effect on the safety score of AVs then does touching another car. A motorcycle can lose balance and the chance of grave injury or death is much higher. Optimizing for this safety score thus results in avoiding such manoeuvres around motorcycles.

  • Tactical: Not entering an intersection because the crowdedness creates a too complex scenario for the AV to handle (P2).

  • Strategical: Avoiding specific areas such as school zones (P2, P5). Due to the high likelihood of unpredictable children in these areas, creating complex situations, AVs can be programmed to avoid these areas altogether.

  • Strategical: Only drive in restricted areas (P3). Some shuttle, for example, are only designed to drive in a designated area where interaction with other road users is limited.

4.4. Process phase III: Evaluation

Both during and after developments of the AV an evaluation on their safety performance is done. We made a distinction between participant input regarding assessment during the development or type approval of the AV, and monitoring of their safety after they are allowed on the road. Monitoring can take place at the level of an individual vehicle, and at the level of the traffic system.

Participant P5 mentioned the multi-pillar approach for assessment of automated driving systems (UNECE, 2022). In this approach the three original pillars for safety assessment, track testing, real world testing and audits/assessment, are extended with two additional pillars: simulation/virtual testing and in-service monitoring and reporting. These five pillars all use a scenario catalogue which contains descriptions of real-world driving situations that may occur during a given trip.

4.4.1. Assessment

Similar to the challenges mentioned during the requirement phase, also for assessment of social driving behaviour the challenge is on how to define the behaviour of a competent and careful driver (P4, P6, P7). On top of that, assessment of adhering to such definition requires a measurement method and threshold definition to determine if the vehicle’s safety performance is adequate (P5).

Examples of choices to be made regarding assessment

  • Determining the acceptable threshold for errors of AVs, whether it is 1% or 1‰, is a societal discussion. It is a technological discussion how reliable the calculated error likelihood is. (P3)

  • What is a competent, social, adequate driver differs per country, per time of the day, etc. In Italy it means something else than in Sweden. (P7)

4.4.2. Monitoring of vehicles

Many things are tested in simulation, but as participant P3 mentioned, not everything can be tested in simulation. Edge cases (e.g., circumstances that the system may not have been specifically trained or programmed to handle) are often only noticed when driving on the public road (P4). Descriptions of edge cases can be insightful into understanding what current technological challenges and capabilities are.

Examples of edge cases:

  • Some AVs require an unobstructed trajectory and have limited capacities to circumvent objects on the road. An actual example of this limitation was demonstrated when the front wheel of a bicyclist waiting to cross is partially on the road caused an AV, intending to move into the intersection leg where the bicyclist was waiting, to stop mid-intersection. As a result, the intersection was blocked for other traffic. (P7) (personal experience)

  • Driving on the highway, there are cones and flares, it is raining and it is night, causing flashes and reflections on the lines on the road. In this edge case scenario, the AV must detect the police officer who is signalling. (P7)

  • A pedestrian walking faster than the expected range of walking speeds, i.e., the characteristics don’t match up with the data. (P1)

Participant P5 mentioned the monitoring process as it is tested with two automotive companies currently. Data should be logged during accidents using the Data Storage System for Automated Driving, after which an automated process is started to improve the system and avoid such acci-dents the future. Also, incidents should be monitored, but only detectable incidents are consid-ered. The examples below highlight the stark differences in interpretations among companies on what constitutes a detectable incident requiring monitoring. P5 noted that the Company 1 would not receive EU type approval for their AV system with their definition of detectable incidents.

Example incident definitions:

  • Company 1 only wanted to define detectable incidents as occurrences when an airbag was deployed (P5).

  • Company 2 wanted to define it as all occurrences when contact between two road us-ers is registered, as well as, for example, hard braking. Also near misses, where an-other road user was very close by the AV, could be registered as incidents (P5).

4.4.3. Monitoring of the traffic system

Several participants highlighted the difference between a safe vehicle and a safe traffic system (P5, P6, P7). They highlighted the need for looking at traffic safety, i.e., all collisions happening on the road, rather than vehicle safety of specific vehicles alone, when inserting AVs in the traffic system. Two participants commented: “Traffic safety is a consequence of all interactions” (P7), and “All actions, even if you do them for traffic safety, might increase risk (in another way), the question is, how much will it increase risk as a total” (P6). Several participants mentioned the expected impact of AVs on traffic safety. Some expect improvements, some see that updates make AVs better, some are not convinced of improvement and some expect initial improvement and then a deterioration. As participant P7 put it: “There is an equilibrium at the moment that yields the number of traffic accidents we have now […] If we disrupt the system with this new vehicle, we do not know where we end up.”

Examples of effects at the level of the traffic system:

  • Risky driving of human drivers due to overreliance on the safe keeping capabilities of AVs (P6)

  • Risky driving of human drivers out of frustration with the AVs (P4)

  • Human drivers adapting to divergent behaviour of AVs (P3, P5)

  • Human road users bullying AVs (P2, P3)

  • Chain instability when the penetration rate of AVs increases, due to lack of anticipation capabilities of AVs (P5)

  • Lack of rule braking of AVs: Matrix signs might not switch off after a traffic jam has dissolved if AVs always keep to the speed limit communicated by the sign. These signs use the speed of the traffic to determine if a traffic jam is dissolving and the sign can thus be shut off. (P5)

5. Discussion

The present study seeks to accommodate interdisciplinary collaboration between social scientists and engineers, by providing insights into current AV technological capabilities regarding social driving behaviour and road safety, and the related challenges from an engineering perspective. To that effect, seven interviews were held with engineers working in the AV domain. The research question “how can social scientists use their knowledge to improve social driving behaviour of AVs?” was leading for the interviews. Direct answers to this question are shown in italic font and further elaborated on in the text below.

5.1. Main takeaways

The interviews highlighted that the materialization of social driving behaviour in AVs is influenced at all stages of the development process, i.e., during the requirements phase, the development phase and the evaluation phase. As demonstrated in previous research (Brown et al., 2022, 2023; Vinkhuyzen & Cefkin, 2016) social scientists can use their knowledge in each of these stages to improve social behaviour of AVs on the road. Main takeaways and challenges are discussed next, and visualized in Figure 3 (numbered circles are referred to in the text as #<number>).

Figure 3
Figure 3 Takeaway points and challenges across all development process phases

SDB = Social driving behaviour. 1) Consensus that implementing SDB can improve safety, 2) A handbook of SDB is needed with clear definitions and measures (based) on what constitutes a competent and careful human driver, 3) Safety involves the ego vehicle within a traffic system, 4) Lack of communication between disciplines, 5) Consensus that a top-down incentive is needed for implementing SDB, 6) SDB aspects are not a goal but an emergent property when evaluating safety, 7) Proactive actions generally aim at making the task easier for the AV without accounting for secondary effects on other road users, 8) Large differences exist across companies in how they deal with safety, 9) Other road users are viewed as obstacles without accounting for their needs and reactions, 10) Reliably predicting behaviour of other is a challenge, 11) Reliable recognition of contextual elements is a challenge, 12) A system that can generalize across situations is a challenge.

5.1.1. Communication between disciplines

Initial ideas on the term social driving behaviour elicited positive associations of driving behaviour in the participants, such as yielding the right of way and taking into account other road users. At the start of the interview instead a value-free definition of social aspects of driving behaviour was introduced: “Driving behaviour that directly or indirectly influences and/or takes into account other road users, i.e. their state, behaviour or goals.” This definition allows for both pro-social and anti-social interpretations of behaviour and is similar to the use of the term ‘social behaviour’ in the social sciences (Schmitt, 1998). Social scientists should be aware that these definitions can differ when interacting with those outside of this field, such as engineering (#4).

Pursuing similar definitions across disciplines for social driving behaviour is important, as it affects the perceived relevance of the subject for AVs. For example, at the end of the interview two participants explicitly mentioned how the interview highlighted the importance and the complexity of social driving behaviour for safe participation in traffic. It seems that the broad definition of social driving behaviour is underlying to this understanding. When discussing the need to implement social driving behaviour in AVs, focusing solely on pro-social behaviour such as yielding the right of way may make it appear as a trivial or low-priority feature to implement. However, considering social driving behaviour in its full scope reveals that, without intentional design, AVs will not simply lack social driving behaviour but may (inadvertently) exhibit anti-social behaviour on the road (Brown et al., 2023; Landolfi & Dragan, 2018). Given this full scope of the definition, a consensus was revealed among the participating engineers on the beneficial effect of implementing social driving behaviour in AVs on traffic safety (#1). One participant even deemed social driving behaviour a prerequisite for reaching the required safety targets.

Next to a shared understanding of basic social science knowledge, collaboration between disciplines is also fostered by sharing basic engineering knowledge, such as the capabilities and limitations of AV technology. For example, AVs can perceive their surroundings more accurately and on a wider range than human road users, but the interviews revealed that basic interpretation of situations and their elements (e.g., children, blind people) is still a challenge (11#). For example, including contextual elements, such as the relevance of a ball rolling onto the street when predicting behaviour of a child, might seem natural for the human road user, but is still a challenge for AVs (#11). Consequentially, the interviews revealed that prediction of the behaviour of these elements, their joint prediction and predicting multiple cycles ahead (i.e., potential action and reactions) is a major challenge (#10). This is in sharp contrast to how vehicle performance is often portrayed by the automotive industry [e.g., Waymo (n.d.-b)]. To avoid the pitfall of proposing utopian standards that are unlikely to be realized in practice, one should define requirements that are not only theoretically sound, but also practically achievable. Therefore, in line with the study by Vinkhuyzen and Cefkin (2016), interdisciplinary collaboration on implementing social driving behaviour may benefit from social scientists’ awareness of engineering challenges.

5.1.2. Requirements regarding social driving behaviour

The interviews revealed that many higher-level requirements are formulated prior to the development process, but as the development progresses additional lower-level requirements (i.e., for specific technical implementation) may emerge when trying to meet, for example, the safety targets. Two approaches to incorporate requirements on social driving behaviour were identified. First, participants believe that a top-down incentive (i.e., legislation) is the easiest way in which requirements related to social driving behaviour would be adopted at the start of the development process (#5). Here lies a vital role for social scientists in using their knowledge on social driving behaviour to inform legislators and aid in developing appropriate regulatory requirements, as also reflected in previous literature (Brown et al., 2022; Straub & Schaefer, 2019).

The second approach identified aims to influence the lower-level requirements. AV engineers could be educated on preferred behaviour and potential consequences of AV behaviour on social interactions in traffic, with the aim that such knowledge is used when defining low level requirements. Here lies a vital role for social scientists as the educators.

Both approaches though, were said to not include social driving behaviour as a goal, but rather as an emergent property when aiming to reach certain safety goals set by the company or regulatory bodies (#6). The interviews made clear that a major challenge for both approaches is finding a way to translate the qualitative requirements defined from a social sciences perspective to quantitative requirements that can be used by engineers. The ‘handbook’ on social driving behaviour that results from these translations requires a common understanding between the social science and engineering disciplines and is currently lacking (#2). Social scientists play a vital role in developing such a handbook in collaboration with engineers, as also suggested in previous research (Brown et al., 2022; Quante et al., 2024; Vinkhuyzen & Cefkin, 2016).

Automotive companies’ requirements for AVs differ per manufacturer (#8) and are shaped through input from various stakeholders, including legal authorities, customers, and society. The interviews revealed that legislation typically outlines mandatory requirements, whereas customer preferences inform additional, company-specific requirements. For instance, robotaxi companies may prioritize cautious driving behaviour (Rahmani et al., 2024), while manufacturers of personal vehicles might require options for the customer to select more aggressive driving styles (Tesla, 2024). The interviews further highlighted that companies can also differ in the safety targets they set for themselves on top of the legal requirements, which is reflected in the broad spectrum of star ratings for different vehicle models by Euro NCAP (Euro NCAP, 2024). The interviews also revealed that when it comes to design choices there are large differences between companies. For example, Tesla is designing their SAE level 2 system “Full Self Driving (supervised)”, which can be used everywhere, but lacks the reliability to drive without human supervision. Mercedes, on the other hand, focusses on high reliability of their SAE level 3 Drive Pilot feature that does not require human supervision, but has a very limited operational design domain. The interviews furthermore revealed that societal requirements, such as creating a safe and comfortable driving environment for all road users, are usually only adopted when they align with legal and customer demands. For effective collaboration, social scientists should recognize these differences when discussing social driving behaviour with specific automotive companies.

5.1.3. Evaluation of traffic safety

Traffic is a social system in which behaviour is governed by, amongst others, the norms and values of a large group of road users (Tennant et al., 2021). Changing the behaviour of a significant portion of this group by introducing AVs can influence these norms, values and corresponding behaviours of all road users and with that, the safety within the complete traffic system (Cohen et al., 2020; Straub & Schaefer, 2019). From the interviews, however, a focus on ‘single vehicle safety’ emerged (#7), as opposed to safety of the traffic system (referred to as the ‘second vehicle problem’ by Straub and Schaefer (2019)). Single vehicle safety generally considers only direct interactions between the ego vehicle and other road users, whereas in the second vehicle problem another road user’s unexpected actions while interacting with an AV has a cascading negative effect on the second or third road user behind them. Another example of indirect effects of AV behaviour on traffic safety was found by Knoop et al. (2019), where the behaviour of a platoon of AVs causes frustration in other road users who then initiate dangerous manoeuvres. Knowledge on social driving behaviour can aid in highlighting the difference between vehicle and traffic safety and developing requirements to improve safety at both levels.

Although the interviews revealed consensus on the relevance of social driving behaviour for traffic safety, the magnitude of its impact and the most influential behaviours remained unclear. One way to highlight which social behaviours are relevant could be to put more focus on traffic safety, rather than individual vehicle safety, in the evaluation process of AVs (#3). Social scientists can play an important role in this evaluation process by providing input for performance metrics to assess the quality of social interactions (e.g., Quante et al. (2024)). The evaluation process is currently based on scenario-based testing, making the selection of scenarios crucial for adequate evaluation (Riedmaier et al., 2020; Sánchez et al., 2022). The interviews revealed that it is currently unclear which scenarios should be tested and what exactly is meant in these scenarios with the UNECE requirement to minimize risks to at least the level of a competent and careful human driver (UNECE 2021). Selecting relevant scenarios based on crash databases, which is often done for ADAS evaluation (Aleksa et al., 2024), might not suffice for evaluating AVs, as these databases mainly contain crashes between human driven vehicles. Situations which are at high risk for AVs might be significantly different (Liu et al., 2021). Social scientists could aid the development of new scenarios through identification of troublesome interactions between AVs and other road users in data of both pre-deployment on-road tests and post-deployment monitoring, in line with previous research (Brown et al., 2023; Brown & Laurier, 2017; Cleij et al., 2024; de Gelder et al., 2024).

5.1.4. Development of social driving behaviour

Social scientists can also use their knowledge in the development phase to impact social driving behaviour of AVs, for example by providing domain expert knowledge to the white box part of hybrid automated driving algorithms (Paardekooper et al., 2021). A hybrid algorithm may use machine learning on a perception level for lane marking recognition (black box), while implementing hardcoded rules on a cognitive level, such as expecting participants to stand still on a sidewalk and not in the middle of the road (white box). The previously mentioned handbook with quantitative descriptions of social driving behaviour (#2) that social scientists could develop, would be beneficial for the development of the white box features of such systems. In Vinkhuyzen and Cefkin (2016), for example, the social driving behaviour of following a car at an intersection called “piggybacking”, was put forth as good candidate to program into the perception and possible action of the AVs of Nissan.

Social scientists can also aid in understanding and improving the effect that AV behaviour has on other road users. The interviews suggest that the reason for implementing proactive safety measures is often due to the notion that AVs are still far from performing well in all traffic situations (e.g., avoiding school zones or increasing safety margins in the presence of cyclists). These measures focus on making the task of the automation easier, but do not always take secondary effects on other road users into account (#7). For example, driving with a large time headway or reducing speed to increase safety margins on a crowded highway can agitate the drivers of following vehicles who in turn might perform dangerous overtaking manoeuvres (Knoop et al., 2019). Also, stopping in the middle of the road when a situation is too complex for the AV to handle can block traffic, confuse other road users, and consequently decrease efficiency and traffic safety (Brown et al., 2023). Social scientists’ knowledge on real world interactions between road users can be further developed and used to predict implications of such proactive safety measures so that they can be optimized towards minimal negative impact on other road users in terms of safety, comfort and efficiency.

Similarly, social scientists’ knowledge on real world interactions between road users could improve social driving behaviour of AVs, in terms of direct and indirect communications. The interviews revealed that AVs are often programmed to see other road users merely as obstacles without accounting for their needs and possible reactions (#9). However, as stated by Brown et al. (2022) "If we are to understand, design, regulate and critique autonomous systems it is important that we also understand how they act interactionally in practice, in our pre-existing social world.". Somewhat surprisingly, eHMIs (i.e., external communication devices on AVs that communicate the intentions of the AV to other road users) did not emerge as a theme from the interviews. Several studies have examined effects of eHMIs on pedestrian crossing behaviour (e.g., Eisele & Petzoldt, 2022; Faas et al., 2020; Kooijman et al., 2019) Possibly, the engineers interviewed in the present study have been working on algorithm development, whereas eHMIs may be developed in other (design oriented) departments. Alternatively, eHMIs were not addressed if the engineers deemed them unnecessary, in line with arguments raised in de Winter and Dodou (2022).

And finally, social scientists could potentially aid the challenge of generalization of behaviour over different traffic scenarios (#12). Traffic is highly dynamic, and different traffic situations require dealing with different sets of parameters (e.g., different road users, different environments). Small variations in these parameters, such as the position of puddles on the road or the age group of a pedestrian, can significantly impact the (predicted) behaviour of other road users and consequently the desired behaviour of the AV. For example, a pedestrian generally will take the most direct path to cross the street, but will adjust their trajectory if a large puddle is present on this path. This inherent variability makes it challenging to generalize automated driving algorithms. Several study approaches that can help tackle this challenge are already found in literature, especially related to pedestrian crossings. Zhao et al. (2024), for example, identified the presence and behaviour of other pedestrians through a virtual reality experiment as a factor that influences the understanding of and reaction to an AV’s eHMI. Sarker et al. (2024) identified significant factors influencing the drivers’ yielding behaviour from results of a survey among drivers and focus group interviews in combination with behavioural change theory. And Li et al. (2025) collected and published naturalistic driver–pedestrian-yielding data from 18 unsignalized intersections across Minnesota. The data was used to investigate the impact of the built environment on driver-yielding behaviour based on 50 distinct contextual variables. In line with these studies, social scientists can address the challenge of generalization in AV algorithms by conducting further research into identifying and categorizing factors that influence human road user behaviour in different traffic situations.

5.2. Limitations and recommendations

A notable limitation of this study is the small sample size, limiting the possibilities to perform quantitative analysis on verbal data. The goal of the study, however, was not to provide an exhaustive list of examples and challenges relating to social driving behaviour, but rather to explore the phenomenon from a broad engineering perspective. For this reason, engineers working in different domains of AV development (i.e., industry, academia, research institutes, vehicle authorities) were selected for the interviews. In future research more engineers could be interviewed to gain an even broader perspective. For example, none of the engineers mentioned that vehicle to infrastructure communication could aid the AV with reading the road. Another recommendation is to identify priorities for tackling the challenges put forth in this paper. This was not a goal of this study, and due to the limited number of participants and the subjectivity that comes with semi-structured interviews, such priorities can also not be deduced from the presented results. Instead, the presented results can, for example, form the basis for a large-scale questionnaire that specifically probes priorities in this set of challenges. The qualitative insights gathered provide a preliminary exploration into the development of AVs, particularly regarding social driving behaviour. The findings offer an initial view into this emerging field, helping to bridge the gap between the technical complexities of AV algorithms and their manifestation in real-world driving scenarios. By translating intricate concepts, such as cost functions and recognition and prediction algorithms, into observable road behaviours, this study enhances the accessibility of these technical aspects for social scientists, facilitating interdisciplinary understanding and collaboration. It is recommended that in future research such translations from the technical to the social domain are also adopted, to further facilitate interdisciplinary research and development.

The interviews highlighted that social scientists can contribute to the development of social driving behaviour in AVs by creating a handbook that provides formal and quantifiable descriptions of these behaviours. Whether human behaviour in traffic is indeed quantifiable at the level of detail desired by engineers is a challenge on its own. As a start, we imagine that a handbook could be based on probabilities of other road user behaviour as function of context, road user characteristics, and corresponding ranges of acceptable AV behaviour. Developing this handbook will likely require substantial collaboration across various disciplines. Future research should therefore investigate how to best organise such collaboration and further define what such a handbook should look like.

A final recommendation applies to the regulatory domain. As AVs can considerably change the dynamics of the complete traffic system, their evaluation should also be expanded from single vehicle safety to their effect on safety (and comfort) in the larger traffic system. Such evaluations will furthermore provide a top-down incentive to design automatic vehicles that benefit the whole of society.

6. Conclusion

As AVs will likely take part in traffic more and more often in the near future, their ability to safely interact in this social environment will become more important. To ascertain safe interactions between AVs and other road users in traffic it is essential that parties responsible for developing and testing these vehicles have a proper understanding of manifestations of social driving behaviour and their consequences for traffic safety. Such understanding is crucial as it may in turn guide the implementation of algorithms and cost functions underlying vehicle automation technology. This interview study takes a first step toward this goal by highlighting engineering challenges and identifying how social scientists can contribute.

The main finding is a lack of well-defined high-level requirements for social driving behaviour and the consequent reliance on engineers to establish these requirements through trial and error during on-road testing. Social scientists can support the establishment of these requirements earlier in the development process by compiling a handbook of social driving behaviour, which includes formal and quantifiable descriptions of these behaviours.

A technological challenge that engineers still face is recognizing other road users in traffic and predicting their behaviour, a fundamental aspect of social driving behaviour. To encourage more research and development efforts in this domain and, consequently, support the safe introduction of AVs on our roads, requirements regarding social driving behaviour should be incorporated in the evaluation processes of AVs.


Acknowledgement

The authors would like to thank their colleague Celina Mons, MSc., for her valuable input on the draft of this manuscript.

CRediT contribution statement

Diane Cleij: Conceptualization, Data curation, Investigation, Methodology, Project administration, Supervision, Validation, Writing—original draft, Writing—review & editing. Rins de Zwart: Conceptualization, Data curation, Investigation, Methodology, Validation, Writing—original draft, Writing—review & editing. Reinier J. Jansen: Conceptualization, Investigation, Methodology, Visualization, Writing—original draft, Writing—review & editing.

Declaration of competing interests

The authors declare that they have no conflict of interest.

Ethics statement

The methods for data collection in the present study have been approved by the ethical committee of SWOV.

Funding

This study was funded by the Dutch Ministry of Infrastructure and Water Management. The funder played no role in the study design, collection, analysis, interpretation of data, writing of the report, or in the decision to submit the paper for publication. The funder accepts no responsibility for the contents.

Declaration of generative AI use in writing

During the preparation of this work the authors used ChatGPT to improve readability. After using this tool, the authors reviewed and edited the content as needed and take full responsibility for the content of the publication.

Editorial information

Handling editor: Haneen Farah, Delft University of Technology, the Netherlands.

Reviewers: Laurent Carnis, Université Gustave Eiffel, France; Andrea Paliotto, University of Florence, Italy.

Submitted: 2 December 2024; Accepted: 3 March 2025; Published: 25 March 2025.

Appendix

Pre-interview survey

A.1. Background questions

  1. What is your current age?

  2. What is your current profession?

  3. What is your professional experience in the development of autonomous vehicles?

A.2. Scenarios

Three scenarios were prepared for the survey: a right-angle conflict with a bicycle (see Figure A.1), merging on a highway (Figure A.2), and approaching a school (Figure A.3). For each scenario, the corresponding panels appeared one-at-a-time, accompanied by a short description of the panel (see captions in Figures A.1, A.2, A.3), and the following three questions:

  1. What behaviour of (the driver of) the blue vehicle ensures that a conflict is avoided?

  2. Which cues and/or information facilitate(s) the above behaviour?

  3. Would your answers differ if the blue vehicle is an automated vehicle? And if so, how?

Figure A.1
Figure A.1 Scenario 1: right-angle conflict with a bicyclist

Panel A: “Driving in a city.” Panel B: “You need to turn into a street ahead by crossing a bicycle path. Another car is driving close behind.” Panel C: “There is a bicyclist on the bicycle path next to you.”

Figure A.2
Figure A.2 Scenario 2: merging on a highway

Panel A: “Driving on the highway during rush hour with weaving section in the near future.” Panel B: “Approaching a busy weaving section (Route: stay on current road).” Panel C: “Behind a slow vehicle when passing the weaving section.”

Figure A.3
Figure A.3 Scenario 3: approaching a school

Panel A: “Driving home during school hours.” Panel B: “Entering an area with a school (no official school zone). Kids are walking outside.” Panel C: “A ball rolls onto the street. After a while a kid appears on the side of the street.”

A.3. Questions on definition and application of social driving behaviour

  1. How would you define ‘social driving behaviour’ from the perspective of an autonomous vehicle?

  2. What is your professional experience with ‘social driving behaviour’ from the perspective of an autonomous vehicle? Please tick all boxes that apply.

  • Research

  • Development

  • Policy

  • None

  • Other

  1. Do you foresee market-ready implementations of ‘social driving behaviour’ in autonomous vehicles within 5 years from now?
  • Implementations are already on the road.

  • Within 5 years is feasible

  • More than 5 years are required.

  • Social driving behaviour by autonomous vehicles is not feasible.

Interview structure

B.1. Questions on social aspects of driving behaviour

  1. Is this definition of social driving behaviour clear, do you have any questions? “Driving behaviour that directly or indirectly influences and/or takes into account other road users, i.e. their state, behaviour or goals.”

  2. Can you give some examples of implementations of social driving behaviour? And do you expect these to be on the market within 5 years?

B.2. Questions on the development process

  1. Are there requirements to ensure that certain social driving behaviour is implemented? For example, is social driving behaviour part of the design brief? Can you relate this to the examples?

  2. Can you elaborate on how algorithms for social driving behaviour are established? E.g., what kind of decisions are implemented within an algorithm, and which inputs are used to inform those decisions?

  3. All technology has a certain operational design domain. Which limits would you expect regarding social driving behaviour? In which real-world driving scenarios would you expect these limits to emerge? How would these limitations and/or scenarios relate to the examples?

  4. Humans make mistakes, and -hopefully- learn from them. Likewise, AVs may make mistakes. How do you expect AVs will learn from their mistakes, and what would be the role of humans in this learning process?

B.3. Questions on the relation with traffic safety

  1. How would you expect that social driving behaviour by AVs will influence traffic safety? Can you relate this to any of the examples you mentioned before?

  2. Social driving behaviour by AVs may influence the behaviour of other road users. How would you expect this to affect traffic safety? Again, can you relate this to any of the examples previously mentioned?

B.4. Questions on challenges

  1. Which challenges do you see towards realizing the potential benefits of social driving behaviour by AVs on traffic safety? How do these challenges relate to the previous examples?

  2. How would you classify the challenges? Technology oriented, human factors oriented, policy and evaluation oriented, and/or other types of challenges? Can you elaborate why?

References

This Is an open-access article distributed under the terms Of the Creative Commons Attribution 4.0 International License (CCBY-4.0). View this license's legal deed at https://​creativecommons.​org/​licenses/​by/​4.0 and legal code at https://​creativecommons.​org/​licenses/​by/​4.0/​legalcode for more information.