Study suggests self-driving cars may need to adapt to share roads safely with runners

A new study on how runners may choose to interact with self-driving cars is challenging assumptions on how automated vehicles will navigate safely on the roads of the future.

Researchers at the University of Glasgow and KAIST in South Korea led the study, which used augmented reality tech to explore for the first time how runners’ behaviour differs from walkers when they are crossing roads and junctions.

Their research revealed that runners are much more likely to take risks when negotiating traffic than walkers are. During the study, the runners often took less time to process the road conditions around them and
sometimes chose to run ahead of oncoming cars in order to maintain their target pace. On several occasions, they were ‘struck’ by virtual vehicles in the team’s simulated road tests.

The team’s findings could help develop better safety systems for automated vehicles, or AVs, which have largely been trained to expect the people around them to behave much more cautiously while walking.

The team suggest that displays of lights on the exteriors of cars called ‘external Human-Machine Interfaces’, or eHMIs, could enable them to communicate their intentions more quickly and effectively.

Well-designed eHMIs could replace the usual non-verbal cues like waves, eye movements, and deceleration which human drivers currently use to show people around them how they intend to proceed. Based on their
research, the team suggest an eHMI design called DualBeam, which uses new types of lights on the vehicle to help runners make swift but well-informed crossing decisions.

Professor Stephen Brewster, of the University of Glasgow’s School of Computing Science said: “In our research group, we’ve been working for several years now on developing eHMIs which could help self-driving
cars share the roads safely with vulnerable road users such as cyclists. In doing so, we realised that there’s been very little research into how runners might expect to interact with driverless cars.

“That’s surprising on a couple of levels. One is that running is the most popular physical activity in the world, with more than 600 million people estimated to run recreationally. The other is that self-driving car journeys are booming, with a million AV trips a month in the USA alone. Clearly, it will be increasingly important to ensure that runners and AVs can share the roads safely in the years to come. We were keen to explore how self-driving cars could use eHMIs to speak the language of runners as well as bike riders to help maximise road safety.”

In order to test how walkers and runners might interact and communicate with self-driving cars, the team used an augmented-reality headset to mix real-world conditions with a life-size simulated AV. The virtual environment enabled them to safely test how self-driving cars might behave around runners and walkers in the future.

The study’s 24 participants went outdoors wearing the headset, which showed them a virtual urban environment overlaid on the real world. The participants either walked or ran towards a junction with a simulated AV approaching.

The simulated car had either no safety features displayed on its exterior, or one of two eHMIs. One eHMI, a ring of lights around the car the team called a LightRing, showed green to indicate it was safe for the participant to continue or red to help them decide to stop. The other, a strip of animated cyan lights called CyanBand, swept the lights inward to show it is slowing and in the opposite direction to display acceleration. Sometimes the car stopped for the participants, and sometimes it did not.

Although both runners and walkers reported that they found it easier to determine the car’s intentions if it used an eHMI, the differences between their behaviour in all cases was striking. When walking, participants had time to slow down or stop and verify the eHMI signals with the vehicle’s driving or braking behaviour. This
increased their trust in the AV’s intentions and allowed them to make safe and informed crossing decisions.

In contrast, when running, participants felt more motivated to cross to maintain exercise. They were unlikely to slow down or stop when approaching the crossing. This gave them less time to make informed decisions, resulting in an overreliance on the display, riskier crossing decisions, and a lack of trust on the AV’s intentions.

The runners struggled to process the animation of the CyanBand lights in the limited time they had to focus on them while approaching the junction, while walkers found it useful. The LightRing display’s simple red-and-green colour scheme, on the other hand, was much more immediately legible to both groups.

Runners also collided with the vehicle three times during the study, while walkers avoided any contact with the cars. In two of the collisions, the runners saw a red light but chose to run ahead anyway, misjudging the vehicle’s pace and causing a simulated collision.

The team’s paper, titled ‘Running into Traffic: Investigating External Human-Machine Interfaces for Automated Vehicle-Runner Interaction’, will be presented at the CHI 2026 conference in Barcelona on Friday 17 April.

(Pictures: University of Glasgow(

Facebook
Twitter
LinkedIn
Email
Print

Related Stories

HIGHWAYS... DAILY

All the latest highways news direct to your inbox every week day

Subscribe now