Language Model Enhances Road Navigation, Provides Commentary

Wayve’s AI system is designed to improve driving behavior of autonomous vehicles

Ben Wodecki, Junior Editor - AI Business

April 29, 2024

2 Min Read
An in-vehicle view of a road journey in central London

British autonomous vehicle startup Wayve has developed an AI system that combines a language model with visual data to better navigate roads while explaining its decisions in real time.

LINGO-2 is a vision-language-action model (VLAM), that combines language with vision capabilities. It can perceive the world around it while controlling the vehicle.

The latest iteration of the LINGO-1 model, which was released in November 2023, was designed to improve the driving behavior of an autonomous vehicle.

The model processes images from a vehicle’s cameras, feeding that data along with information about a road’s speed rules into a language model to provide a continuous commentary of its driving decisions.

For example, it can explain that it is slowing down for pedestrians on the road or executing an overtaking maneuver.

The model can also be prompted to perform commands like “pull over” or “turn right.” LINGO-2 can also predict and respond to questions about the road ahead and the decisions it is making while driving.

Wayve conducted the first VLAM test on a public road, deploying LINGO-2 on a drive through central London.

LINGO-2 followed a route and was able to change lanes, reduce speed to adjust to traffic conditions, instruct on how to safely pass a bus and stop at red lights.

Related:Senators Demand Crackdown on Self-Driving Tech

LINGO-2’s underlying mechanics combine Wayve’s vision model with an auto-regressive language model.

The visual model processes camera images into a sequence of tokens, combining those tokens with variables like current speed and the active route which are fed into the language model.

The language model, trained to predict a driving trajectory then produces commentary text in which the car’s controller then performs the next driving action.


The startup said models like LINGO-2 that combine visual capabilities with language “opens up new possibilities for accelerating learning with natural language.”

“Natural language interfaces could, even in the future, allow users to engage in conversations with the driving model, making it easier for people to understand these systems and build trust,” according to a Wayve blog post.

Wayve plans to conduct further research on the safety of controlling the car’s behavior with language.

Having conducted off-road tests in its virtual simulator platform Ghost Gym, Wayve wants to conduct more real-world experiments to ensure it's safe to use on the roads.

Founded in 2017, Wayve’s backers include Microsoft, online supermarket Ocado, Virgin Group founder Sir Richard Branson and the inventor of the internet Sir Tim Berners-Lee.

Related:China Reveals Self-Driving Vehicles Safety Guidelines

This article first appeared in IoT World Today's sister site, AI Business.

About the Author(s)

Ben Wodecki

Junior Editor - AI Business

Ben Wodecki is the junior editor of AI Business, covering a wide range of AI content. Ben joined the team in March 2021 as assistant editor and was promoted to junior editor. He has written for The New Statesman, Intellectual Property Magazine, and The Telegraph India, among others.

Sign Up for the Newsletter
The most up-to-date news and insights into the latest emerging technologies ... delivered right to your inbox!

You May Also Like