The visual language could be used in little to no-access areas, such as space or search-and-rescue missions

Scarlett Evans, Assistant Editor, IoT World Today

July 11, 2022

2 Min Read

The ‘waggle dance’ used by bees to communicate with one another where pollen is located is the inspiration for a new motion-based robotic language.

A study published this month in Frontiers in Robotics and AI demonstrated the successful use of this unusual method of communication for robot-to-robot interaction, allowing them to communicate with one another without requiring a network connection. 

Bee colonies can communicate the location of nearby flowers in both distance and direction by wiggling their behinds in the direction of the pollen source. The length of the dance indicates the distance to the source, while the direction of the dance indicates the flowers’ location. 

To test this method for robotic communication, researchers from the University of Maryland and the Indian Institute of Science had two warehouse robots arrange the movement of a package interacting with only vision-based cues.

A human worker initially gave instructions to one robot using hand gestures, which the robot interpreted using cameras and skeletal tracking algorithms. This bot then passed along the details to a handling robot by tracing a symbol on the ground; with the symbol’s shape representing the direction of travel, and the time it takes to draw the shape representing the distance to the package’s destination. 


A diagram of how bees’ movements show direction and distance to pollen source.
Credit: Frontiers in Robotics and AI

This novel method was shown to have a 93.3% accuracy in simulations and a 90% accuracy in physical practice. 

Such a method could be useful in remote areas where network connection is intermittent or unreliable, in search-and-rescue missions where networks are damaged or destroyed, or even in space missions where networks are not yet established. 

As the technique relies solely on a simple camera, it can be deployed on any kind of robot and is both scalable and customizable, making its real-world applications limitless, according to the researchers.

Next, the team will be looking to fine-tune the method, increasing the robots’ ability to recognize different human gestures as well as the varying intensities of human gestures. 

About the Author(s)

Scarlett Evans

Assistant Editor, IoT World Today

Scarlett Evans is the assistant editor for IoT World Today, with a particular focus on robotics and smart city technologies. Scarlett has previous experience in minerals and resources with Mine Australia, Mine Technology and Power Technology. She joined Informa in April 2022.

Sign Up for the Newsletter
The most up-to-date news and insights into the latest emerging technologies ... delivered right to your inbox!

You May Also Like