Controlling Drones with the Human Brain
A researcher from Arizona State University wants to command machines with the human brain. In fact, within ten years, Panagiotis Artemiadis, professor of mechanical and aerospace engineering at Arizona State University, envisions a swarm of brain-controlled drones playing a critical role in a range of applications, including agriculture and search-and-rescue operations.
The essential technology component of mind-controlled UAVs, or unmanned aerial vehicles, is the interface: a combination of hardware and algorithms that maps one’s brain activation to commands for a robotic system. Using electrodes placed on the scalp to record electrical activity, the non-invasive Electroencephalography (EEG) method measures voltage fluctuations resulting from the neurons’ ionic current. The hardware then measures brain activation as the subject thinks about an intended motion for the machine. The algorithms decode those activations to control commands for the robotic system.
At ASU’s Human-Oriented Robotics and Control (HORC) Lab, Artemiadis and his fellow researchers have evaluated the brain's capacity to direct non-human behaviors, including the flight of drones. In 2016, Artemiadis tested the ability of a single operator to control three UAVs, guiding each through a narrow loop. The subject used a hybrid brain-machine interface, combining EEG-recorded brain activity with input from a joystick. To control the multiple drones, the user had to think about shrinking or expanding the shape of the UAV grouping as a whole. By developing algorithms that extracted specific information from the brain, the experiment demonstrated the ability for humans to steer not just a single machine, but a swarm.
Collective behaviors are abundant in nature, but the idea of a flock or herd has only recently been adopted to robotic concepts. A robotic “swarm” system consists of a large group of interchangeable vehicles that use information, obtained via local sensing and communication, to execute autonomous decisions. Artemiadis and his team tested a “human-swarm interface” to, in effect, decode the brain and extract information related to desired collective behaviors, like the expansion of a swarm’s coverage or the implementation of a specific formation.
The algorithms developed at the HORC lab extract brain signals. The ASU team applies machine learning and pattern recognition algorithms to translate the brain’s electrical signals into control commands that are wirelessly transmitted to the drones.
The integration of very large teams of robots into comprehensive systems enables new tasks and missions, including search, exploration, and surveillance, according to the ASU researcher. Armed with infrared imaging equipment, a drone swarm, for example, could provide real-time tracking of a forest fire, allowing responders to adjust their plans on the fly. Teams of drones could also be called on to create topographic maps for soil analysis and irrigation planning.
Top Stories
INSIDERManned Systems
How Airbus is Using w-DED to 3D Print Larger Titanium Airplane Parts
INSIDERAerospace
FAA to Replace Aging Network of Ground-Based Radars
NewsUnmanned Systems
Microvision Aquires Luminar, Plans Relationship Restoration, Multi-industry Push
NewsDesign
CES 2026: Bosch is Ready to Bring AI to Your (Likely ICE-powered) Vehicle
NewsSoftware
Accelerating Down the Road to Autonomy
INSIDERManned Systems
Can This Self-Healing Composite Make Airplane and Spacecraft Components Last...
Webcasts
Power
Hydrogen Engines Are Heating Up for Heavy Duty
Transportation
Advantages of Smart Power Distribution Unit Design for Automotive...
Automotive
Quiet, Please: NVH Improvement Opportunities in the Early Design...
Power
A FREE Two-Day Event Dedicated to Connected Mobility
Connectivity
E/E Architecture Redefined: Building Smarter, Safer, and Scalable Vehicles
Manufacturing & Prototyping
How Sift's Unified Observability Platform Accelerates Drone Innovation



