Drone Behavior Team

What We Do

We are responsible for planning the drone's movements and actions when searching a farm for deer and when deterring deer once they are detected. Patrols (or flight paths) are waypoint missions that are assigned to each drone by the base computer. To guarantee that the entire geofenced area, with the exception of no-fly zones, is scanned for deer, we overlay a grid across the area with grid cell sizes slightly smaller than the drone camera's ground footprint, then place a waypoint at the center of each grid cell.

Energy-Efficient Coverage Path Planning

Once waypoints are assigned to drones, our algorithm decides the order in which each drone will visit its waypoints to minimize the drone's energy expenditure during the mission. Our algorithm generates energy-efficient coverage paths that account for the energy costs of straight-line flight, turns, and flying in wind.

Energy-efficient flight paths for two drones on Lantern Farm and Saint Paul Farm.
Energy-efficient flight paths for two drones on Lantern Farm (left) and Saint Paul Farm (right), optimized for conditions with no wind using our energy-efficient coverage path planning algorithm. Waypoints were assigned to drones according to our load balancing approach.
GTSP formulation of the energy-efficient coverage path planning algorithm.
The energy-efficient coverage path planning algorithm formulated as a generalized traveling salesman (GTSP) problem. Each cluster contains GTSP nodes which represent flying from the original node to each of its adjacent nodes, allowing the path cost to model the energy of straight-line flight, turns, and wind.

Deer Shepherding

Because the shepherding drones only know where deer are located if they are within the drones' line of sight, we need a way to represent their belief states—what they believe to be true about the world. When a deer is detected in a grid cell, the value of that cell is set to the probability of the computer vision model being correct. If the drones look away from a grid cell where they had detected a deer, then the deer probability value assigned to that grid cell diffuses slightly to the neighboring cells for every time step that the cell is unobserved. This embeds memory into their belief state so that they remember that an area where a deer was recently observed has some probability to still contain a deer.

Our approach uses multi-agent Deep-Q learning to train the shepherding policy used by all drones. We use a centralized learning architecture in which all shepherding drones update a single learned policy during training, with fully decentralized execution where each drone has its own copy of the policy.

A simulation rollout of drones shepherding deer.
A simulation rollout of drones shepherding deer.

Shepherding Algorithm Training Status

Training

The shepherding algorithm is still training and improving. Shepherding training metrics across 1.142 billion environmental steps are shown below. The vertical dotted lines indicate when we interrupted training to adjust parameters and reward weights before resuming where it left off.

Shepherding training metrics across 1.142 billion environmental steps.

If you're interested in learning the details of our implementation, please get in touch with us!

Note: The paper below is from the previous year. We are working on our next publication.

UAV Coverage Path Planning for Autonomous Deer Detection in the Wild Driven by Ant Colony Optimization

Workshop: Robots in the Wild , ICRA 2025

Meet the Team

Mario Jerez

Mario Jerez

PhD Student

Computer Science

Ebasa Temesgen

Ebasa Temesgen

PhD Candidate

Computer Science

Greta Brown

Greta Brown

Undergraduate

Computer Science

Khoi Duong

Khoi Duong

Undergraduate

Computer Science

Serena Bovee

Serena Bovee

Undergraduate

Sociology