The group’s inspiration is the pigeon. Hardy, plentiful and receptive to training, the birds are easy to study. In flight, they estimate the distance between themselves and objects ahead by quickly processing blurry, low-resolution images, just as a drone will need to do. And, crucially, they have a tendency to make decisions at the last moment—within five feet of an obstacle.
The first step is to teach robots to differentiate between obstacles and empty space. Engineers have already figured out how to train point-and-shoot cameras to spot faces in a photo: In a process called supervised learning, a technician feeds millions of images into a computer and tells it to output a “1” when the image contains a human face and a “0” when it does not. But this style of supervised learning would be an impossibly labor-intensive way to train a drone. A human would have to label not just faces but every possible object the robot might encounter. Instead, Yann LeCun, a professor of computer and neural science at New York University who leads the drone’s vision team, is developing software that will allow the drone to draw conclusions about what it’s seeing with much less human coaching.