Submitted by L. Millard on Sat, 23/01/2016 - 18:20
Research led by CSIC Co-Investigator, Professor Roberto Cipolla, is developing two technologies that use deep learning techniques to help machines to see and recognise their location and surroundings.
Image courtesy of University of Cambridge Visual Localisation
The technologies, which can be used on a regular camera or smartphone, could be used for the development of driverless cars and autonomous robotics.
The separate but complementary systems have been designed by researchers from the Department of Engineering and demonstrations are freely available online.
Although the systems cannot currently control a driverless car, the ability to make a machine ‘see’ and accurately identify where it is and what it’s looking at is a vital part of developing autonomous vehicles and robotics.
“Vision is our most powerful sense and driverless cars will also need to see but teaching a machine to see is far more difficult than it sounds,” said Professor Cipolla.
“Work in the field of artificial intelligence and robotics has really taken off in the past few years,” said Alex Kendall, a PhD student in the Department of Engineering. “But what’s cool about our group is that we’ve developed technology that uses deep learning to determine where you are and what’s around you – this is the first time this has been done using deep learning.”
Read the full article, which originally appeared on the University of Cambridge and Department of Engineering websites, here.