This website uses cookies primarily for visitor analytics. Certain pages will ask you to fill in contact details to receive additional information. On these pages you have the option of having the site log your details for future visits. Indicating you want the site to remember your details will place a cookie on your device. To view our full cookie policy, please click here. You can also view it at any time by going to our Contact Us page.

Toilet or chair? Robots that ‘see’ in 3D can tell

20 July 2017

Scientists at Duke University are closer to a new technology that lets machines make sense of 3D objects in a richer and more human-like way.

When fed 3D models of household items in bird's-eye view (left), a new algorithm is able to guess what the objects are. This image shows the guess in the centre, and the actual 3D model on the right. (Credit: Duke University)

Autonomous robots, for example, can inspect nuclear power plants, clean up oil spills in the ocean, accompany fighter planes into combat, and explore the surface of Mars. Yet for all their talents, robots still can’t make a cup of tea.

That’s because tasks such as turning on the stove, fetching the kettle, and finding the milk and sugar require perceptual abilities that, for most machines, are still a fantasy.

Among them is the ability to make sense of 3D objects. While it’s relatively straightforward for robots to “see” objects with cameras and other sensors, interpreting what they see, from a single glimpse, is more difficult.

The most sophisticated robots in the world can’t yet do what most children do automatically, but scientists may be closer to a solution.

A robot that clears dishes off a table, for example, must be able to adapt to an enormous variety of bowls, platters, and plates in different sizes and shapes, left in disarray on a cluttered surface.

Humans can glance at a new object and intuitively know what it is, whether it is right side up, upside down, or sideways, in full view, or partially obscured by other objects. Even when an object is partially hidden, we mentally fill in the parts we can’t see.

The new robot perception algorithm can simultaneously guess what a new object is, and how it’s oriented, without examining it from multiple angles first. It can also “imagine” any parts that are out of view.

A robot with this technology wouldn’t need to see every side of a teapot, for example, to know that it probably has a handle, a lid, and a spout, and whether it is sitting upright or off-kilter on the stove.

The approach makes fewer mistakes and is three times faster than the best current methods.

The design is an important step toward robots that function alongside humans in homes and other real-world settings, which are less orderly and predictable than the highly controlled environment of the lab or the factory floor, says Ben Burchfiel, a graduate student at Duke University.

The framework gives the robot a limited number of training examples and uses them to generalise to new objects.

“It’s impractical to assume a robot has a detailed 3D model of every possible object it might encounter, in advance,” Burchfiel says.

The researchers trained their algorithm on a dataset of roughly 4,000 complete 3D scans of common household objects: an assortment of bathtubs, beds, chairs, desks, dressers, monitors, nightstands, sofas, tables, and toilets.

Each 3D scan was converted into tens of thousands of little cubes, or voxels, stacked on top of each other like LEGO blocks to make them easier to process.

The algorithm learned categories of objects by combing through examples of each one and figuring out how they vary and how they stay the same, using a version of a technique called probabilistic principal component analysis.

When a robot spots something new - say a bunk bed - it doesn’t have to sift through its entire mental catalogue for a match. It learns, from prior examples, what characteristics beds tend to have.

Based on that prior knowledge, it has the power to generalise like a person would - to understand that two objects may be different, yet share properties that make them both a particular type of furniture.

To test the approach, researchers fed the algorithm 908 new 3D examples of the same 10 kinds of household items, viewed from the top.

From this single vantage point, the algorithm correctly guessed what most objects were, and what their overall 3D shapes should be, including the concealed parts, about 75 percent of the time - compared with just over 50 percent for the state-of-the-art alternative.

It was also capable of recognising objects that were rotated in various ways, which the best competing approaches can’t do.

While the system is reasonably fast - the whole process takes about a second - it is still a far cry from human vision, Burchfiel says.

For one, both their algorithm and previous methods were easily fooled by objects that, from certain perspectives, look similar in shape. They might see a table from above, and mistake it for a dresser.

“Overall, we make a mistake a little less than 25 percent of the time, and the best alternative makes a mistake almost half the time, so it is a big improvement,” Burchfiel says “But it still isn’t ready to move into your house. You don’t want it putting a pillow in the dishwasher.”

Now the team is working on scaling up their approach to enable robots to distinguish between thousands of types of objects at a time.

“Researchers have been teaching robots to recognise 3D objects for a while now,” Burchfield says. What’s new, is the ability to both recognise something and fill in the blind spots in its field of vision, to reconstruct the parts it can’t see.

“That has the potential to be invaluable in a lot of robotic applications.”

Source: Duke University

The original article can be found on the Futurity website

Print this page | E-mail this page

Coda Systems