This website uses cookies primarily for visitor analytics. Certain pages will ask you to fill in contact details to receive additional information. On these pages you have the option of having the site log your details for future visits. Indicating you want the site to remember your details will place a cookie on your device. To view our full cookie policy, please click here. You can also view it at any time by going to our Contact Us page.

‘ANYmal’ robot stalks dark sewers to test its navigation

10 January 2019

Researchers are working to make sure a seeing, hearing, door-opening robot called “ANYmal” can also function in extreme conditions.

ANYmal on its first tour of inspection in Zurich’s sewer system (Credit: Daniel Winkler/ETH Zurich)

Their aim is to determine whether ANYmal – a robot that ANYbotics, an ETH Zurich spin-off company, jointly developed with Robotic Systems Lab – could one day be useful in sewer systems. It might, for instance, help city of Zurich employees who regularly have to walk or crawl through the some 62 miles of accessible shafts and drains underneath the city and whose job it is to check the walls and floors for damage.

This work not only poses a health risk, but is also potentially lethal, given that the drains can fill up with water very quickly without warning. Another advantage of robots in such an environment is that they could operate in narrow sewers that are not accessible with current technology.

Going down

The researchers place the robot upright at the bottom of the shaft. It is about 19.5 inches tall and has four articulated legs as well as something resembling a head that consists of a camera and various sensors.

Peter Fankhauser, cofounder of ANYbotics, radios his colleagues on the surface, who are responsible for coordinating the test and sending commands to the robot. Fankhauser then twiddles with a joystick and the robot plods forward. As this is the first test in unknown terrain, he takes partial control of the robot even though it is capable of moving autonomously.

“It’s a precautionary measure,” says Fankhauser, “Just because something works in the lab doesn’t always mean it will in the real world.”

After all, the conditions underground are not what the robot is used to: the chamber is wet and slippery, with lower temperatures and higher humidity than in the lab. What’s more, it’s very, very dark.

“It’s hard to distinguish much down here,” says Fankhauser, almost with a hint of resignation in his voice, as the robot moves at a slow pace through the roughly 10-foot by 16-foot tunnel.

The robot emits a uniform electromechanical sound – a kind of rhythmic whirring – that blends with the sound of rushing water emanating from the main sewer nearby. Given that the robot is on its maiden test run 13 feet below ground ¬level, the researchers have taken the precaution of avoiding large volumes of water.

In the dark

The goal of the three-year research project, called THING (sub-Terranean Haptic InvestiGator), is to design robots that can move about on their own and are better able to identify their surroundings.

(Credit: Daniel Winkler/ETH Zurich)

Robots generally use 3D cameras and laser sensors for orientation. But such devices can malfunction in adverse conditions – such as when the ground surface is wet or the air full of dust. That’s why the researchers consider enhanced haptic perception – orientation by touch – to be a possible solution.

Researchers at other institutions are also working with and testing ANYmal. In addition to the tests in the sewerage system, next year researchers will deploy the robot in a Polish copper mine. That will determine whether it can function in an entirely different microclimate, one characterised by hot, dusty air and gravel surfaces.

One of the key questions on this first day of testing is whether the robot can find its way around at all in the darkness of the sewerage system. Initially, two helpers with big LED lamps illuminate the surroundings so that the researchers can clearly see what’s going on.

Then, Fankhauser asks the helpers to turn off the lamps and radios his colleagues on the surface to tell the robot to use its own lights. The robot’s sense of touch isn’t the only thing that helps it find its way in the dark, as Marco Hutter, who has been conducting research into legged robots for many years at ETH Zurich, explains: “The robot uses laser sensors and cameras to scan its surroundings. By identifying irregularities in the surface of the concrete, it can determine where it is at any given moment.”

All that the researchers can see in the darkness now are the small round LEDs in the robot’s “head.” The atmosphere is other-worldly: the darkness, the sound of rushing water, the electromechanical whirring, the robot’s LED eyes. Then someone breaks the eerie silence with a droll comment: “Its eyes are a bit like a Rottweiler.”

ANYmal’s future missions

Researchers completed the first ANYmal prototype in 2015. They carry out on-site tests two or three times a month. For instance, Fankhauser and some members of his team recently headed to an offshore platform in the middle of the North Sea. The hope is that robots could one day perform inspections on such platforms. On its pilot run at least, ANYmal autonomously completed several inspection routes with flying colours.

After almost ten years of research, there’s a lot ANYmal can do. It can not only walk autonomously, but also boasts the sensory capabilities of sight, hearing, and touch. These enable it, for instance, to read the air pressure display on a machine, identify sounds, and recognise objects – for example to determine whether or not a fire extinguisher is in the right place. The robot can even perform certain manual tasks on its own.

Equipped with an additional gripping arm, it can open doors, dispose of refuse, or press a lift button. It also delivers data that is more precise than our own eyes, ears, and noses can perceive. It can identify the ambient temperature and detect the presence of gases in the air. Its latest trick is recognising the composition of the ground beneath it. “Some of its powers are superhuman,” says Fankhauser.

Despite the lack of light in the sewer, the robot seems to be finding its way quite well, plodding through the shallow channel at a leisurely pace. When the high-tech machine reaches a 20cm high ledge (just under 8 inches) in a dry side arm of the sewer, Fankhauser brings it to a halt with a flick of the joystick. Initially, he is reluctant to give the ¬robot the command to climb over the ledge. Although it has easily mastered this manoeuvre in laboratory conditions, down here it is a risky undertaking.

“It’s an expensive machine,” says Fankhauser. But he gives it a try anyway. ANYmal doesn’t manage it at its first attempt. It stops at the ledge like a horse balking at a jump. “Default, start again,” radios Fankhauser. Now the robot elegantly places one leg after another over the ledge.

Watching from the surface

Sitting on a bench under a white canopy on the surface, two assistants have their eyes firmly fixed on a laptop. A generator is buzzing and a router is blinking – and many a cyclist passes by along the main road looking on in bemusement at the hubbub around the open manhole at the side of the road.

Looking over the researchers’ shoulders, you can see an almost constant stream of data flickering across the screen. And thanks to state-of-the-art 3D and laser technologies, live images the robot constantly transmits from underground are visible on a separate monitor.

When Fankhauser radios from below that he wants the robot to touch the wall of the sewer with one of its legs, the two assistants have their work cut out for them. The software they are using hasn’t been programmed for this. They respond quickly, however, taking an algorithm originally programmed to teach ANYmal to shake hands. But to make sure the robot doesn’t hit the wall with force, the researchers have to adapt the parameters.

In this case, the problem is the angle at which the robot is to raise its leg. One of the assistants types in 100 and then gradually ratchets up the number. At 180 the perfect level is reached and the robot’s manoeuvre is successful.

“The robot was in nonstop operation and collected a lot of data,” says Fankhauser as he undoes his high rubber boots and removes his protective clothing.

Hutter is satisfied, too: “All the teams will be taking home a huge volume of data to incorporate in their research.”

They are now one step closer to their goal of delivering a robot that can function properly in challenging conditions underground. But their work is far from finished. The robot recorded 500,000 measurements per second over the course of the day. “That’s enough data to keep us busy for six months,” says Fankhauser with a laugh.

Hutter received support from ETH soon after embarking on this research in the form of an ESOP scholarship and a Pioneer fellowship.

Source: ETH Zurich

The original article can be found on Futurity 


Print this page | E-mail this page