Helping robots zero in on the objects that matter

-

Imagine having to straighten up a messy kitchen, starting with a counter suffering from sauce packets. In case your goal is to wipe the counter clean, you would possibly sweep up the packets as a bunch. If, nevertheless, you desired to first select the mustard packets before throwing the remainder away, you’d sort more discriminately, by sauce type. And if, among the many mustards, you had a hankering for Grey Poupon, finding this specific brand would entail a more careful search.

MIT engineers have developed a way that permits robots to make similarly intuitive, task-relevant decisions.

The team’s recent approach, named Clio, enables a robot to discover the parts of a scene that matter, given the tasks at hand. With Clio, a robot takes in an inventory of tasks described in natural language and, based on those tasks, it then determines the extent of granularity required to interpret its surroundings and “remember” only the parts of a scene which might be relevant.

In real experiments starting from a cluttered cubicle to a five-story constructing on MIT’s campus, the team used Clio to robotically segment a scene at different levels of granularity, based on a set of tasks laid out in natural-language prompts reminiscent of “move rack of magazines” and “get first aid kit.”

The team also ran Clio in real-time on a quadruped robot. Because the robot explored an office constructing, Clio identified and mapped only those parts of the scene that related to the robot’s tasks (reminiscent of retrieving a dog toy while ignoring piles of office supplies), allowing the robot to know the objects of interest.

Clio is known as after the Greek muse of history, for its ability to discover and remember only the weather that matter for a given task. The researchers envision that Clio could be useful in lots of situations and environments by which a robot would must quickly survey and make sense of its surroundings within the context of its given task.

“Search and rescue is the motivating application for this work, but Clio can even power domestic robots and robots working on a factory floor alongside humans,” says Luca Carlone, associate professor in MIT’s Department of Aeronautics and Astronautics (AeroAstro), principal investigator within the Laboratory for Information and Decision Systems (LIDS), and director of the MIT SPARK Laboratory. “It’s really about helping the robot understand the environment and what it has to recollect with a view to perform its mission.”

The team details their ends in a study appearing today within the journal . Carlone’s co-authors include members of the SPARK Lab: Dominic Maggio, Yun Chang, Nathan Hughes, and Lukas Schmid; and members of MIT Lincoln Laboratory: Matthew Trang, Dan Griffith, Carlyn Dougherty, and Eric Cristofalo.

Open fields

Huge advances within the fields of computer vision and natural language processing have enabled robots to discover objects of their surroundings. But until recently, robots were only capable of accomplish that in “closed-set” scenarios, where they’re programmed to work in a fastidiously curated and controlled environment, with a finite variety of objects that the robot has been pretrained to acknowledge.

In recent times, researchers have taken a more “open” approach to enable robots to acknowledge objects in additional realistic settings. In the sector of open-set recognition, researchers have leveraged deep-learning tools to construct neural networks that may process billions of images from the web, together with each image’s associated text (reminiscent of a friend’s Facebook picture of a dog, captioned “Meet my recent puppy!”).

From thousands and thousands of image-text pairs, a neural network learns from, then identifies, those segments in a scene which might be characteristic of certain terms, reminiscent of a dog. A robot can then apply that neural network to identify a dog in a very recent scene.

But a challenge still stays as to learn how to parse a scene in a useful way that’s relevant for a selected task.

“Typical methods will pick some arbitrary, fixed level of granularity for determining learn how to fuse segments of a scene into what you’ll be able to consider as one ‘object,’” Maggio says. “Nonetheless, the granularity of what you call an ‘object’ is definitely related to what the robot has to do. If that granularity is fixed without considering the tasks, then the robot may find yourself with a map that isn’t useful for its tasks.”

Information bottleneck

With Clio, the MIT team aimed to enable robots to interpret their surroundings with a level of granularity that might be robotically tuned to the tasks at hand.

For example, given a task of moving a stack of books to a shelf, the robot should have the option to  determine that your complete stack of books is the task-relevant object. Likewise, if the duty were to maneuver only the green book from the remainder of the stack, the robot should distinguish the green book as a single goal object and disrespect the remainder of the scene — including the opposite books within the stack.

The team’s approach combines state-of-the-art computer vision and enormous language models comprising neural networks that make connections amongst thousands and thousands of open-source images and semantic text. Additionally they incorporate mapping tools that robotically split a picture into many small segments, which might be fed into the neural network to find out if certain segments are semantically similar. The researchers then leverage an idea from classic information theory called the “information bottleneck,” which they use to compress numerous image segments in a way that picks out and stores segments which might be semantically most relevant to a given task.

“For instance, say there’s a pile of books within the scene and my task is simply to get the green book. In that case we push all this information concerning the scene through this bottleneck and find yourself with a cluster of segments that represent the green book,” Maggio explains. “All the opposite segments that are usually not relevant just get grouped in a cluster which we are able to simply remove. And we’re left with an object at the correct granularity that is required to support my task.”

The researchers demonstrated Clio in numerous real-world environments.

“What we thought could be a extremely no-nonsense experiment could be to run Clio in my apartment, where I didn’t do any cleansing beforehand,” Maggio says.

The team drew up an inventory of natural-language tasks, reminiscent of “move pile of garments” after which applied Clio to pictures of Maggio’s cluttered apartment. In these cases, Clio was capable of quickly segment scenes of the apartment and feed the segments through the Information Bottleneck algorithm to discover those segments that made up the pile of garments.

Additionally they ran Clio on Boston Dynamic’s quadruped robot, Spot. They gave the robot an inventory of tasks to finish, and because the robot explored and mapped the inside an office constructing, Clio ran in real-time on an on-board computer mounted to Spot, to pick segments within the mapped scenes that visually relate to the given task. The tactic generated an overlaying map showing just the goal objects, which the robot then used to approach the identified objects and physically complete the duty.

“Running Clio in real-time was an enormous accomplishment for the team,” Maggio says. “Quite a lot of prior work can take several hours to run.”

Going forward, the team plans to adapt Clio to have the option to handle higher-level tasks and construct upon recent advances in photorealistic visual scene representations.

“We’re still giving Clio tasks which might be somewhat specific, like ‘find deck of cards,’” Maggio says. “For search and rescue, it’s worthwhile to give it more high-level tasks, like ‘find survivors,’ or ‘get power back on.’ So, we wish to get to a more human-level understanding of learn how to accomplish more complex tasks.”

This research was supported, partially, by the U.S. National Science Foundation, the Swiss National Science Foundation, MIT Lincoln Laboratory, the U.S. Office of Naval Research, and the U.S. Army Research Lab Distributed and Collaborative Intelligent Systems and Technology Collaborative Research Alliance.

ASK ANA

What are your thoughts on this topic?
Let us know in the comments below.

0 0 votes
Article Rating
guest
0 Comments
Oldest
Newest Most Voted
Inline Feedbacks
View all comments

Share this article

Recent posts

0
Would love your thoughts, please comment.x
()
x