Currently robots are very limited in what they can do. Their inability to understand the nuances of human language makes them mostly useless for more complicated requests.
For example, if you put a specific tool in a toolbox and ask a robot to "pick it up," it would be completely lost.
Picking it up means being able to see and identify objects, understand commands, recognise that the "it" in question is the tool you put down, go back in time to remember the moment when you put down the tool, and distinguish the tool you put down from other ones of similar shapes and sizes.
They have developed an Alexa-like system "ComText" - for "commands in context" - that allows robots to understand a wide range of commands that require contextual knowledge about objects and their environments.
"Where humans understand the world as a collection of objects and people and abstract concepts, machines view it as pixels, point-clouds, and 3D maps generated from sensors," said Rohan Paul,
"This semantic gap means that, for robots to understand what we want them to do, they need a much richer representation of what we do and say," Paul said.
ComText can observe a range of visuals and natural language to learm about an object's size, shape, position, type and even if it belongs to somebody.
From this knowledge base, it can then reason, infer meaning and respond to commands.
"The main contribution is this idea that robots should have different kinds of memory, just like people," said Barbu.
With ComText, Baxter was successful in executing the right command about 90 per cent of the time.
In the future, the team hopes to enable robots to understand more complicated information, such as multi-step commands, the intent of actions, and using properties about objects to interact with them more naturally.
"This work is a nice step towards building robots that can interact much more naturally with people," said Luke Zettlemoyer, an associate professor at the University of Washington in the US, who was not involved in the research.
"In particular, it will help robots better understand the names that are used to identify objects in the world, and interpret instructions that use those names to better do what users ask," Zettlemoyer said.
Disclaimer: No Business Standard Journalist was involved in creation of this content
You’ve reached your limit of {{free_limit}} free articles this month.
Subscribe now for unlimited access.
Already subscribed? Log in
Subscribe to read the full story →
Smart Quarterly
₹900
3 Months
₹300/Month
Smart Essential
₹2,700
1 Year
₹225/Month
Super Saver
₹3,900
2 Years
₹162/Month
Renews automatically, cancel anytime
Here’s what’s included in our digital subscription plans
Exclusive premium stories online
Over 30 premium stories daily, handpicked by our editors


Complimentary Access to The New York Times
News, Games, Cooking, Audio, Wirecutter & The Athletic
Business Standard Epaper
Digital replica of our daily newspaper — with options to read, save, and share


Curated Newsletters
Insights on markets, finance, politics, tech, and more delivered to your inbox
Market Analysis & Investment Insights
In-depth market analysis & insights with access to The Smart Investor


Archives
Repository of articles and publications dating back to 1997
Ad-free Reading
Uninterrupted reading experience with no advertisements


Seamless Access Across All Devices
Access Business Standard across devices — mobile, tablet, or PC, via web or app
