• Skip to primary navigation
  • Skip to main content
  • Skip to primary sidebar
  • Skip to secondary sidebar
  • About
    • Contact
    • Privacy
    • Terms of use
  • Advertise
    • Advertising
    • Case studies
    • Design
    • Email marketing
    • Features list
    • Lead generation
    • Magazine
    • Press releases
    • Publishing
    • Sponsor an article
    • Webcasting
    • Webinars
    • White papers
    • Writing
  • Subscribe to Newsletter

Robotics & Automation News

Where Innovation Meets Imagination

  • Home
  • News
  • Features
  • Editorial Sections A-Z
    • Agriculture
    • Aircraft
    • Artificial Intelligence
    • Automation
    • Autonomous Vehicles
    • Business
    • Computing
    • Construction
    • Culture
    • Design
    • Drones
    • Economy
    • Energy
    • Engineering
    • Environment
    • Health
    • Humanoids
    • Industrial robots
    • Industry
    • Infrastructure
    • Investments
    • Logistics
    • Manufacturing
    • Marine
    • Material handling
    • Materials
    • Mining
    • Promoted
    • Research
    • Robotics
    • Science
    • Sensors
    • Service robots
    • Software
    • Space
    • Technology
    • Transportation
    • Warehouse robots
    • Wearables
  • Press releases
  • Events

Can robots learn from ‘machine dreams’?

November 12, 2024 by Mark Allinson

Scientists from MIT’s Computer Science and Artificial Intelligence Laboratory used AI-generated images to train a robot dog in parkour, without real-world data.

Parkour is a non-competitive physical discipline where participants move through an environment using their bodies to overcome obstacles – basically like navigating an obstacle course.

CSAIL’s LucidSim system demonstrates the potential of GenAI – generative AI, or generative AI – for creating relevant robotics training data, enabling expert-level performance in complex tasks like obstacle navigation and stair climbing.

For roboticists, one challenge towers above all others: generalization – the ability to create machines that can adapt to any environment or condition. Since the 1970s, the field has evolved from writing sophisticated programs to using deep learning, teaching robots to learn directly from human behavior.

But a critical bottleneck remains: data quality. To improve, robots need to encounter scenarios that push the boundaries of their capabilities, operating at the edge of their mastery.

This process traditionally requires human oversight, with operators carefully challenging robots to expand their abilities. As robots become more sophisticated, this hands-on approach hits a scaling problem: the demand for high-quality training data far outpaces humans’ ability to provide it.

Now, a team of MIT CSAIL researchers have developed a novel approach to robot training that could significantly accelerate the deployment of adaptable, intelligent machines in real-world environments.

The new system, called “LucidSim”, uses recent advances in generative AI and physics simulators to create diverse and realistic virtual training environments, helping robots achieve expert-level performance in difficult tasks without any real-world data.

LucidSim combines physics simulation with generative AI models, addressing one of the most persistent challenges in robotics: transferring skills learned in simulation to the real world.

MIT CSAIL postdoctoral associate Ge Yang, a lead researcher on LucidSim, says: “A fundamental challenge in robot learning has long been the ‘sim-to-real gap’ – the disparity between simulated training environments and the complex, unpredictable real world.

“Previous approaches often relied on depth sensors, which simplified the problem but missed crucial real-world complexities.”

The multipronged system is a blend of different technologies. At its core, LucidSim uses large language models to generate various structured descriptions of environments.

These descriptions are then transformed into images using generative models. To ensure that these images reflect real-world physics, an underlying physics simulator is used to guide the generation process.

The birth of an idea: from burritos to breakthroughs

The inspiration for LucidSim came from an unexpected place: a conversation outside Beantown Taqueria in Cambridge.

Alan Yu, an undergraduate student at MIT and co-lead on LucidSim, says: “We wanted to teach vision-equipped robots how to improve using human feedback. But then, we realized we didn’t have a pure vision-based policy to begin with.

“We kept talking about it as we walked down the street, and then we stopped outside the taqueria for about half an hour. That’s where we had our moment.”

To cook up their data, the team generated realistic images by extracting depth maps, which provide geometric information, and semantic masks, which label different parts of an image, from the simulated scene.

They quickly realized, however, that with tight control on the composition of the image content, the model would produce similar images that weren’t different from each other using the same prompt. So, they devised a way to source diverse text prompts from ChatGPT.

This approach, however, only resulted in a single image. To make short, coherent videos which serve as little “experiences” for the robot, the scientists hacked together some image magic into another novel technique the team created, called “Dreams In Motion (DIM)”.

The system computes the movements of each pixel between frames, to warp a single generated image into a short, multi-frame video. Dreams In Motion does this by considering the 3D geometry of the scene and the relative changes in the robot’s perspective.

Yu says: “We outperform domain randomization, a method developed in 2017 that applies random colors and patterns to objects in the environment, which is still considered the go-to method these days.

“While this technique generates diverse data, it lacks realism. LucidSim addresses both diversity and realism problems. It’s exciting that even without seeing the real world during training, the robot can recognize and navigate obstacles in real environments.”

The team is particularly excited about the potential of applying LucidSim to domains outside quadruped locomotion and parkour, their main testbed. One example is mobile manipulation, where a mobile robot is tasked to handle objects in an open area, and also, color perception is critical.

Yang says: “Today, these robots still learn from real-world demonstrations. Although collecting demonstrations is easy, scaling a real-world robot teleoperation setup to thousands of skills is challenging because a human has to physically set up each scene.

“We hope to make this easier, thus qualitatively more scalable, by moving data collection into a virtual environment.”

The team put LucidSim to the test against an alternative, where an expert teacher demonstrates the skill for the robot to learn from.

The results were surprising: robots trained by the expert struggled, succeeding only 15 percent of the time – and even quadrupling the amount of expert training data barely moved the needle.

But when robots collected their own training data through LucidSim, the story changed dramatically. Just doubling the dataset size catapulted success rates to 88 percent.

Yang says: “And giving our robot more data monotonically improves its performance – eventually, the student becomes the expert.”

Stanford University Assistant Professor of Electrical Engineering Shuran Song, who wasn’t involved in the research, says: “One of the main challenges in sim-to-real transfer for robotics is achieving visual realism in simulated environments.

“The LucidSim framework provides an elegant solution by using generative models to create diverse, highly realistic visual data for any simulation. This work could significantly accelerate the deployment of robots trained in virtual environments to real-world tasks.”

From the streets of Cambridge to the cutting edge of robotics research, LucidSim is paving the way toward a new generation of intelligent, adaptable machines – ones that learn to navigate our complex world without ever setting foot in it.

Yu and Yang wrote the paper with four fellow CSAIL affiliates: mechanical engineering postdoc Ran Choi; undergraduate researcher Yajvan Ravan; John Leonard, Samuel C. Collins Professor of Mechanical and Ocean Engineering in the MIT Department of Mechanical Engineering; and MIT Associate Professor Phillip Isola.

Their work was supported, in part, by a Packard Fellowship, a Sloan Research Fellowship, the Office of Naval Research, Singapore’s Defence Science and Technology Agency, Amazon, MIT Lincoln Laboratory, and the National Science Foundation Institute for Artificial Intelligence and Fundamental Interactions.

The researchers presented their work at the recent Conference on Robot Learning (CoRL).

Print Friendly, PDF & Email

Share this:

  • Click to print (Opens in new window) Print
  • Click to share on Facebook (Opens in new window) Facebook
  • Click to share on LinkedIn (Opens in new window) LinkedIn
  • Click to share on Reddit (Opens in new window) Reddit
  • Click to share on X (Opens in new window) X
  • Click to share on Tumblr (Opens in new window) Tumblr
  • Click to share on Pinterest (Opens in new window) Pinterest
  • Click to share on WhatsApp (Opens in new window) WhatsApp
  • Click to share on Telegram (Opens in new window) Telegram
  • Click to share on Pocket (Opens in new window) Pocket

Related stories you might also like…

Filed Under: Features, Science Tagged With: ai, csail, dreams, genai, generative, images, machine, mit, parkour, robots

Primary Sidebar

Search this website

Latest articles

  • Neura Robotics and HD Hyundai partner to bring humanoid robots to shipbuilding
  • Exclusive interview with Vorago Technologies: The risks of failure in home robots
  • MIT engineers design new handheld interface that gives ‘anyone the ability to train a robot’
  • Glīd launches ‘world-first’ unmanned autonomous road-to-rail dual mobility platform
  • A robot renaissance with a human focus: Hexagon’s 2035 vision for intelligent autonomy
  • Automotive testing UTAC embarks on global expansion and upgrade program
  • The evolving role of the human worker alongside cobots: Upskilling and new job creation
  • Enterprise Sales Courses That Help Close Bigger Deals
  • Mitsubishi Electric says its robots are ‘bridging skills gaps’ in automation technology
  • Pusan National University scientists develop ‘game-changing method’ to create safer, long-lasting lithium-ion batteries

Secondary Sidebar

Copyright © 2025 · News Pro on Genesis Framework · WordPress · Log in

We use cookies on our website to give you the most relevant experience by remembering your preferences and repeat visits. By clicking “Accept”, you consent to the use of ALL the cookies.
Do not sell my personal information.
Cookie SettingsAccept
Manage consent

Privacy Overview

This website uses cookies to improve your experience while you navigate through the website. Out of these, the cookies that are categorized as necessary are stored on your browser as they are essential for the working of basic functionalities of the website. We also use third-party cookies that help us analyze and understand how you use this website. These cookies will be stored in your browser only with your consent. You also have the option to opt-out of these cookies. But opting out of some of these cookies may affect your browsing experience.
Necessary
Always Enabled
Necessary cookies are absolutely essential for the website to function properly. These cookies ensure basic functionalities and security features of the website, anonymously.
CookieDurationDescription
cookielawinfo-checkbox-analytics11 monthsThis cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Analytics".
cookielawinfo-checkbox-functional11 monthsThe cookie is set by GDPR cookie consent to record the user consent for the cookies in the category "Functional".
cookielawinfo-checkbox-necessary11 monthsThis cookie is set by GDPR Cookie Consent plugin. The cookies is used to store the user consent for the cookies in the category "Necessary".
cookielawinfo-checkbox-others11 monthsThis cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Other.
cookielawinfo-checkbox-performance11 monthsThis cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Performance".
viewed_cookie_policy11 monthsThe cookie is set by the GDPR Cookie Consent plugin and is used to store whether or not user has consented to the use of cookies. It does not store any personal data.
Functional
Functional cookies help to perform certain functionalities like sharing the content of the website on social media platforms, collect feedbacks, and other third-party features.
Performance
Performance cookies are used to understand and analyze the key performance indexes of the website which helps in delivering a better user experience for the visitors.
Analytics
Analytical cookies are used to understand how visitors interact with the website. These cookies help provide information on metrics the number of visitors, bounce rate, traffic source, etc.
Advertisement
Advertisement cookies are used to provide visitors with relevant ads and marketing campaigns. These cookies track visitors across websites and collect information to provide customized ads.
Others
Other uncategorized cookies are those that are being analyzed and have not been classified into a category as yet.
SAVE & ACCEPT