Can robots learn from machine dreams?
MIT CSAIL researchers used AI-generated images to train a robot dog in parkour, without real-world data. Their LucidSim system demonstrates generative AI’s potential for creating robotics training data.
MIT CSAIL researchers used AI-generated images to train a robot dog in parkour, without real-world data. Their LucidSim system demonstrates generative AI’s potential for creating robotics training data.
Inspired by large language models, researchers develop a training technique that pools diverse data to teach robots new skills.
A new method can train a neural network to sort corrupted data while anticipating next steps. It can make flexible plans for robots, generate high-quality video, and help AI agents navigate digital environments.
MIT CSAIL researchers created an AI-powered method for low-discrepancy sampling, which uniformly distributes data points to boost simulation accuracy.
A new method called Clio enables robots to quickly map a scene and identify the items they need to complete a given set of tasks.
New algorithm helps robots practice skills like sweeping and placing objects, potentially helping them improve at important tasks in houses, hospitals, and factories.
CSAIL researchers introduce a novel approach allowing robots to be trained in simulations of scanned home environments, paving the way for customized household automation accessible to anyone.
Neural network controllers provide complex robots with stability guarantees, paving the way for the safer deployment of autonomous vehicles and industrial machines.
The method uses language-based inputs instead of costly visual data to direct a robot through a multistep navigation task.
With generative AI models, researchers combined robotics data from different sources to help robots learn better.