Researchers use large language models to help robots navigate
The method uses language-based inputs instead of costly visual data to direct a robot through a multistep navigation task.
The method uses language-based inputs instead of costly visual data to direct a robot through a multistep navigation task.
DenseAV, developed at MIT, learns to parse and understand the meaning of language just by watching videos of people talking, with potential applications in multimedia search, language learning, and robotics.
With generative AI models, researchers combined robotics data from different sources to help robots learn better.
A new approach could streamline virtual training processes or aid clinicians in reviewing diagnostic videos.
“Alchemist” system adjusts the material attributes of specific objects within images to potentially modify video game models to fit different environments, fine-tune VFX, and diversify robotic training.
Fifteen new faculty members join six of the school’s academic departments.
A new technique that can automatically classify phases of physical systems could help scientists investigate novel materials.
A new “consensus game,” developed by MIT CSAIL researchers, elevates AI’s text comprehension and generation skills.
A new algorithm learns to squish, bend, or stretch a robot’s entire body to accomplish diverse tasks like avoiding obstacles or retrieving items.
Associate Professor Jonathan Ragan-Kelley optimizes how computer graphics and images are processed for the hardware of today and tomorrow.