MIT researchers make language models scalable self-learners
The scientists used a natural language-based logical inference dataset to create smaller language models that outperformed much larger counterparts.
The scientists used a natural language-based logical inference dataset to create smaller language models that outperformed much larger counterparts.
A new multimodal technique blends major self-supervised learning methods to learn more similarly to humans.
Researchers develop an algorithm that decides when a “student” machine should follow its teacher, and when it should learn on its own.
Selecting the right method gives users a more accurate picture of how their model is behaving, so they are better equipped to correctly interpret its predictions.
This machine-learning method could assist with robotic scene understanding, image editing, or online recommendation systems.
Senior Ananya Gurumurthy adds her musical talents to her math and computer science studies to advocate using data for social change.
With the artificial intelligence conversation now mainstream, the 2023 MIT-MGB AI Cures conference saw attendance double from previous years.
The CSAIL scientist describes natural language processing research through state-of-the-art machine-learning models and investigation of how language can enhance other types of artificial intelligence.
Models trained using common data-collection techniques judge rule violations more harshly than humans would, researchers report.
Researchers identify a property that helps computer vision models learn to represent the visual world in a more stable, predictable way.