Data
Data

MIT Sea Grant students explore the intersection of technology and offshore aquaculture in Norway

AquaCulture Shock program, in collaboration with MIT-Scandinavia MISTI, offers international internships for AI and autonomy in aquaculture

Researchers discover a shortcoming that makes LLMs less reliable

Large language models can learn to mistakenly link certain sentence patterns with specific topics — and may then repeat these patterns instead of reasoning.

Charting the future of AI, from safer answers to faster thinking

MIT PhD students who interned with the MIT-IBM Watson AI Lab Summer Program are pushing AI tools to be more flexible, efficient, and grounded in truth.

3 Questions: How AI is helping us monitor and support vulnerable ecosystems

MIT PhD student and CSAIL researcher Justin Kay describes his work combining AI and computer vision systems to monitor the ecosystems that support our planet.

Creating AI that matters

How the MIT-IBM Watson AI Lab is shaping AI-sociotechnical systems for the future.

Method teaches generative AI models to locate personalized objects

After being trained with this technique, vision-language models can better identify a unique item in a new scene.

Optimizing food subsidies: Applying digital platforms to maximize nutrition

An algorithm can change the face of food assistance policy in the Global South, says MIT assistant professor and J-WAFS researcher Ali Aouad.

Using generative AI to diversify virtual training grounds for robots

New tool from MIT CSAIL creates realistic virtual kitchens and living rooms where simulated robots can interact with models of real-world objects, scaling up training data for robot foundation models.

MIT affiliates win AI for Math grants to accelerate mathematical discovery

Department of Mathematics researchers David Roe and Andrew Sutherland seek to advance automated theorem proving; four additional MIT alumni also awarded.

How to build AI scaling laws for efficient LLM training and budget maximization

MIT-IBM Watson AI Lab researchers have developed a universal guide for estimating how large language models will perform based on smaller models in the same family.