Study: Platforms that rank the latest LLMs can be unreliable
Removing just a tiny fraction of the crowdsourced data that informs online ranking platforms can significantly change the results.
Removing just a tiny fraction of the crowdsourced data that informs online ranking platforms can significantly change the results.
New research detects hidden evidence of mistaken correlations — and provides a method to improve accuracy.
While the growing energy demands of AI are worrying, some techniques can also help make power grids cleaner and more efficient.
New research demonstrates how AI models can be tested to ensure they don’t cause harm by revealing anonymized patient health data.
The technique can help scientists in economics, public health, and other fields understand whether to trust the results of their experiments.
This new technique enables LLMs to dynamically adjust the amount of computation they use for reasoning, based on the difficulty of the question.
With insect-like speed and agility, the tiny robot could someday aid in search-and-rescue missions.
MIT CSAIL and LIDS researchers developed a mathematically grounded system that lets soft robots deform, adapt, and interact with people and objects, without violating safety limits.
Large language models can learn to mistakenly link certain sentence patterns with specific topics — and may then repeat these patterns instead of reasoning.
MIT PhD students who interned with the MIT-IBM Watson AI Lab Summer Program are pushing AI tools to be more flexible, efficient, and grounded in truth.