Augmenting citizen science with computer vision for fish monitoring
MIT Sea Grant works with the Woodwell Climate Research Center and other collaborators to demonstrate a deep learning-based system for fish monitoring.
MIT Sea Grant works with the Woodwell Climate Research Center and other collaborators to demonstrate a deep learning-based system for fish monitoring.
After being trained with this technique, vision-language models can better identify a unique item in a new scene.
This new machine-learning model can match corresponding audio and visual data, which could someday help robots interact in the real world.
The CausVid generative AI tool uses a diffusion model to teach an autoregressive (frame-by-frame) system to rapidly produce stable, high-resolution videos.
A new method can train a neural network to sort corrupted data while anticipating next steps. It can make flexible plans for robots, generate high-quality video, and help AI agents navigate digital environments.
Researchers find large language models make inconsistent decisions about whether to call the police when analyzing surveillance videos.
Multimedia artist Jackson 2bears reimagines the Haudenosaunee longhouse and creation story.
A new approach could streamline virtual training processes or aid clinicians in reviewing diagnostic videos.
Associate Professor Jonathan Ragan-Kelley optimizes how computer graphics and images are processed for the hardware of today and tomorrow.