A “scientific sandbox” lets researchers explore the evolution of vision systems
The AI-powered tool could inform the design of better sensors and cameras for robots or autonomous vehicles.
The AI-powered tool could inform the design of better sensors and cameras for robots or autonomous vehicles.
AquaCulture Shock program, in collaboration with MIT-Scandinavia MISTI, offers international internships for AI and autonomy in aquaculture
The innovations map the ocean floor and the brain, prevent heat stroke and cognitive injury, expand AI processing and quantum system capabilities, and introduce new fabrication approaches.
This technique could lead to safer autonomous vehicles, more efficient AR/VR headsets, or faster warehouse robots.
MIT engineers developed a tag that can reveal with near-perfect accuracy whether an item is real or fake. The key is in the glue on the back of the tag.
Jonathan How and his team at the Aerospace Controls Laboratory develop planning algorithms that allow autonomous vehicles to navigate dynamic environments without colliding.
The dataset, being collected as part of a US Coast Guard science mission, will be released open source to help advance naval mission planning and climate change studies.
A new computer vision system turns any shiny object into a camera of sorts, enabling an observer to see around corners or beyond obstructions.
J-WAFS researchers are using remote sensing observations to build high-resolution systems to monitor drought.
Worried about the possibility of flooding near your home in Iowa? Soon, you can just ask Alexa. The move to merge flood sensor data with artificially intelligent chatbots marks the next generation of flood data analysis available all across Iowa. The project is being led by the Iowa Flood Center, based at the University of Iowa, […]