Shafaqna English | International Shia News & Fatwas
Featured 2Other NewsSci-Tech

MIT Develops AI Linking Sight and Sound Without Human Guidance

Shafaqna English- Researchers at MIT have created an AI model capable of independently connecting specific video frames with corresponding audio segments, advancing machine perception closer to human-like understanding.

The newly developed AI system, named CAV-MAE Sync, learns to associate visual and auditory information from unlabeled video data, eliminating the need for human annotations. By precisely segmenting audio and balancing multiple learning objectives, the model significantly enhances performance in tasks such as video retrieval and event classification.

This self-supervised approach enables the AI to extract meaningful patterns from raw sensory inputs, similar to how humans integrate information from their environment. Potential applications for this technology span a wide range of fields, including robotics, immersive media, journalism, and accessibility tools.

The breakthrough marks a substantial step forward in building machines capable of richer, more intuitive understanding of real-world audiovisual content, promising to impact industries that rely on multimedia analysis and human-computer interaction.

Source: MIT News

www.shafaqna.com

 

Related posts

Not All Plant-Based Diets Are Heart-Healthy, Study Warns

parniani

Study finds protecting forests key to stabilizing crucial rainfall

parniani

Ancient Pacific migrations could be due to “chase the rain”

parniani

Study Finds Silent Kidney Disease Surges

parniani

Weight-Loss Drugs Show Promise in “Curbing Addiction”

parniani

Animals getting human diseases

parniani

Leave a Comment