Meta's new architecture helps robots interact in environments they've never seen before
Briefly

Meta has introduced the Video Joint Embedding Predictive Architecture 2 (V-JEPA 2), an open-source model trained primarily on video to enhance AI's capabilities in robotics and automation. This advanced model, trained on over one million hours of video, is designed for tasks like predictive analytics in various sectors, including manufacturing and surveillance. Unlike traditional computer vision systems, V-JEPA 2 excels in operating within unstructured environments, allowing robots to navigate and adapt to new settings efficiently. Experts emphasize this development as a significant shift in the AI vision landscape, crucial for future enterprise implementations.
Meta's recent unveiling of V-JEPA 2 marks a quiet but significant shift in the evolution of AI vision systems, and it's one enterprise leaders can't afford to overlook.
Trained for predictive tasks on more than 1 million hours of video, V-JEPA 2 achieves state-of-the-art performance on visual understanding and prediction in physical environments.
Read at Computerworld
[
|
]