Particle.news

Download on the App Store

Meta Unveils V-JEPA2 to Power Predictive Robotics in New Environments

Operating in a streamlined latent space, it learns from over one million hours of raw video to power predictive action in new environments.

Image
From delivery robots to self-driving cars, Meta’s V-JEPA 2 aims to power a new era of intelligent machines that think before they act.
Image
Meta V-JEPA 2

Overview

  • V-JEPA2 is a 1.2-billion-parameter world model trained on more than one million hours of unlabeled video to internalize gravity, motion and object interactions.
  • The model’s latent-space architecture delivers physical predictions about 30 times faster than Nvidia’s Cosmos, cutting compute and latency for real-world tasks.
  • In two-stage testing, lab robots using V-JEPA2 achieved 65–80% success rates on zero-shot pick-and-place tasks with previously unseen objects and settings.
  • Meta has published three open benchmarks to enable the research community to evaluate how well video-trained agents learn, predict and plan in physical scenarios.
  • Positioned as a milestone toward Advanced Machine Intelligence, V-JEPA2 holds potential for logistics automation, manufacturing and autonomous systems.