Particle.news
Download on the App Store

Meta Launches SAM 3 and SAM 3D for Text-Driven Segmentation and Single-Image 3D

The company is releasing model assets plus a public Playground to speed adoption across creative and commerce tools.

Overview

  • SAM 3 detects, segments and tracks objects in images and video from natural-language prompts, including conditionals and exclusions.
  • SAM 3D reconstructs 3D objects, scenes and human body shape from a single image and debuts an artist-curated evaluation dataset.
  • Meta published model weights, checkpoints, code, benchmarks and research papers, with access available now through the Segment Anything Playground.
  • Reported results include 47.0 zero-shot mask AP on LVIS and roughly 30 milliseconds per frame on H200 GPUs while handling over 100 objects.
  • Early product uses include selective edits in the Edits app and Vibes, plus a Marketplace ‘View in Room’ feature for previewing items at home.