Meta’s self-supervised video world model for dense prediction, planning, and manipulation.
Official PyTorch release of V-JEPA 2/2.1, offering self-supervised video encoders and latent world models that push motion understanding, action anticipation, and robotic manipulation without task-specific tuning.
Sign in to leave a comment.
No comments yet.