HyperAI超神経
Back to Headlines

Meta’s New Model Outpaces NVIDIA by 30 Times for Real-World Planning Tasks

3日前

Meta has introduced its latest open-source world model, V-JEPA 2, which operates 30 times faster than Nvidia's Cosmos. This advancement showcases the potential for unsupervised machine learning, enabling robots to operate effectively in real-world environments without requiring labeled data or specific training tasks. In demonstrations, Meta utilized V-JEPA 2-AC, an action-conditioned version of the model, on the Franka robot in various experimental setups. The robot successfully performed tasks such as object retrieval and placement by using image targets to guide its planning. Notably, these tasks were achieved without any prior data collection from the physical environment or specific training tasks. Instead, the model leveraged network data and limited interaction data to develop an effective world model through self-supervised learning. However, it’s important to recognize that V-JEPA 2, like other models, has its limitations. For instance, it does not predict action using camera parameters and relies on hand movements to determine the best camera angles. Additionally, error accumulation and the search space explosion can make it difficult to complete extended planning tasks over longer periods. Meta’s research team has explored multiple variants of the JEPA model, focusing on different sensory modalities, including vision, hearing, and touch. By conducting extensive predictions through these modalities, they aim to create a more comprehensive understanding of the physical world, enabling better planning and decision-making in robotics. For more detailed information, readers can refer to the following resources: - Meta AI Blog: V-JEPA 2 World Model Benchmarks - GitHub Repository: facebookresearch/vjepa2 - Hugging Face Collection: V-JEPA 2 - Meta AI Research Paper: V-JEPA 2 Self-Supervised Video Models Enable Understanding, Prediction, and Planning These developments highlight Meta’s commitment to advancing the field of autonomous robotics and making significant strides in creating versatile, efficient, and self-learning models that can navigate complex real-world scenarios.

Related Links