Close Menu
Pineapples Update –Pineapples Update –

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    I tried 0patch as a last resort for my Windows 10 PC – here’s how it compares to its promises

    January 20, 2026

    A PC Expert Explains Why Don’t Use Your Router’s USB Port When These Options Are Present

    January 20, 2026

    New ‘Remote Labor Index’ shows AI fails 97% of the time in freelancer tasks

    January 19, 2026
    Facebook X (Twitter) Instagram
    Facebook X (Twitter) Instagram Pinterest Vimeo
    Pineapples Update –Pineapples Update –
    • Home
    • Gaming
    • Gadgets
    • Startups
    • Security
    • How-To
    • AI/ML
    • Apps
    • Web3
    Pineapples Update –Pineapples Update –
    Home»AI/ML»Meta’s new world model allows robots to manipulate the environment that they have never encountered before
    AI/ML

    Meta’s new world model allows robots to manipulate the environment that they have never encountered before

    PineapplesUpdateBy PineapplesUpdateJune 13, 2025No Comments6 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Meta’s new world model allows robots to manipulate the environment that they have never encountered before
    Share
    Facebook Twitter LinkedIn Pinterest Email

    For nearly two decades, join a reliable event by Enterprise leaders. The VB transform brings people together with real venture AI strategy together. learn more


    While the large language model (LLM) has mastered lessons (and somewhat other methods), they have a lack of physical “general knowledge” to work in the dynamic, real-world environment. This has limited the deployment of AI in areas such as manufacturing and logistics, where understanding and effects are important.

    The latest model of Meta, V-JP 2Learning a world model from video and physical interaction takes a step towards reducing this difference.

    V-JEPA 2 AI can help creating applications, which requires predicting and planning results in unexpected environments with several gains cases. This approach can provide more capable robots in the physical environment and a clear route towards advanced automation.

    How to learn how to plan a ‘world model’

    Human beings develop physical intuition in life by looking at their environment. If you see a ball throwing, you easily know its trajectory and can predict where it will land. V-JEPA 2 learns a uniform “world model”, which is the internal simulation of an AI system of how the physical world operates.

    The model is designed at three main capabilities that are required for enterprise applications: understanding what is happening in a scene, predicting how the view will change on an action basis, and planning a sequence of tasks to achieve a specific goal. As stated in meta BlogIts “long -term vision is that world models will enable AI agents to plan and reasons in the physical world.”

    The architecture of the model, which is called video joint embeding Predictive Architecture (V-JEPA), consists of two major parts. A “encoder” sees a video clip and condenses it in a compact numeric summary, known as an embeding. It catches the necessary information about objects and their relationships in the embeding scene. A second component, “prophet”, then takes this summary and imagines how the scene will develop, what the next summary will look like, it will produce a prediction of what the next summary will look.

    Meta’s new world model allows robots to manipulate the environment that they have never encountered before
    V-JEPA is made of an encoder and a prophet (source: meta blog)

    This architecture is the latest development of the JEPA framework, which was first applied to images with I-JEPA and now progresses for the video, which performs a consistent approach to the creation of the world’s model.

    Unlike the generous AI model that tries to predict the exact color of each pixel in the future frame-a computationally intensive work-V-JEPA 2 operates in an abstract place. This focuses on predicting the high-level characteristics of a view, such as the position and trajectory of an object, instead of its texture or background details, it only becomes far more efficient than other large models on 1.2 billion parameters.

    It translates into low calculation costs and makes it more suitable for deployment in real -world settings.

    Learn from observation and action

    The V-Jepa 2 is trained in two stages. First, it creates its fundamental understanding of physics through self-reviewed learning, watching unbelled internet videos of over a million hours. Just seeing how the object walks and interacts, it develops a common-purpose world model without any human guidance.

    In the second phase, this pre-educated model is fine on a small, special dataset. Showing a robot performing functions by processing only 62 hours of video, with this control command, the V-JEPA 2 learns to connect with its physical results. It is a model that can plan and control tasks in the real world.

    V-JEPA two-step training pipeline (Source: Meta)
    V-JEPA two-step training pipeline (Source: Meta)

    This two-step training enables a significant ability to the real-world automation: zero-shot robot scheme. A robot operated by the V-JEPA2 can be deployed in a new environment and without the need to retreat for that specific setting, items can be successfully manipulated before.

    This is an important advance on previous models that require training data perfect Robots and environment where they will work. The model was trained on an open-source dataset and then successfully deployed on various robots in meta laboratories.

    For example, to complete the task such as taking an object, the robot is given a target image of the desired result. It then uses the V-JEPA2 prophet to simulate a series of potentially potentially potential next moves. It scores each imagined action on the basis of how close it becomes to the target, executing the top-rested action, and repeats the process until the work is completed.

    Using this method, the model achieved success rates between 65% and 80% on pick-end-plas tasks with unfamiliar objects in new settings.

    Real world influence of physical logic

    This ability to plan and work in novel conditions has direct implications for business operations. In logistics and manufacturing, it allows for more adaptable robots that can handle varying in products and warehouses without wide reprograming. This can be particularly useful because companies are searching for the deployment of humanoid robots in factories and assembly lines.

    The same world model can give strength to extremely realistic digital twins, allowing companies to imitate new processes or to train other AIs in an physically accurate virtual environment. In industrial settings, a model can monitor the video feed of machinery and, depending on its learned understanding of physics, before predicting safety issues and failures.

    This research is an important step that Meta says “Advanced Machine Intelligence (AMI)”, where AI system “can learn about the world as humans do, plan to execute unfamiliar tasks, and efficiently suit the changing world around us.

    Meta has released the model and its training code and “expecting a comprehensive community around this research that is progressing towards our ultimate goal of developing the world model that can change the way the AI ​​can interact with the physical world.”

    What does this mean for enterprises technical decision making

    The V-JEPA 2 takes robotics closer to software-defined models that already identify the cloud teams: East-train once, deploy anywhere. Because the model learns general physics from public videos and requires only a few dozen hours of work-specific footage, enterprise can reduce data collection cycles that usually reduce pilot projects. In practice, you can prototype a pick-and-place robot on an inexpensive desktop arm, then roll the same policy on an industrial rig on the floor of the factory without collecting thousands of fresh samples or writing a custom motion script.

    Low training overhead also shapes the cost equation again. In 1.2 billion parameters, the V-JEPA 2 fits comfortably on a single high-end GPU, and its abstract prediction goes forward. This allows teams to run on closed-loop controls or on the edge, to avoid the headache coming out with cloud delay and streaming video outside the plant. The budget that had once gone into a large -scale compute clusters can fund additional sensors, excess, or rapid recurrence cycles.

    Daily insights on business use cases with VB daily

    If you want to impress your boss, VB daily has covered you. We give you the scoop inside what companies are doing with generative AI, from regulatory changes to practical deployment, so you can share insight for maximum ROI.

    Read our privacy policy

    Thanks for membership. See more VB newsletters here.

    There was an error.

    encountered environment manipulate Metas model Robots World
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleWhat is chilpad and how is it different from a heating and cooling mattress?
    Next Article A systematic review suggests that the mouth taping pose ‘serious risk of loss’
    PineapplesUpdate
    • Website

    Related Posts

    Startups

    CES 2026: Follow live for the best, weirdest and most interesting tech as physical AI and robots dominate the event

    January 9, 2026
    Startups

    CES 2026: Follow live with news from Caterpillar and Nvidia, as well as robotaxis, robots, and surprises from the show floor

    January 7, 2026
    Startups

    CES 2026: Follow the first official day live with Ring, Mobileye, Siemens, robots, AI and more

    January 6, 2026
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Microsoft’s new text editor is a VIM and Nano option

    May 19, 2025797 Views

    The best luxury car for buyers for the first time in 2025

    May 19, 2025724 Views

    Massives Datenleck in Cloud-Spichenn | CSO online

    May 19, 2025650 Views
    Stay In Touch
    • Facebook
    • YouTube
    • TikTok
    • WhatsApp
    • Twitter
    • Instagram
    Latest Reviews

    Subscribe to Updates

    Get the latest tech news from FooBar about tech, design and biz.

    Most Popular

    10,000 steps or Japanese walk? We ask experts if you should walk ahead or fast

    June 16, 20250 Views

    FIFA Club World Cup Soccer: Stream Palmirus vs. Porto lives from anywhere

    June 16, 20250 Views

    What do chatbott is careful about punctuation? I tested it with chat, Gemini and Cloud

    June 16, 20250 Views
    Our Picks

    I tried 0patch as a last resort for my Windows 10 PC – here’s how it compares to its promises

    January 20, 2026

    A PC Expert Explains Why Don’t Use Your Router’s USB Port When These Options Are Present

    January 20, 2026

    New ‘Remote Labor Index’ shows AI fails 97% of the time in freelancer tasks

    January 19, 2026

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    Facebook X (Twitter) Instagram Pinterest
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms And Conditions
    • Disclaimer
    © 2026 PineapplesUpdate. Designed by Pro.

    Type above and press Enter to search. Press Esc to cancel.