{"id":5291,"date":"2025-10-09T14:32:11","date_gmt":"2025-10-09T14:32:11","guid":{"rendered":"https:\/\/cybersecurityinfocus.com\/?p=5291"},"modified":"2025-10-09T14:32:11","modified_gmt":"2025-10-09T14:32:11","slug":"mit-built-a-virtual-playground-where-robots-learn-to-think","status":"publish","type":"post","link":"https:\/\/cybersecurityinfocus.com\/?p=5291","title":{"rendered":"MIT Built a Virtual Playground Where Robots Learn to Think"},"content":{"rendered":"<p>Before robots can clean your kitchen, they need to train in one. The Massachusetts Institute of Technology (MIT) just built a virtual world where they can practice\u2026 no broken dishes required.<\/p>\n<p>A team from MIT\u2019s Computer Science and Artificial Intelligence Laboratory (CSAIL) and the Toyota Research Institute developed a new AI method called \u201csteerable scene generation.\u201d It quickly builds realistic 3D spaces \u2014 like virtual kitchens and living rooms \u2014 where robots can safely train at scale.<\/p>\n<p>At the heart of the system is a planning technique called Monte Carlo Tree Search (MCTS) \u2014 the same strategy employed by AI programs like AlphaGo to consider numerous possibilities before selecting the best one.<\/p>\n<p>\u201cWe are the first to apply MCTS to scene generation by framing the scene generation task as a sequential decision-making process,\u201d says Nicholas Pfaff, an MIT PhD student and <a href=\"https:\/\/steerable-scene-generation.github.io\/\" target=\"_blank\" rel=\"noopener\">lead author on the project<\/a>. \u201cWe keep building on top of partial scenes to produce better or more desired scenes over time. As a result, MCTS creates scenes that are more complex than what the diffusion model was trained on.\u201d<\/p>\n<p>Robots need that complexity. While chatbots learn from trillions of words, robots depend on realistic visual and physical demonstrations \u2014 data that\u2019s slow and expensive to create by hand. The new system automates this process, using a <a href=\"https:\/\/www.eweek.com\/artificial-intelligence\/generative-ai-model\/\">generative AI model<\/a> that can be \u201csteered\u201d to build detailed, physically accurate scenes.<\/p>\n<p>In one test, this approach allowed the AI to pack a virtual restaurant table with 34 items, including tall stacks of dim sum dishes, after only being trained on scenes with an average of 17 objects.<\/p>\n<h2 class=\"wp-block-heading\">More than just pretty pictures<\/h2>\n<p>Beyond just creating visuals, the system also understands physical logic.<\/p>\n<p>For example, it ensures that a fork doesn\u2019t float through a bowl or that a cup sits firmly on a table. Users can type in commands like \u201ca kitchen with four apples and a bowl on the table,\u201d and the tool will bring that scene to life with impressive accuracy \u2014 98% for pantry scenes and 86% for messy breakfast tables, <a href=\"https:\/\/news.mit.edu\/2025\/using-generative-ai-diversify-virtual-training-grounds-robots-1008\" target=\"_blank\" rel=\"noopener\">according to the researchers<\/a>.<\/p>\n<p>Pfaff said the real breakthrough lies in the tool\u2019s flexibility.<\/p>\n<p>\u201cA key insight from our findings is that it\u2019s OK for the scenes we pre-trained on to not exactly resemble the scenes that we actually want,\u201d he explained. \u201cUsing our steering methods, we can move beyond that broad distribution and sample from a \u2018better\u2019 one.\u201d<\/p>\n<h2 class=\"wp-block-heading\">Industry experts see big potential<\/h2>\n<p>Experts not involved in the project have praised the approach.<\/p>\n<p>Jeremy Binagia, an applied scientist at Amazon Robotics, told MIT News that steerable scene generation \u201coffers a better approach\u201d to realistic simulations because it ensures physical accuracy and 3D awareness, something most previous models lacked.<\/p>\n<p>Rick Cory, a roboticist at the Toyota Research Institute, added that the framework can create \u201c\u2018never-before-seen\u2019 scenes\u201d important for training robots that can adapt to new situations.<\/p>\n<p>While the project is still a proof of concept, <a href=\"https:\/\/www.eweek.com\/news\/flu-vaccine-mit-ai-vaxseer\/\">the MIT team<\/a> hopes to go further. They plan to use generative AI to invent new objects entirely and create more dynamic environments, complete with moving parts like cabinets, jars, and drawers.<\/p>\n<p>Eventually, they aim to combine their technology with internet-scale image data, building a global platform for robot training that mimics the diversity of real life.<\/p>\n<p><strong>For a look at how Google is advancing robot intelligence, read how <\/strong><a href=\"https:\/\/www.eweek.com\/news\/google-gemini-robotics-1-5-er-1-5-launch\/\"><strong>Gemini 1.5 is changing the game<\/strong><\/a><strong>.<\/strong><\/p>\n<p>The post <a href=\"https:\/\/www.eweek.com\/news\/mit-toyota-builds-robot-virtual-playground\/\">MIT Built a Virtual Playground Where Robots Learn to Think<\/a> appeared first on <a href=\"https:\/\/www.eweek.com\/\">eWEEK<\/a>.<\/p>","protected":false},"excerpt":{"rendered":"<p>Before robots can clean your kitchen, they need to train in one. The Massachusetts Institute of Technology (MIT) just built a virtual world where they can practice\u2026 no broken dishes required. A team from MIT\u2019s Computer Science and Artificial Intelligence Laboratory (CSAIL) and the Toyota Research Institute developed a new AI method called \u201csteerable scene [&hellip;]<\/p>\n","protected":false},"author":0,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[2],"tags":[],"class_list":["post-5291","post","type-post","status-publish","format-standard","hentry","category-news"],"_links":{"self":[{"href":"https:\/\/cybersecurityinfocus.com\/index.php?rest_route=\/wp\/v2\/posts\/5291"}],"collection":[{"href":"https:\/\/cybersecurityinfocus.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/cybersecurityinfocus.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/cybersecurityinfocus.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=5291"}],"version-history":[{"count":0,"href":"https:\/\/cybersecurityinfocus.com\/index.php?rest_route=\/wp\/v2\/posts\/5291\/revisions"}],"wp:attachment":[{"href":"https:\/\/cybersecurityinfocus.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=5291"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/cybersecurityinfocus.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=5291"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/cybersecurityinfocus.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=5291"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}