{"id":35905,"date":"2025-10-20T20:51:26","date_gmt":"2025-10-20T20:51:26","guid":{"rendered":"https:\/\/agooka.com\/news\/technologies\/mit-researchers-have-built-an-ai-that-teaches-itself-how-to-learn\/"},"modified":"2025-10-20T20:51:26","modified_gmt":"2025-10-20T20:51:26","slug":"mit-researchers-have-built-an-ai-that-teaches-itself-how-to-learn","status":"publish","type":"post","link":"https:\/\/agooka.com\/news\/technologies\/mit-researchers-have-built-an-ai-that-teaches-itself-how-to-learn\/","title":{"rendered":"MIT researchers have built an AI that teaches itself how to learn"},"content":{"rendered":"<p><img decoding=\"async\" src=\"https:\/\/dataconomy.com\/wp-content\/uploads\/2025\/10\/MIT-researchers-have-built-an-AI-that-teaches-itself-how-to-learn.jpg\" alt=\"MIT researchers have built an AI that teaches itself how to learn\" title=\"MIT researchers have built an AI that teaches itself how to learn\"\/><\/p>\n<p>Large language models like ChatGPT have a fundamental problem: they\u2019re static. They are trained on a mountain of data and then frozen in time, like a textbook printed in 2023 that knows nothing about 2024. Now, researchers at MIT\u2019s Improbable AI Lab have open-sourced a new framework that could change that. Their paper, presented at the recent NeurIPS 2025 conference, unveils a system called Self-Adapting Language Models (SEAL).<\/p>\n<p>The core idea is simple, but the implications are huge: the AI learns to teach itself. Instead of just passively holding information, SEAL enables a model to generate its own high-quality training data and then use that data to permanently update its own weights. This matters because it\u2019s the first real step away from static, \u201cknow-it-all\u201d bots and toward AI models that can actually evolve, adapt, and incorporate new information over time.<\/p>\n<h2>Why AI models are bad students<\/h2>\n<p>Right now, if you want an LLM to learn a new fact, you have two bad options. You can \u201cstuff\u201d the information into its context window (the prompt), but it will forget that fact the moment the conversation resets. Or, you can perform a massive, expensive retraining, which is like reprinting an entire encyclopedia just to add a new entry. Neither of these methods is true learning.<\/p>\n<p>The MIT team, including Adam Zweiger, Jyothish Pari, and Pulkit Agrawal, looked at how humans learn. When a student prepares for an exam, they don\u2019t just re-read the textbook 50 times. A good student rewrites the information, making flashcards, summarizing chapters, and creating their own notes. This process of reformatting and assimilating information is what cements it in their brain.<\/p>\n<p>SEAL is designed to be that good student. It learns to take the \u201craw textbook\u201d of new information and generate its own \u201cstudy notes\u201d\u2014which the paper calls \u201cself-edits\u201d\u2014in whatever format is most effective for its own learning.<\/p>\n<h3>So, how does it learn to \u2018study\u2019?<\/h3>\n<p>It learns through trial and error, using a process called reinforcement learning. Think of it as an AI holding its own study sessions.<\/p>\n<ol>\n<li>Get the lesson: The AI is given a new piece of information (like a passage of text).<\/li>\n<li>Write the notes: It generates a \u201cself-edit\u201d\u2014its own synthetic notes on that info. This could be a list of key implications, a set of question-and-answer pairs, or just a simple summary.<\/li>\n<li>Take the quiz: The AI is briefly fine-tuned on its own notes and then immediately given a pop quiz on the new information.<\/li>\n<li>Get the grade: If it passes the quiz, it gets a \u201creward.\u201d This positive feedback teaches the model that the \u201cself-edit\u201d notes it just wrote were high-quality and effective.<\/li>\n<li>Study smarter: If it fails, it learns that its notes were bad and tries a different format next time. Over thousands of these loops, the AI doesn\u2019t just learn the new facts; it learns how to learn new facts more efficiently.<\/li>\n<\/ol>\n<h3>And the results?<\/h3>\n<p>The researchers tested SEAL in two key areas, and the results are striking.<\/p>\n<p>First, they tested its ability to incorporate new knowledge. They gave the model text passages and quizzed it on the contents. After training itself with SEAL, the AI\u2019s accuracy jumped to 47.0%. Here\u2019s the kicker: that score outperformed synthetic data generated by the much larger and more powerful GPT-4.1, which only scored 46.3%. The smaller model literally taught itself to be \u201csmarter\u201d than its massive competitor at this specific task.<\/p>\n<p>Second, they tested its ability to learn a new skill from just a few examples. This is a notoriously hard abstract reasoning benchmark called ARC. SEAL\u2019s job wasn\u2019t just to solve the puzzle, but to generate the best learning strategy for itself (e.g., \u201cuse these data augmentations,\u201d \u201cset this learning rate\u201d). The self-adapting AI found a successful strategy 72.5% of the time. The baseline model, without this self-learning, fumbled, succeeding only 20% of the time.<\/p>\n<h2>What\u2019s the catch?<\/h2>\n<p>This all sounds great, but a pragmatist would be right to ask about the downsides. The researchers are transparent about the limitations.<\/p>\n<ul>\n<li>Catastrophic forgetting: The model still suffers from the classic AI problem of \u201ccatastrophic forgetting.\u201d As it crams for new exams, it starts to forget what it learned for the midterms. Learning a new fact can still overwrite old ones.<\/li>\n<li>It\u2019s painfully slow: This process is not fast. The researchers note that the computational overhead is \u201csubstantial.\u201d It takes 30-45 seconds just to grade a single self-edit during the training loop.<\/li>\n<li>It needs an answer key: The current system relies on having a \u201cquiz\u201d with correct answers to provide that all-important reward signal.<\/li>\n<\/ul>\n<p>Despite these hurdles, the team is looking ahead. Experts project that we will run out of high-quality human-generated text to train AI on by 2028. When we hit that \u201cdata wall,\u201d progress will hinge on a model\u2019s ability to generate its own high-utility training data. This research is a crucial roadmap for how that might work, paving the way for future AI \u201cagents\u201d that don\u2019t just answer your questions, but actively learn from their interactions with the world and get smarter every day.<\/p>\n<p><a href=\"https:\/\/unsplash.com\/photos\/chart-treemap-chart-GjBPLkTDzt4\" rel=\"noreferrer\" target=\"_blank\"><strong>Featured image credit<\/strong><\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Large language models like ChatGPT have a fundamental problem: they\u2019re static. They are trained on a mountain of data and then frozen in time, like a textbook printed in 2023 that knows nothing about 2024. Now, researchers at MIT\u2019s Improbable AI Lab have open-sourced a new framework that could change that. Their paper, presented at [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":35906,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[37],"tags":[],"class_list":{"0":"post-35905","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-technologies"},"_links":{"self":[{"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/posts\/35905","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/comments?post=35905"}],"version-history":[{"count":0,"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/posts\/35905\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/media\/35906"}],"wp:attachment":[{"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/media?parent=35905"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/categories?post=35905"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/agooka.com\/news\/wp-json\/wp\/v2\/tags?post=35905"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}