{"id":6294,"date":"2025-07-08T06:00:00","date_gmt":"2025-07-08T04:00:00","guid":{"rendered":"https:\/\/aitrends.center\/mit-researchers-boost-llm-reasoning-with-test-time-training\/"},"modified":"2025-07-24T13:13:16","modified_gmt":"2025-07-24T11:13:16","slug":"naukowcy-z-mit-zwiekszaja-zdolnosc-rozumowania-na-poziomie-lm-dzieki-treningowi-w-czasie-testu","status":"publish","type":"post","link":"https:\/\/aitrendscenter.eu\/pl\/mit-researchers-boost-llm-reasoning-with-test-time-training\/","title":{"rendered":"Naukowcy z MIT usprawniaj\u0105 rozumowanie LLM dzi\u0119ki szkoleniom w czasie test\u00f3w"},"content":{"rendered":"<p>Large language models have made major leaps in how computers understand and use human language\u2014you see their handiwork in everything from rapid-fire document summaries to instant translations and slick Q&#038;A bots. But as smart as they are, these models can hit a wall when faced with challenges that call for genuine reasoning or a touch of human logic. Imagine an AI that spins out financial summaries without breaking a sweat, but struggles if you ask it to sniff out suspicious transactions or predict where the market is headed next. The reason? Once these models are released into the wild, they can\u2019t really pick up new skills or make themselves smarter on their own.<\/p>\n<h5>Test-Time Training: Giving AI a Second Wind<\/h5>\n<p>Researchers at MIT decided that wasn\u2019t good enough, and developed something called \u201ctest-time training.\u201d Instead of updating the entire model or retraining it from scratch, this clever approach tweaks certain parts of the AI on the fly, just when you need it. So the model stays the same at its core, but can temporarily adapt to solve trickier or totally new problems. Ekin Aky\u00fcrek, a member of the team, explained it neatly: these AI models can\u2019t genuinely learn by themselves after deployment\u2014but with a small, smart push, their abilities take a big leap forward.<\/p>\n<p>Before this, most people relied on what\u2019s called \u201cin-context learning\u201d to improve how these models handle new problems: you show them a handful of examples and let them pick up the pattern. It works\u2014up to a point. The trouble is, if a task really needs reasoning or flexibility, this approach often falls flat, making the results underwhelming.<\/p>\n<h5>A Big Boost from Hybrid Thinking<\/h5>\n<p>The MIT team took things up another notch by combining in-context learning with test-time training. Here\u2019s what makes it unique: during inference\u2014that stage when the AI is actually solving a problem\u2014the model gets a mini-crash course using a batch of examples tailored just for the job. Instead of overhauling the AI\u2019s massive set of parameters, only a tiny, critical subset gets updated. This trick, known as \u201clow-rank adaptation,\u201d keeps things light and fast, delivering huge performance gains with surgical efficiency. In fact, the combo can make the AI perform as much as six times better than using in-context learning alone.<\/p>\n<p>To further toughen up this quick-training set, the researchers didn\u2019t just reuse examples as-is. They jazzed up the data, flipping it, mixing it, and modifying inputs to create an even richer batch. What really stands out is that all these changes happen just in time for the new problem, without disturbing the model\u2019s core knowledge. Sure, this real-time learning can make each query take a little longer\u2014perhaps a minute turning into ten\u2014but the payoff is a dramatic jump in accuracy, which matters most when tackling especially complex or high-stakes questions.<\/p>\n<h5>Where Could This Go Next?<\/h5>\n<p>The team put their hybrid method through its paces on some seriously tough challenges\u2014think tests filled with IQ puzzles and hard-to-spot patterns. The results? Tasks requiring nuanced reasoning or making sense of never-before-seen data saw clear, measurable improvements. As graduate student Mehul Damani puts it, in-context learning is fine for straightforward jobs, but when you allow the model to tweak itself on the fly, it\u2019s like teaching it a brand-new skill on demand.<\/p>\n<p>The next frontier? Building models that can keep learning as they go, figuring out for themselves whether to stick to examples or jump into real-time training, all without human help. If accomplished, this would mark a significant step toward AIs that are not just bigger, but genuinely smarter and more flexible.<\/p>\n<p>The research was a collaborative effort, supported by the MIT-IBM Watson AI Lab and the National Science Foundation, and it\u2019ll debut at the International Conference on Machine Learning. For all the technical nuts and bolts, check out the complete research writeup at <a href=\"https:\/\/news.mit.edu\/2025\/study-could-lead-llms-better-complex-reasoning-0708\" target=\"_blank\" rel=\"noopener\">MIT News<\/a>.<\/p>","protected":false},"excerpt":{"rendered":"<p>Large language models have made major leaps in how computers understand and use human language\u2014you see their handiwork in everything from rapid-fire document summaries to instant translations and slick Q&#038;A bots. But as smart as they are, these models can hit a wall when faced with challenges that call for genuine reasoning or a touch of human logic. Imagine an AI that spins out financial summaries without breaking a sweat, but struggles if you ask it to sniff out suspicious transactions or predict where the market is headed next. The reason? Once these models are released into the wild, they [&hellip;]<\/p>\n","protected":false},"author":4,"featured_media":6295,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[47],"tags":[],"class_list":["post-6294","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-ai-news","post--single"],"_links":{"self":[{"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/posts\/6294","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/users\/4"}],"replies":[{"embeddable":true,"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/comments?post=6294"}],"version-history":[{"count":1,"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/posts\/6294\/revisions"}],"predecessor-version":[{"id":6493,"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/posts\/6294\/revisions\/6493"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/media\/6295"}],"wp:attachment":[{"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/media?parent=6294"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/categories?post=6294"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/tags?post=6294"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}