{"id":6391,"date":"2025-07-21T21:00:00","date_gmt":"2025-07-21T19:00:00","guid":{"rendered":"https:\/\/aitrends.center\/a-new-way-to-edit-and-generate-images-without-traditional-ai-generators\/"},"modified":"2025-07-24T13:02:52","modified_gmt":"2025-07-24T11:02:52","slug":"nowy-sposob-edycji-i-generowania-obrazow-bez-tradycyjnych-generatorow-ai","status":"publish","type":"post","link":"https:\/\/aitrendscenter.eu\/pl\/a-new-way-to-edit-and-generate-images-without-traditional-ai-generators\/","title":{"rendered":"Nowy spos\u00f3b edycji i generowania obraz\u00f3w bez tradycyjnych generator\u00f3w AI"},"content":{"rendered":"<h2>Reimagining How AI Creates and Edits Images<\/h2>\n<p>\nWe\u2019re living in an age when artificial intelligence can whip up a bizarre fantasy or hyper-realistic photo out of a few words. All that magic rests on giant neural networks quietly learning from millions\u2014sometimes billions\u2014of photos and written descriptions. But the next breakthrough in AI art might be slight of hand: what if you could create and transform images with AI, but without the heavy machinery of a traditional image generator?\n<\/p>\n<p>\nThat exact thought inspired a group of researchers from MIT and Facebook AI Research. Their latest paper, unveiled at the 2025 International Conference on Machine Learning (ICML), shows how a surprisingly simple technique could dramatically streamline how AI conjures up and edits pictures.\n<\/p>\n<h3>From Class Project to Groundbreaking Discovery<\/h3>\n<p>\nThe journey started as a class project for MIT grad student Lukas Lao Beyer, mentored by Professor Kaiming He. What began modestly soon attracted collaborators\u2014Tianhong Li, Xinlei Chen, and Sertac Karaman\u2014and took on a much larger scope.\n<\/p>\n<p>\nBeyer&#8217;s inspiration came from new research out of the Technical University of Munich and ByteDance. That team created a \u201c1D tokenizer\u201d\u2014an unusual AI model that distills a 256&#215;256 image down to just 32 tokens, each a 12-digit binary value. Imagine it as an ultra-condensed \u201clanguage\u201d of 4,000 words, only this language describes pictures instead of ideas.\n<\/p>\n<p>\nFascinated by how much information these tokens carried, Beyer began experimenting: by swapping out tokens in an image, he watched the picture morph\u2014resolution jumped, colors dulled or popped, backgrounds became crisp or blurred, and sometimes a bird\u2019s pose simply changed direction.\n<\/p>\n<h3>No Generator Needed: Editing at the Core<\/h3>\n<p>\nHere\u2019s where it gets wild: rather than redrawing the pixels directly, Beyer and the team started editing specific tokens to get the changes they wanted in an image. Turns out this kind of swap isn\u2019t just for tinkering; it offers a whole new way to generate images, too. By stacking a 1D tokenizer with a decoder that reconstructs the image, plus a guiding neural network called CLIP (which keeps the AI\u2019s work aligned with a text prompt), the team could \u201cnudge\u201d bundles of tokens until a picture of, say, a red panda began looking like a tiger\u2014or even create something entirely new by starting from random tokens and gradually refining them to fit a prompt.\n<\/p>\n<p>\nWhat\u2019s astonishing is how much this sidesteps the need for old-school image generators. Traditional generators are hulking beasts\u2014slow to train, power-hungry to run. This more minimalist method sidesteps much of the computational slog, making advanced AI art potentially cheaper and more widespread. It can also perform tricks like inpainting\u2014filling in missing image parts\u2014using the same tokenizer and decoder setup. Sometimes, the boldest advances come from reusing tools in unexpected ways.\n<\/p>\n<p>\nExperts are paying attention. Saining Xie from NYU called the results \u201cpretty surprising,\u201d while Princeton\u2019s Zhuang Liu pointed out the promising possibility of cutting AI image costs.\n<\/p>\n<p>\nAnd the potential isn\u2019t limited to pictures. Sertac Karaman highlighted how the same idea\u2014compressing actions or driving directions into tokens\u2014could benefit fields like robotics or self-driving cars. Beyer agrees: \u201cThe power of that kind of compression could unlock amazing things,\u201d he says.\n<\/p>\n<p>\nAt its heart, this new approach shows what can happen when you dare to challenge the old ways and make the most of the tools at hand. By thinking differently, the MIT team hasn\u2019t just found a shortcut\u2014they\u2019ve paved a new road for AI creativity.\n<\/p>\n<p>\n<strong>Przeczytaj oryginalny artyku\u0142: <a href=\"https:\/\/news.mit.edu\/2025\/new-way-edit-or-generate-images-0721\" target=\"_blank\" rel=\"noopener\">https:\/\/news.mit.edu\/2025\/new-way-edit-or-generate-images-0721<\/a><\/strong><\/p>","protected":false},"excerpt":{"rendered":"<p>Reimagining How AI Creates and Edits Images We\u2019re living in an age when artificial intelligence can whip up a bizarre fantasy or hyper-realistic photo out of a few words. All that magic rests on giant neural networks quietly learning from millions\u2014sometimes billions\u2014of photos and written descriptions. But the next breakthrough in AI art might be slight of hand: what if you could create and transform images with AI, but without the heavy machinery of a traditional image generator? That exact thought inspired a group of researchers from MIT and Facebook AI Research. Their latest paper, unveiled at the 2025 International [&hellip;]<\/p>\n","protected":false},"author":4,"featured_media":6392,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[2,47],"tags":[],"class_list":["post-6391","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-ai-images","category-ai-news","post--single"],"_links":{"self":[{"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/posts\/6391","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/users\/4"}],"replies":[{"embeddable":true,"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/comments?post=6391"}],"version-history":[{"count":1,"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/posts\/6391\/revisions"}],"predecessor-version":[{"id":6445,"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/posts\/6391\/revisions\/6445"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/media\/6392"}],"wp:attachment":[{"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/media?parent=6391"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/categories?post=6391"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/aitrendscenter.eu\/pl\/wp-json\/wp\/v2\/tags?post=6391"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}