<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Tutorial &#8211; aitrendscenter.eu</title>
	<atom:link href="https://aitrendscenter.eu/category/tutorial/feed/" rel="self" type="application/rss+xml" />
	<link>https://aitrendscenter.eu</link>
	<description>Written by AI, about AI, for humans (and occasionally confused robots)</description>
	<lastBuildDate>Thu, 18 Dec 2025 21:20:00 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>

<image>
	<url>https://aitrendscenter.eu/wp-content/uploads/2025/05/cropped-Sygnet-32x32.png</url>
	<title>Tutorial &#8211; aitrendscenter.eu</title>
	<link>https://aitrendscenter.eu</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>Guided Learning Helps &#8216;Untrainable&#8217; Neural Networks Reach New Potential</title>
		<link>https://aitrendscenter.eu/guided-learning-helps-untrainable-neural-networks-reach-new-potential/</link>
		
		<dc:creator><![CDATA[Max Krawiec]]></dc:creator>
		<pubDate>Thu, 18 Dec 2025 21:20:00 +0000</pubDate>
				<category><![CDATA[News]]></category>
		<category><![CDATA[Tutorial]]></category>
		<guid isPermaLink="false">https://aitrendscenter.eu/guided-learning-helps-untrainable-neural-networks-reach-new-potential/</guid>

					<description><![CDATA[It&#8217;s easy to write off certain neural networks as &#8220;untrainable&#8221; when they fall short of modern machine learning tasks. But a team of researchers at MIT&#8217;s Computer Science and Artificial Intelligence Laboratory (CSAIL) reminds us that we just might be wrong. They&#8217;ve unravelled a new method known as guidance. It&#8217;s a brief phase of alignment that can dramatically enhance the learning ability of previously dismissed neural network architectures. Revamping the Underdogs Traditionally, we&#8217;ve considered some network architectures as inherently flawed or limited in dealing with complex tasks. However, these neural networks might be victim to an unfavorable starting point in [&#8230;]]]></description>
										<content:encoded><![CDATA[<p>It&#8217;s easy to write off certain neural networks as &#8220;untrainable&#8221; when they fall short of modern machine learning tasks. But a team of researchers at MIT&#8217;s Computer Science and Artificial Intelligence Laboratory (CSAIL) reminds us that we just might be wrong. They&#8217;ve unravelled a new method known as <em>guidance</em>. It&#8217;s a brief phase of alignment that can dramatically enhance the learning ability of previously dismissed neural network architectures.</p>
<h5>Revamping the Underdogs</h5>
<p>Traditionally, we&#8217;ve considered some network architectures as inherently flawed or limited in dealing with complex tasks. However, these neural networks might be victim to an unfavorable starting point in the parameter space rather than a lack of potential. The researchers found that by pairing these networks with a structured &#8220;guide&#8221; network for a short while, they could steer the struggling networks towards effective learning.</p>
<p>This technique, unlike knowledge distillation—in which a student model emulates the outputs of a teacher—relies on internal representations. Here, the target network absorbs how the guide network organizes information across its layers instead of mimicking its predictions. Even when the guide network is untrained, this process facilitates a meaningful transfer of knowledge, thereby enhancing the learning ability.</p>
<p>This theory was put to the test using deep fully connected networks (FCNs). The researchers briefly aligned the networks with a guide network using random noise before the actual training. The results were astonishing–notorious for overfitting, these networks became more stable, avoided the usual FCN pitfalls, displayed lower training losses and improved their performances. “It’s impressive that we could use representational similarity to make these traditionally ‘crappy’ networks actually work,&#8221; says Vighnesh Subramaniam ’23, MEng ’24, a PhD student in MIT’s Department of Electrical Engineering and Computer Science and lead author of the study.</p>
<h5>Game-Changer for Neural Networks</h5>
<p>The study reveals that guidance, unlike knowledge distillation, doesn&#8217;t falter when using an untrained teacher network. As guidance relies on the network&#8217;s internal structure which carries valuable architectural biases. These biases function like a compass, directing the network towards better learning paths.</p>
<p>But the implications of this research don&#8217;t stop at performance improvements. It suggests that the success of a network may be more dependent on its starting point in the learning space than the data it’s trained on. By pairing networks with a guide, the impact of architectural design can be isolated from learned experience. This introduction of guidance provides a new perspective for evaluating network structures&#8217; contribution to effective learning. It also gives scientists a way to understand the differences between architectures, helping refine theories about neural network optimization and identifying which components are consequential for learning.</p>
<p>The highlight, however, is that no network is beyond redemption. Even those once branded as ineffective can be brought up to par with modern standards through guidance. Currently, the CSAIL team is studying which architectural elements are major contributors to these improvements, intending to influence future neural network design.</p>
<p>“It’s generally assumed that different neural network architectures have particular strengths and weaknesses,” noted Leyla Isik, an assistant professor of cognitive science at Johns Hopkins University, who was not involved in the study. “This exciting research shows that one type of network can inherit the advantages of another architecture, without losing its original capabilities.”</p>
<p>The research, a collective effort of Subramaniam and his MIT CSAIL collaborators, was supported by organizations such as the Center for Brains, Minds, and Machines, the National Science Foundation, the MIT-IBM Watson AI Lab, and the U.S. Department of the Air Force Artificial Intelligence Accelerator. Their game-changing findings were recently presented at the Conference and Workshop on Neural Information Processing Systems (NeurIPS).</p>
<p>Read the original article from MIT News here: <a href="https://news.mit.edu/2025/guided-learning-lets-untrainable-neural-networks-realize-their-potential-1218" target="_blank" rel="noopener">https://news.mit.edu/2025/guided-learning-lets-untrainable-neural-networks-realize-their-potential-1218</a></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Helping K-12 Schools Navigate the Complex World of AI</title>
		<link>https://aitrendscenter.eu/helping-k-12-schools-navigate-the-complex-world-of-ai/</link>
		
		<dc:creator><![CDATA[Max Krawiec]]></dc:creator>
		<pubDate>Mon, 03 Nov 2025 21:45:00 +0000</pubDate>
				<category><![CDATA[News]]></category>
		<category><![CDATA[Tutorial]]></category>
		<guid isPermaLink="false">https://aitrendscenter.eu/helping-k-12-schools-navigate-the-complex-world-of-ai/</guid>

					<description><![CDATA[Emerging technologies can often pose a sea of challenge and excitement, especially in our classrooms. As generative artificial intelligence advances, educators nationwide are eagerly trying to bring this powerhouse technology into their lessons without hindering student growth. The sudden influx of AI tools in schools has teachers and leaders urgently looking for ways to responsibly and effectively welcome this change. Enter Justin Reich, associate professor at MIT’s Comparative Media Studies/Writing program. Reich has written a lifeboat for educators in the form of a guidebook titled &#8220;A Guide to AI in Schools: Perspectives for the Perplexed&#8221;, through the MIT Teaching Systems [&#8230;]]]></description>
										<content:encoded><![CDATA[<p>Emerging technologies can often pose a sea of challenge and excitement, especially in our classrooms. As generative artificial intelligence advances, educators nationwide are eagerly trying to bring this powerhouse technology into their lessons without hindering student growth. The sudden influx of AI tools in schools has teachers and leaders urgently looking for ways to responsibly and effectively welcome this change.</p>
<p>Enter Justin Reich, associate professor at MIT’s Comparative Media Studies/Writing program. Reich has written a lifeboat for educators in the form of a guidebook titled <a href="https://tsl.mit.edu/ai-guidebook/" target="_blank" rel="noopener">&#8220;A Guide to AI in Schools: Perspectives for the Perplexed&#8221;</a>, through the MIT Teaching Systems Lab. The publication aims to hold the hands of K–12 educators as they sail the unexplored waters of AI in education. &#8220;Over my career, I’ve aimed to be a bridge between educational research and practical insights. My goal is to simplify and share insights about complex concepts like AI&#8221;, Reich shares.</p>
<h5>A Compass, Not a Map</h5>
<p>In this guidebook, you won&#8217;t find strict rules or &#8216;best practices&#8217;. The content is diverse, featuring thought-provoking experiences and insights from over 100 students and teachers across America. Instead of pushing for a &#8216;one-size-fits-all&#8217; approach, it stimulates dialogue and encourages a culture of experimentation in classrooms. Reich&#8217;s ethos aligns with this approach, advocating a humble acceptance that some strategies might work better than others—an understanding that will become clearer over time.</p>
<h5>Important Questions and Conversations</h5>
<p>The advent of AI has caused an influx of questions. How do we safeguard academic integrity? How do we protect student data privacy? And how does learning change when students can use AI to bypass the hard work of thinking? Reich states that we&#8217;re only starting to wrap our minds around these questions. The scenario, he points out, is akin to us sitting in the teacher&#8217;s lounge, conversing about AI, before inviting students, parents and others to join the discussion. Comparing this to the early days of aviation, the guidebook authors note that predicting what&#8217;s going to work would be premature at this stage.</p>
<h5>Learning from History</h5>
<p>Reich is open about the potential problems that lie ahead. &#8220;We’re finding our way in the dark,&#8221; he candidly admits, drawing parallels with past blunders with classroom technology. Reflecting on flawed strategies such as discouraging students from using Wikipedia or looking for antiquated credibility markers on websites, Reich urges caution. He espouses an evidence-based approach over sweeping assumptions about AI, advocating for trials, observation, and learning from the outcomes.</p>
<p>Thus, Reich&#8217;s guidebook and corresponding podcast series <a href="https://www.teachlabpodcast.com/" target="_blank" rel="noopener">&#8220;The Homework Machine&#8221;</a>, a collaboration with journalist Jesse Dukes, provide vital resources for educators navigating AI in the classroom. Reich encourages a diverse conversation around AI, involving voices from various stakeholders. This dialog has further been amplified through an hour-long radio special that aired on public radio stations nationwide, thereby reaching a wider audience.</p>
<p>While the demands of navigating new technology like AI can be high, Reich underlines the need for caution, collaboration, and a collective willingness to learn from our past experiences—as, unlike previous technologies, AI has seeped into students&#8217; lives, not through traditional channels but through their smartphones. This has placed even more pressure on educators to respond without clear directions or tried-and-tested practices. But this doesn&#8217;t mean they should be denied a voice in shaping its use. As Reich says, “AI is already here. Teachers don’t get to choose whether it exists, but they should have strong input on shaping its future.”</p>
<p>For more insights and to join this ongoing conversation, you can visit the original article from MIT News: <a href="https://news.mit.edu/2025/helping-k-12-schools-navigate-complex-world-of-ai-1103" target="_blank" rel="noopener">Helping K-12 Schools Navigate the Complex World of AI</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>New Research Enhances Confidence in Long-Term Nuclear Waste Storage</title>
		<link>https://aitrendscenter.eu/new-research-enhances-confidence-in-long-term-nuclear-waste-storage/</link>
		
		<dc:creator><![CDATA[Max Krawiec]]></dc:creator>
		<pubDate>Fri, 18 Jul 2025 04:00:00 +0000</pubDate>
				<category><![CDATA[News]]></category>
		<category><![CDATA[Tutorial]]></category>
		<guid isPermaLink="false">https://aitrends.center/new-research-enhances-confidence-in-long-term-nuclear-waste-storage/</guid>

					<description><![CDATA[As nuclear power regains popularity in the push for cleaner energy, a big challenge remains unresolved: what do we do with all the nuclear waste? In the United States, plans for a permanent underground repository have stalled, leaving a critical gap in the country’s nuclear energy strategy. The problem isn’t just political—it’s scientific, too. Understanding precisely how nuclear waste behaves when stored underground is still a major puzzle. Researchers from MIT, Lawrence Berkeley National Lab, and the University of Orléans are on the case, and their latest findings bring much-needed clarity. In a study recently published in PNAS, the team [&#8230;]]]></description>
										<content:encoded><![CDATA[<p>As nuclear power regains popularity in the push for cleaner energy, a big challenge remains unresolved: what do we do with all the nuclear waste? In the United States, plans for a permanent underground repository have stalled, leaving a critical gap in the country’s nuclear energy strategy. The problem isn’t just political—it’s scientific, too. Understanding precisely how nuclear waste behaves when stored underground is still a major puzzle.</p>
<p>Researchers from MIT, Lawrence Berkeley National Lab, and the University of Orléans are on the case, and their latest findings bring much-needed clarity. In a study recently published in <em>PNAS</em>, the team managed to create computer models that mirror real-world experiments with impressive accuracy. That might sound technical, but it’s a big deal: having models that we can trust makes designing safer storage methods a whole lot easier.</p>
<p>What makes the research groundbreaking is its connection to the Mont Terri laboratory in Switzerland, a site that’s been at the forefront of nuclear waste studies since the 1990s. The scientists focused on Opalinus clay, a natural material thought to be ideal for locking away radioactive material. By running experiments at Mont Terri and then feeding that data into sophisticated computer simulations, they were able to see how artificial barriers (like cement) interact with clay deep underground, not just for months or years, but potentially for centuries.</p>
<p>This progress owes a lot to new tools. The latest computer model, called CrunchODiTi, is a notable upgrade over older programs. Unlike previous versions, CrunchODiTi can account for the tiny electrical charges in clay minerals. Those details matter because they help predict how radioactive particles might move—or stay put—over very long periods of time.</p>
<p>The team zeroed in on a tiny, crucial “skin” layer, just a centimeter thick, where cement and clay meet. This interface plays an outsized role in the slow migration of radioactive elements through the subsurface. Data from a 13-year experiment at Mont Terri gave the researchers a rare, long-term look at changes in this skin zone. Their computer model matched the physical observations, lending real confidence that they’re on the right track.</p>
<p>Dauren Sarsenbayev, the study’s lead author, finds it remarkable to observe these changes unfold. Seeing how the intersection between cement and clay evolves over time helps bridge the gap between theory and reality. The team’s observations also support long-standing ideas about how minerals build up and how the material’s porosity shifts, both of which matter for waste containment over the long term.</p>
<p>What does all this mean for nuclear waste management? For starters, these more accurate models could replace the outdated simulations currently used to judge the safety of nuclear repositories. That’s critical if the U.S.—or any country—wants to move forward with building a permanent disposal site. The models also have the flexibility to assess different rock types, including salt formations, stretching their usefulness far into the future.</p>
<p>Looking ahead, the researchers plan to refine their simulations even more, possibly integrating machine learning to speed up the process. They’re also eager to gather more data from ongoing experiments, inching closer to a time when storing nuclear waste is a solved problem rather than a looming risk.</p>
<p>For Sarsenbayev and the team, the ultimate goal is straightforward: create a scientifically solid, trustworthy solution for nuclear waste that decision-makers—and the public—can have faith in. As he puts it, their work sits at the intersection of science, systems, and society: a genuinely collaborative approach to one of the most daunting engineering challenges of our era.</p>
<p>Read the full story at <a href="https://news.mit.edu/2025/model-predicts-long-term-effects-nuclear-waste-underground-disposal-systems-0718" target="_blank" rel="noopener">MIT News</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Bridging Educational Gaps in the Developing World through Beneficial AGI: Lessons from Ethiopia</title>
		<link>https://aitrendscenter.eu/bridging-educational-gaps-in-the-developing-world-through-beneficial-agi-lessons-from-ethiopia/</link>
		
		<dc:creator><![CDATA[Max Krawiec]]></dc:creator>
		<pubDate>Tue, 10 Jun 2025 17:31:49 +0000</pubDate>
				<category><![CDATA[News]]></category>
		<category><![CDATA[Tutorial]]></category>
		<guid isPermaLink="false">https://aitrends.center/bridging-educational-gaps-in-the-developing-world-through-beneficial-agi-lessons-from-ethiopia/</guid>

					<description><![CDATA[The Future of AI for a More Inclusive World When people talk about Artificial General Intelligence (AGI), it’s easy to imagine it as a distant, high-tech concept only relevant to the world’s most advanced cities. But in reality, AGI—and its younger sibling, everyday AI technology—has the potential to make the biggest difference where resources have traditionally lagged behind. My time working alongside iCog Labs in Ethiopia, a company co-founded by Ben Goertzel and Getnet Aseffa, has shown me just how much promise and how many unique hurdles there are when nurturing AI projects in developing nations. There’s enormous potential here, [&#8230;]]]></description>
										<content:encoded><![CDATA[<h4>The Future of AI for a More Inclusive World</h4>
<p>
When people talk about Artificial General Intelligence (AGI), it’s easy to imagine it as a distant, high-tech concept only relevant to the world’s most advanced cities. But in reality, AGI—and its younger sibling, everyday AI technology—has the potential to make the biggest difference where resources have traditionally lagged behind. My time working alongside <a href="https://icog-labs.com/" target="_blank" rel="noopener">iCog Labs</a> in Ethiopia, a company co-founded by Ben Goertzel and Getnet Aseffa, has shown me just how much promise and how many unique hurdles there are when nurturing AI projects in developing nations. There’s enormous potential here, but also a lot to learn.
</p>
<h4>How AI Can Help Close Education Gaps</h4>
<p>
AI could be a real game changer for education—especially in places where shortages of teachers, textbooks, and even basic infrastructure are everyday challenges. And if we dig deeper, the obstacles aren’t just about missing blackboards: many children in countries like Ethiopia are expected to learn in languages they barely understand, using materials that don’t reflect their real lives or cultures. The answer isn’t just technology for technology’s sake; instead, it’s about solutions born out of local insights, creativity, and hands-on partnerships.
</p>
<p>
UNESCO estimates that about 40% of children worldwide go to school in a language they don’t fully grasp, which means they’re missing out not just on words, but on all kinds of opportunities. This is where AI-powered language tools shine. While there’s often little profit motivation for global tech giants to support small or under-resourced language communities, targeted projects fill the gap. Alongside <a href="https://icog.ghost.io/curious-reader-using-digital-tools-to-support-early-education-in-ethiopia/" target="_blank" rel="noopener">Curious Learning</a>, we’ve built AI-backed reading apps catering to local Ethiopian languages, now used by over 85,000 people. They’re simple, low-resource tools taming the hardest linguistic challenges faced by so many kids.
</p>
<p>
To dig even deeper, we’ve launched Leyu—a decentralized data collection platform—focused on gathering language samples directly from far-flung villages. The data we collect trains AI translation systems, so another under-represented language doesn’t get left out of tomorrow’s digital conversation.
</p>
<h4>Making Learning Exciting and Relevant</h4>
<p>
Of course, language is only part of the equation. True educational progress depends on content that means something locally. AI has a role here, too: it can generate science lessons about crops local kids see every day or math problems built around the rhythms of their own markets. This is education that clicks, because it’s finally about them.
</p>
<p>
Some of our proudest work has rolled out on the road with Digitruck—a full-blown mobile classroom built into a truck. Thanks, in part, to support from <a href="https://singularitynet.io/" target="_blank" rel="noopener">SingularityNET</a>, Digitruck is decked out with computers and devices, and rolls into some of Ethiopia’s most remote spots. Kids test out coding and even get hands-on experiences with AI—all while staying close to their roots.
</p>
<p>
Looking back, there have also been joyful, unforgettable moments. From 2015 to 2019, the RoboSapiens program had university students programming humanoid robots to play soccer—a blend of culture and high-tech excitement that created instant buzz from Ethiopia to Kenya and Nigeria. Sadly, practical obstacles like expensive import tariffs on electronics put the program on hold, a reminder that technical ambition needs support from strong policy and infrastructure.
</p>
<h4>Rooting AI in Real Community Needs</h4>
<p>
All this energy must be matched by caution and care. As AI makes its way into classrooms and daily life, it’s crucial not to let these tools replace essential skills or undermine teachers. Rather, the aim is to help educators do what they do best—foster curiosity and independent thinking—while supporting them with new resources.
</p>
<p>
It’s not about one-size-fits-all rules. Instead, the communities themselves need to guide what AI does for them. That kind of meaningful, grassroots engagement is what will ensure AI enhances education, rather than becoming just another barrier.
</p>
<p>
Equitable, responsible AI in education isn’t simple or quick. It means working side-by-side with local leaders, keeping cultural nuances front and center, and planning for ethical risks. But it’s worth every bit of effort. When a child in rural Ethiopia codes for the very first time on a Digitruck laptop, that’s when the big themes of equity and empowerment stop being buzzwords—and become real, lived experiences.
</p>
<p>
<a href="https://www.unite.ai/bridging-educational-gaps-in-the-developing-world-through-beneficial-agi-lessons-from-ethiopia/" target="_blank" rel="noopener">Original article on Unite.AI</a></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Why AI Courses Must Teach Students to Recognize Bias in Medical Data</title>
		<link>https://aitrendscenter.eu/why-ai-courses-must-teach-students-to-recognize-bias-in-medical-data/</link>
		
		<dc:creator><![CDATA[Max Krawiec]]></dc:creator>
		<pubDate>Mon, 02 Jun 2025 14:30:00 +0000</pubDate>
				<category><![CDATA[News]]></category>
		<category><![CDATA[Tutorial]]></category>
		<guid isPermaLink="false">https://aitrends.center/why-ai-courses-must-teach-students-to-recognize-bias-in-medical-data/</guid>

					<description><![CDATA[An influx of students pursuing courses in artificial intelligence (AI) for healthcare brightens the prospect of revolutionizing medical diagnostics and treatment recommendations yearly. Despite this excitement, one area of concern remains under-addressed—educating students on the critical importance of assessing the quality and biases inherent in the training data used to develop these AI models. The Unseen Shortcomings in Healthcare Data and the Role of AI Education Wonderfully highlighted by Leo Anthony Celi, an accomplished physician and senior research scientist at MIT&#8217;s Institute for Medical Engineering and Science, this oversight in AI education may create issues down the line. In a [&#8230;]]]></description>
										<content:encoded><![CDATA[<p>An influx of students pursuing courses in artificial intelligence (AI) for healthcare brightens the prospect of revolutionizing medical diagnostics and treatment recommendations yearly. Despite this excitement, one area of concern remains under-addressed—educating students on the critical importance of assessing the quality and biases inherent in the training data used to develop these AI models.</p>
<p><h4>The Unseen Shortcomings in Healthcare Data and the Role of AI Education</h4>
<p>Wonderfully highlighted by Leo Anthony Celi, an accomplished physician and senior research scientist at MIT&#8217;s Institute for Medical Engineering and Science, this oversight in AI education may create issues down the line. In a <a href="https://dl.acm.org/doi/10.1145/3737650" target="_blank" rel="noopener">recent article</a>, he explains how data bias, especially in clinical data primarily collected from white males, might lead to inefficiencies in AI systems when applied to more diverse populations. For instance, pulse oximeters often overestimate oxygen saturation levels in people of color due to their underrepresentation in clinical trials. This is only the tip of the iceberg, with countless more cases where medical equipment and data systems overlook population diversity, producing skewed results and potentially harmful decisions.</p>
<p>Another crucial issue orbits around the use of electronic health records (EHRs) as a basis for AI models. Although an essential part of medicine, EHRs were never intended to serve as learning systems, and they are ripe with inconsistencies and biases. Celi, however, is not all doom and gloom, advocating for ingenious ways of responsibly using the existing data instead of replacing the entire EHR infrastructure—something currently not feasible. Thankfully, innovative approaches such as transformer models are being explored to understand correlations between lab results, vital signs, and treatments better. This fascinating approach could help lessen the impact of missing or biased data, often affected by social determinants of health and implicit provider biases.</p>
<p><h4>Addressing the Imperfections and Maximizing Learning</h4>
<p>The challenges become apparent when you consider Celi&#8217;s experiences teaching AI in healthcare. Since starting in 2016, his MIT team realized that students were being taught to optimize models in terms of statistical performance rather than question the data&#8217;s integrity. A review of 11 online courses revealed the scale of this problem; only five courses talked about data bias and a mere two offered substantial discussions on the subject. As AI continues to establish its footprint in healthcare, the onus lies with educators to ensure students can not only build models but also scrutinize the data fueling them. Bridging this divide will require a shift in focus from solely model building to also understanding the data &#8211; an area Celi believes should account for at least half the course content.</p>
<p>One initiative helping to solve this puzzle is the MIT Critical Data consortium. They&#8217;ve been hosting international datathons since 2014. These sessions unite clinicians, data scientists, and healthcare professionals to collaboratively examine local datasets, aiming to understand health and disease within the unique cultural and systematic context of each region. These collaborations inspire an environment where critical thinking organically thrives.</p>
<p>Embracing the imperfections in data can also be a step towards improvement, albeit a challenging one. A good example is the MIMIC database, which took over 10 years to formulate a usable schema, largely due to users acknowledging and pointing out its flaws. An insightful reminder from Celi at this point is that even without all the answers, inspiring people to start asking the right questions can be a game-changer. As students and researchers engage with AI development in healthcare, they need to remain aware of its transformative potential and the ethical responsibilities that come with that.</p>
<p>For a more in-depth discussion on this subject with Leo Anthony Celi, visit <a href="https://news.mit.edu/2025/3-questions-recognizing-potential-bias-in-ai-datasets-0602" target="_blank" rel="noopener">MIT News</a>.</p>
]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>
