<?xml version="1.0" encoding="UTF-8" ?><!-- generator=Zoho Sites --><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><atom:link href="https://www.nownextlater.ai/Insights/tag/weakly-supervised-learning/feed" rel="self" type="application/rss+xml"/><title>Now Next Later AI - Blog #Weakly supervised learning</title><description>Now Next Later AI - Blog #Weakly supervised learning</description><link>https://www.nownextlater.ai/Insights/tag/weakly-supervised-learning</link><lastBuildDate>Wed, 26 Nov 2025 21:24:10 +1100</lastBuildDate><generator>http://zoho.com/sites/</generator><item><title><![CDATA[Weakly Supervised AI: Overpromising, Underdelivering]]></title><link>https://www.nownextlater.ai/Insights/post/weakly-supervised-ai-overpromising-underdelivering</link><description><![CDATA[Weakly supervised learning is a popular technique. The goal is to alleviate the costly data annotation bottleneck. But new research suggests these methods may be significantly overstating their capabilities.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_kHlMHFhcSHWm0l6M_LA5Rg" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_dejLJsr8QAqCAHwG_pfRTw" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_Cza8iv2yQ26ed5vINlmcoQ" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_8rj2V4DLyvWmVToQqItlCQ" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_8rj2V4DLyvWmVToQqItlCQ"] .zpimage-container figure img { width: 800px ; height: 655.66px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_8rj2V4DLyvWmVToQqItlCQ"] .zpimage-container figure img { width:500px ; height:409.79px ; } } @media (max-width: 767px) { [data-element-id="elm_8rj2V4DLyvWmVToQqItlCQ"] .zpimage-container figure img { width:500px ; height:409.79px ; } } [data-element-id="elm_8rj2V4DLyvWmVToQqItlCQ"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-large zpimage-tablet-fallback-large zpimage-mobile-fallback-large hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-08%20at%207.06.14%20pm.png" width="500" height="409.79" loading="lazy" size="large" alt="Relative performance gain over weak la- bels when varying validation conditions." data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_yZHFit1TQ2Sg3YpfPPO74g" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_yZHFit1TQ2Sg3YpfPPO74g"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Weakly supervised learning is a popular technique in AI research that aims to train models using noisy, imperfect data instead of clean, human-labeled data. The goal is to alleviate the costly data annotation bottleneck. But new research suggests these methods may be significantly overstating their capabilities.</p><p><br></p><p>Researchers from Saarland University thoroughly evaluated several widely-used weakly supervised learning techniques on diverse natural language tasks. They found that with no clean data available, the methods completely failed to improve over basic weak labels. More concerning, commonly used clean validation data for model selection could be more efficiently leveraged by simply training models directly on it instead.</p><p><br></p><p>On benchmark datasets, sophisticated weakly supervised approaches were handily outperformed by fine-tuning models on as little as 5-10 clean examples per class. Even combining weak and limited clean data, a simple two-stage fine-tuning approach worked best. The purported benefits of complex weakly supervised methods largely vanished in realistic low-data situations.</p><p><br></p><p>The findings cast doubt on claims of progress in weak supervision and reveal flaws in how the techniques are evaluated. The reliance on sizable validation data underscores their limitations for real-world application. While weak supervision remains appealing for bypassing human labeling, current techniques seem to overpromise.</p><p><br></p><p>For business leaders, this research highlights pitfalls of trusting &quot;state-of-the-art&quot; AI before rigorously testing if it applies to your use case and data constraints. Clever weakly supervised approaches may not live up to claims on small datasets. But when designed and validated appropriately, they can still unlock unique value. Disillusionment with AI is often simply due to mismatch with practical realities. Aligning research with real-world needs is vital to deliver robust AI that businesses can confidently leverage.</p><p><br></p><p>Sources:</p><p><a href="https://arxiv.org/abs/2305.17442" title="arxiv" rel="">arxiv</a><br></p><p></p></div><p></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 07:59:24 +1000</pubDate></item></channel></rss>