<?xml version="1.0" encoding="UTF-8" ?><!-- generator=Zoho Sites --><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><atom:link href="https://www.nownextlater.ai/Insights/tag/stochastic-parrots/feed" rel="self" type="application/rss+xml"/><title>Now Next Later AI - Blog #stochastic parrots</title><description>Now Next Later AI - Blog #stochastic parrots</description><link>https://www.nownextlater.ai/Insights/tag/stochastic-parrots</link><lastBuildDate>Wed, 26 Nov 2025 21:23:39 +1100</lastBuildDate><generator>http://zoho.com/sites/</generator><item><title><![CDATA[Measuring the Truthfulness of Large Language Models: Benchmarks, Challenges, and Implications for Business Leaders]]></title><link>https://www.nownextlater.ai/Insights/post/Measuring-the-Truthfulness-of-Large-Language-Models</link><description><![CDATA[<img align="left" hspace="5" src="https://www.nownextlater.ai/Screenshot 2024-04-29 at 12.56.35 pm.png"/>LLMs currently face significant challenges when it comes to truthfulness. Understanding these limitations is essential for any business considering leveraging LLMs.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_GVCHdVe6Q5O7K4Wm7HPPvg" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_b1wUWYpxS3yvxfwmmwjHMQ" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_n7uk0luTQAe-ggvG9QuZqg" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_As1QnLrVLRrkRjbbxVG6lw" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_As1QnLrVLRrkRjbbxVG6lw"] .zpimage-container figure img { width: 500px ; height: 477.31px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_As1QnLrVLRrkRjbbxVG6lw"] .zpimage-container figure img { width:500px ; height:477.31px ; } } @media (max-width: 767px) { [data-element-id="elm_As1QnLrVLRrkRjbbxVG6lw"] .zpimage-container figure img { width:500px ; height:477.31px ; } } [data-element-id="elm_As1QnLrVLRrkRjbbxVG6lw"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202024-04-29%20at%2012.56.35%E2%80%AFpm.png" width="500" height="477.31" loading="lazy" size="medium" alt="LLM misinformation" data-lightbox="true"/></picture></span><figcaption class="zpimage-caption zpimage-caption-align-center"><span class="zpimage-caption-content">LLM Misinformation</span></figcaption></figure></div>
</div><div data-element-id="elm_PUyJ5oo1S8KD7u631BYLWA" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_PUyJ5oo1S8KD7u631BYLWA"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-center " data-editor="true"><div style="color:inherit;text-align:left;"><p>In recent years, large language models (LLMs) like GPT-3, ChatGPT, and others have made stunning breakthroughs in natural language processing. These powerful AI systems can engage in human-like conversations, answer questions, write articles, and even generate code. Their potential to transform industries from customer service to content creation has captured the imagination of business leaders worldwide.</p><p><br></p><p>However, as companies rush to adopt LLMs, a critical question often goes overlooked - just how truthful and reliable are these systems? Can we trust the outputs of LLMs to be factual and free of misinformation or deception? As it turns out, LLMs currently face significant challenges when it comes to truthfulness. Understanding these limitations is essential for any business considering leveraging LLMs.</p><p><br></p><p><span style="font-family:&quot;Archivo Black&quot;, sans-serif;">The Hallucination Problem&nbsp;</span></p><p><br></p><p>One of the biggest issues with LLMs today is their tendency to &quot;hallucinate&quot; information - that is, to generate content that seems plausible but is not actually true. Because LLMs are trained on vast amounts of online data, they can pick up and parrot back common misconceptions, outdated facts, biases and outright falsehoods mixed in with truth.</p><p><br></p><p>An LLM may confidently assert something that sounds right but does not match reality. For example, an LLM might claim a fictional event from a book or movie actually happened in history. Or it may invent realistic-sounding but untrue details when asked about a topic it lacks knowledge of.</p><p><br></p><p>LLMs do not have a true understanding of the information they process - they work by recognizing and reproducing patterns of text. So they can combine ideas in seemingly coherent but inaccurate ways. This makes it difficult to always separate LLM fact from fiction.</p><p><br></p><p><span style="font-family:&quot;Archivo Black&quot;, sans-serif;">Benchmarking LLM Truthfulness&nbsp;</span></p><p><br></p><p>To quantify just how prone LLMs are to truthful vs untruthful outputs, AI researchers have developed benchmark datasets to test these models. Two notable examples are:</p><ol><li><a href="https://arxiv.org/abs/2109.07958" title="TruthfulQA" rel="">TruthfulQA</a> (2022) - Contains 817 questions designed to elicit false answers that mimic human misconceptions across topics like health, law, and finance. Models are scored on how often they generate truthful responses.</li><li><a href="https://arxiv.org/abs/2305.11747" title="HaluEval" rel="">HaluEval</a> (2023) - Includes 35,000 examples of human-annotated or machine-generated &quot;hallucinated&quot; outputs for models to detect, across user queries, Q&amp;A, dialog and summarization. Measures model ability to discern truthful vs untruthful text.</li></ol><p><br></p><p>When tested on these benchmarks, even state-of-the-art LLMs struggle with truthfulness:</p><ul><li>On TruthfulQA, the best model was truthful only 58% of the time (vs 94% for humans). Larger models actually scored worse.</li><li>On HaluEval, models frequently failed to detect hallucinations, with accuracy barely above random chance in some cases. Hallucinated content often covered entities and topics the models lacked knowledge of.</li></ul><p><br></p><p>While providing knowledge or adding reasoning steps helped models somewhat, truthfulness remains an unsolved challenge. Models today are not reliable oracles of truth.</p><p><br></p><p><span style="font-family:&quot;Archivo Black&quot;, sans-serif;">Implications for Businesses&nbsp;</span></p><p><br></p><p>The current limitations of LLMs in generating consistently truthful outputs has major implications for their practical use in business:</p><ol><li>Careful human oversight of LLM content is a must. Outputs cannot be blindly trusted as true without verification from authoritative sources.</li><li>LLMs are not suitable for high-stakes domains like healthcare, finance, or legal advice where inaccuracies pose unacceptable risks. More narrow, specialized and validated knowledge bases are needed.</li><li>Using LLMs for content generation requires clear disclosure that output may not be entirely factual. Audiences should be informed on the role and limitations of AI.</li><li>&quot;Prompt engineering&quot; and other filtering techniques to coax more truthful responses have limits. Changes to underlying training data and architectures are needed for major improvements.</li></ol><p><br></p><p>As research continues to progress, we can expect to see more truthful and dependable LLMs over time. Providing models with curated factual knowledge, better reasoning abilities, and alignment with human values are promising directions.</p><p><br></p><p>But for now, business leaders eager to harness the power of LLMs must temper their expectations around truthfulness. Treating these AIs as helpful assistants to augment and accelerate human knowledge work, while keeping a human in the loop to validate outputs, is the prudent approach. The truth is, LLMs still have a ways to go before they can be fully trusted as reliably truthful.</p></div><p></p></div>
</div><div data-element-id="elm_TIg9LdfKuCtOhXsFxtQFeA" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_TIg9LdfKuCtOhXsFxtQFeA"] .zpimage-container figure img { width: 500px ; height: 500.00px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_TIg9LdfKuCtOhXsFxtQFeA"] .zpimage-container figure img { width:500px ; height:500.00px ; } } @media (max-width: 767px) { [data-element-id="elm_TIg9LdfKuCtOhXsFxtQFeA"] .zpimage-container figure img { width:500px ; height:500.00px ; } } [data-element-id="elm_TIg9LdfKuCtOhXsFxtQFeA"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium "><figure role="none" class="zpimage-data-ref"><a class="zpimage-anchor" href="/introduction-to-large-language-models-for-business-leaders-book" target="" rel=""><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/12.png" width="500" height="500.00" loading="lazy" size="medium" alt="Intro to LLMs for Business Leaders"/></picture></a></figure></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Mon, 29 Apr 2024 13:00:10 +1000</pubDate></item><item><title><![CDATA[The Risks of Ever-Larger AI Language Models]]></title><link>https://www.nownextlater.ai/Insights/post/the-risks-of-ever-larger-ai-language-models</link><description><![CDATA[A thought-provoking paper from computer scientists raises important concerns about the AI community's pursuit of ever-larger language models.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_Z8F3OtCsQCWRTAJ4k7C-mQ" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm__EH9mVpOTaKqsM990ikoag" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_KSuxIIF1QLyRqNzFLRN03A" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_0T9Fm5FVTLW5nrYOa10dWQ" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_0T9Fm5FVTLW5nrYOa10dWQ"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><div style="color:inherit;">A thought-provoking paper from computer scientists raises important concerns about the AI community's pursuit of ever-larger language models. It argues this dominant research direction has significant downsides and risks that demand urgent attention.</div><div style="color:inherit;"><br><div><p>In recent years, models like Google's BERT, OpenAI's GPT-3, and others have achieved impressive performance gains in language tasks through scaling up to hundreds of billions of parameters trained on massive text datasets. However, the authors argue the environmental, financial, and social costs of this approach outweigh the benefits, and more focus should go towards better understanding models rather than simply making them bigger.</p><p><br></p><p>On the environmental front, training these models requires prohibitive amounts of computing power, racking up massive carbon footprints. This compounds inequality when the benefits accrue mainly to wealthy nations but the environmental consequences are borne globally. The financial costs of training also centralize progress in a few well-resourced labs.</p><p><br></p><p>The authors also highlight problems with training data. Web-scale datasets amplify dominant viewpoints and encode harmful biases against marginalized groups. Attempting to filter out toxic content is insufficient and risks suppressing minority voices. More investment is needed in thoughtful data curation versus simply amassing unfathomable quantities.</p><p><br></p><p>Additionally, while larger models post impressive scores on NLP leaderboards, they don't actually perform true language understanding. Their inner workings remain opaque and they succeed by picking up on spurious statistical cues. This risks misdirecting research efforts away from real progress on AI interpretability and accountability.</p><p><br></p><p>When deployed, huge models can generate remarkably fluent but meaningless and incoherent text. The authors liken them to &quot;stochastic parrots&quot; given their tendency to amplify toxic patterns in training data. The term refers to how these models randomly stitch together linguistic patterns they have observed, without any grounding in meaning or intent. If people interpret their outputs as credible despite lack of grounding, it enables spreading misinformation and abuse.</p><p><br></p><p>Given these downsides, the authors advocate rethinking the goal of ever-larger models. They recommend prioritizing energy efficiency, curating training data carefully, engaging stakeholders to shape ethical systems, and exploring alternative research directions not dependent on unfathomable data quantities.</p><p><br></p><p>While large models can sometimes benefit applications like speech recognition, risks need balancing with harm mitigation measures like watermarking their outputs. Overall, the paper compellingly argues that continuing blindly on the path of scaling up carries severe risks that require urgent attention.</p><p><br></p><p>This paper became controversial when some authors published it while working at Google Research. Google allegedly requested they withdraw the paper for internal review, then fired several of the co-authors, including well-known AI ethics researcher Timnit Gebru. The incident highlighted risks of speaking out against dominant research paradigms, especially when papers critique an employer's technology direction. It increased scrutiny of research freedom and ethics in AI.</p></div></div><p><br></p><p>Sources:</p><div style="color:inherit;"><div><div><div><p><a href="https://dl.acm.org/doi/10.1145/3442188.3445922" title="On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? " rel="">On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? </a></p><p></p></div>
</div></div></div><div style="color:inherit;"><div style="color:inherit;"><div><div><div><p>Emily M. Bender,&nbsp; Timnit Gebru</p><div style="color:inherit;"><div><div><div><p><span style="font-size:12pt;"><br></span></p></div>
</div></div></div></div></div></div></div></div><p></p></div><p></p></div></div></div>
</div></div></div></div> ]]></content:encoded><pubDate>Sun, 13 Aug 2023 21:46:24 +1000</pubDate></item></channel></rss>