<?xml version="1.0" encoding="UTF-8" ?><!-- generator=Zoho Sites --><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><atom:link href="https://www.nownextlater.ai/Insights/tag/explainability/feed" rel="self" type="application/rss+xml"/><title>Now Next Later AI - Blog #Explainability</title><description>Now Next Later AI - Blog #Explainability</description><link>https://www.nownextlater.ai/Insights/tag/explainability</link><lastBuildDate>Wed, 26 Nov 2025 21:38:47 +1100</lastBuildDate><generator>http://zoho.com/sites/</generator><item><title><![CDATA[Unlocking the Power of Interpretable AI with InterpretML: A Guide for Business Leaders]]></title><link>https://www.nownextlater.ai/Insights/post/unlocking-the-power-of-interpretable-ai-with-interpretml-a-guide-for-business-leaders</link><description><![CDATA[<img align="left" hspace="5" src="https://www.nownextlater.ai/interactive-visualization-dashboard-icon.png"/>InterpretML is a valuable tool for unlocking the power of interpretable AI in traditional machine learning models. While it may have limitations when it comes to directly interpreting LLMs, the principles of interpretability and transparency remain crucial in the age of generative AI.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_vv2ADhj1T_qpnc2ndrUpmQ" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_8YFDElmaRhqHfl1XzkasfA" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_6_hvn9CZSByLATJKzfuW8A" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_mylyszm4zJaJlxC7fPV1cw" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_mylyszm4zJaJlxC7fPV1cw"] .zpimage-container figure img { width: 648px !important ; height: 356px !important ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_mylyszm4zJaJlxC7fPV1cw"] .zpimage-container figure img { width:648px ; height:356px ; } } @media (max-width: 767px) { [data-element-id="elm_mylyszm4zJaJlxC7fPV1cw"] .zpimage-container figure img { width:648px ; height:356px ; } } [data-element-id="elm_mylyszm4zJaJlxC7fPV1cw"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-original zpimage-tablet-fallback-original zpimage-mobile-fallback-original hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/interactive-visualization-dashboard-icon.png" width="648" height="356" loading="lazy" size="original" alt="InterpretML Interactive Visualisation Dashboard" data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_jS7fiVqiTGmG8PihrE5y1A" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_jS7fiVqiTGmG8PihrE5y1A"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-center " data-editor="true"><div style="color:inherit;text-align:left;"><p style="font-weight:400;text-indent:0px;">In today's fast-paced business world, artificial intelligence (AI) has become a game-changer, enabling organizations to make data-driven decisions and gain a competitive advantage. However, as machine learning models grow more complex, the need for transparency and interpretability becomes increasingly important. <a href="https://interpret.ml/" title="InterpretML" rel="">InterpretML</a>, an open-source Python package developed by Microsoft, empowers businesses to explain and understand the behavior of their AI models. In this article, we will explore the key capabilities and benefits of InterpretML, discuss its limitations when it comes to interpreting advanced language models, and delve into the current research efforts in the field of interpretability for generative AI.</p><p style="font-weight:400;text-indent:0px;"></p><p style="font-weight:400;text-indent:0px;"><br></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:24px;color:rgb(41, 77, 135);">Key Capabilities of InterpretML</span></p><p style="font-weight:400;text-indent:0px;"></p><ol><li><span><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Global and Local Explanations: </span>InterpretML offers a comprehensive set of tools to explain model behavior at both high-level (global) and individual (local) perspectives. Global explanations provide insights into overall patterns and trends, allowing business leaders to grasp the general decision-making process of their models. Local explanations, on the other hand, focus on specific predictions, enabling a detailed analysis of individual cases. This dual approach empowers organizations to gain a holistic understanding of their AI systems.</span></li><li><span><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Compatibility with Various Models:</span> One of the standout features of InterpretML is its ability to work with a wide range of machine learning models, including decision trees, linear models, neural networks, random forests, gradient boosting machines, and support vector machines. This versatility ensures that businesses can apply interpretation techniques to their existing AI workflows while enhancing transparency and interpretability.</span></li><li><span><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Feature Importance and What-If Scenarios: </span>InterpretML provides powerful techniques to identify the most influential factors in a model's predictions. By determining the importance of different features, business leaders can gain valuable insights into the key drivers behind the model's decisions. Additionally, InterpretML can generate &quot;what-if&quot; scenarios, showing how changes in input features would impact the model's output. This capability allows organizations to explore different possibilities and make informed decisions based on the model's behavior.</span></li><li><span><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Clear Visualizations:</span> Effective communication is crucial when it comes to interpreting and explaining AI models. InterpretML recognizes this need and offers a range of visualization tools to present explanations in a clear and accessible manner. From feature importance plots to graphs showing the model's behavior, these visualizations help business leaders and stakeholders understand the inner workings of their AI systems without requiring deep technical expertise.</span></li></ol><div><br></div><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:24px;color:rgb(41, 77, 135);">Limitations of InterpretML with Advanced Language Models</span></p><p style="font-weight:400;text-indent:0px;"></p><p style="font-weight:400;text-indent:0px;"><br></p><p style="font-weight:400;text-indent:0px;">While InterpretML is a powerful tool for interpreting various types of machine learning models, it may have limitations when it comes to explaining the behavior of advanced language models, such as GPT-3, BERT, and T5. These models, known as large language models (LLMs) or transformers, are highly complex and have millions or billions of parameters. Their intricate inner workings and decision-making processes can be challenging to interpret due to their scale and complexity.</p><p style="font-weight:400;text-indent:0px;"><br></p><p style="font-weight:400;text-indent:0px;">InterpretML's techniques, such as SHAP (SHapley Additive exPlanations) and LIME (Local Interpretable Model-agnostic Explanations), are primarily designed for interpreting more traditional machine learning models. SHAP assigns importance scores to each feature based on its contribution to the model's prediction, while LIME generates local explanations by approximating the model's behavior around a specific instance using a simpler, interpretable model. These techniques may not directly translate to the complexities of LLMs and transformers, which have more sophisticated architectures and capture nuanced patterns in natural language.</p><p style="font-weight:400;text-indent:0px;"><br></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:24px;color:rgb(41, 77, 135);">Current Research in Interpretability for Generative AI</span></p><p style="font-weight:400;text-indent:0px;"></p><p style="font-weight:400;text-indent:0px;"><br></p><p style="font-weight:400;text-indent:0px;">Although InterpretML may not be the perfect fit for interpreting LLMs and transformers, the field of interpretability for advanced language models is an active area of research. Scientists and researchers are developing new techniques specifically tailored to understanding and explaining the behavior of these models. Some of the current research efforts include:</p><ol><li><span><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Attention Analysis:</span> Researchers are studying the attention mechanisms of transformer models to understand which parts of the input the model focuses on during prediction. By visualizing and analyzing these attention patterns, we can gain insights into how the model processes and prioritizes information.</span></li><li><span><span><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Probing Tasks: </span>Designing specific tasks to test the model's understanding of language properties, such as grammar, meaning, and common sense, can help uncover the knowledge and capabilities of LLMs. These probing tasks provide a targeted evaluation of the model's behavior and decision-making process.</span></span></li><li><span><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Perturbation-based Methods:</span> By slightly modifying the input or internal representations of the model and observing how the outputs change, researchers can gain insights into the model's sensitivity to specific changes and its decision-making process. Perturbation-based methods help identify the most influential factors in the model's predictions.</span></li><li><span><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Interpretable Architectures: </span>Some researchers are exploring the development of new architectures for LLMs and transformers that are inherently more interpretable. By designing models with built-in interpretability mechanisms, such as attention-based explanations or modular components, we can achieve a better understanding of their inner workings.</span></li><li><span><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Other Approaches:</span> Researchers are also investigating techniques such as layer-wise relevance propagation (LRP), which assigns relevance scores to input features based on their contribution to the model's output, and integrated gradients, which attribute the model's prediction to input features by calculating the path integral of the gradients.</span></li></ol><p style="font-weight:400;text-indent:0px;"><br></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:24px;color:rgb(41, 77, 135);">The Importance of Interpretability in the Age of Generative AI</span></p><p style="font-weight:400;text-indent:0px;"></p><p style="font-weight:400;text-indent:0px;"><br></p><p style="font-weight:400;text-indent:0px;">As generative AI models become more prevalent and influential in various industries, the need for transparency and accountability becomes paramount. These models have the potential to generate human-like text, images, and even code, revolutionizing the way businesses operate. However, the complexity and autonomy of generative AI models also raise concerns about biased outputs or potential misuse.</p><p style="font-weight:400;text-indent:0px;"><br></p><p style="font-weight:400;text-indent:0px;">Interpretability plays a crucial role in mitigating these risks and building trust in AI systems. By providing clear explanations of how models arrive at their outputs, businesses can ensure fairness, detect and address biases, and maintain ethical standards. Interpretability also enables organizations to comply with regulatory requirements and demonstrate the reasoning behind AI-driven decisions.</p><p style="font-weight:400;text-indent:0px;"><br></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:24px;color:rgb(41, 77, 135);">Key Takeaways for Business Leaders</span></p><p style="font-weight:400;text-indent:0px;"><br></p><p style="font-weight:400;text-indent:0px;">InterpretML is a valuable tool for unlocking the power of interpretable AI in traditional machine learning models. While it may have limitations when it comes to directly interpreting advanced language models, the broader principles of interpretability and transparency remain crucial in the age of generative AI. As research in this field advances, business leaders should stay informed about the latest developments and adopt new tools and techniques that enable them to explain and understand the behavior of their AI systems. By prioritizing interpretability and transparency, organizations can build trust, mitigate risks, ensure compliance, and harness the full potential of AI technologies while maintaining ethical and responsible practices.</p></div></div>
</div><div data-element-id="elm_IbU3obgEOW8xtM5pyFMaqg" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_IbU3obgEOW8xtM5pyFMaqg"] .zpimage-container figure img { width: 500px ; height: 500.00px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_IbU3obgEOW8xtM5pyFMaqg"] .zpimage-container figure img { width:500px ; height:500.00px ; } } @media (max-width: 767px) { [data-element-id="elm_IbU3obgEOW8xtM5pyFMaqg"] .zpimage-container figure img { width:500px ; height:500.00px ; } } [data-element-id="elm_IbU3obgEOW8xtM5pyFMaqg"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium "><figure role="none" class="zpimage-data-ref"><a class="zpimage-anchor" href="/responsible-ai-in-the-age-of-generative-models-ai-governance-ethics-and-risk-management" target="" rel=""><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Navy%20and%20Blue%20Modern%20We%20Provide%20Business%20Solutions%20Facebook%20Ad%20-1200%20x%201200%20px-.png" width="500" height="500.00" loading="lazy" size="medium"/></picture></a></figure></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 04 Apr 2024 12:45:20 +1100</pubDate></item><item><title><![CDATA[Why AI is Harder Than We Think: Key Takeaways for Business Leaders]]></title><link>https://www.nownextlater.ai/Insights/post/Why-AI-is-Harder-Than-We-Think-Key-Takeaways-for-Business-Leaders</link><description><![CDATA[A new study has shown that transformers can be expressed in a simple logic formalism. This finding challenges the perception that transformers are inscrutable black boxes and suggests avenues for interpreting how they work.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_sfUGQRrPRMOb-zqryhN6kA" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_aI0rE4RHRAGxFiFOSq2e8g" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_KQc6QKC-TUS9k5ADouakpw" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_lLX8ye2SQte64RLyZASIwA" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_lLX8ye2SQte64RLyZASIwA"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Artificial intelligence (AI) has seen remarkable advances in recent years, from self-driving cars to defeating humans at games like chess and Go. However, leading AI expert Melanie Mitchell argues that achieving truly human-like &quot;general&quot; AI is much harder than many experts predict. Here are the key takeaways from her paper for business leaders:</p><ul><li>Narrow AI versus general AI: Success in specialized applications like chess or image recognition does not necessarily translate to broader capabilities. We cannot assume today's AI systems are on a steady path to human-level intelligence.</li><li>&quot;Easy&quot; things are hard: Basic human skills like perceiving the world and carrying on a conversation have proven very difficult to replicate in machines. Conversely, AI can excel at things that are quite hard for humans.</li><li>Wishful vocabulary: Terms like &quot;learn,&quot; &quot;understand,&quot; and &quot;think&quot; are often applied to today's AI, but these systems do not have the same underlying capabilities as humans. Using human vocabulary can misleadingly imply advanced intelligence.</li><li>Intelligence is embodied: Human intelligence relies heavily on our experiences and interactions using our entire bodies, not just abstract reasoning. Attempts to achieve human-level intelligence must consider embracing human-like bodies and environments.</li><li>Common sense is key: To operate successfully in the real world, AI needs the vast background knowledge humans accumulate about how the world works. We still do not understand how to enable machines to acquire this &quot;common sense.&quot;</li></ul><p><br></p><p>Key implications:</p><ul><li>Avoid overconfidence about timelines for achieving human-level AI based on hype and narrow successes. True general intelligence likely remains far off.</li><li>Focus investment on applications of existing AI capabilities, not attempts to replicate human thinking. Manage expectations of near-term outcomes.</li><li>Monitor advances in embodied AI and research on common sense reasoning as indicators of progress toward general AI.</li><li>Ensure AI systems have transparent workings, clear objectives, and human oversight. The orthogonality thesis that intelligence can be coupled to any goals does not hold for human-like general intelligence.</li></ul><p><br></p><p>The path to human-level AI is long with much still unknown. By avoiding unfounded assumptions and acknowledging the challenges ahead, business leaders can make wise strategic decisions about how to apply AI technology today and anticipate what may come tomorrow.</p><p><br></p><p>Sources:</p><div style="color:inherit;"><div><div><div><div><div><p><span style="font-size:14px;"><a href="https://arxiv.org/abs/2104.12871" title="Why AI is Harder Than We Think " rel="">Why AI is Harder Than We Think </a></span></p><p></p><span style="font-size:14px;"></span><p><span style="font-size:14px;">Melanie Mitchell </span></p></div></div>
</div></div></div></div><p></p></div><p></p></div></div></div></div></div></div></div>
 ]]></content:encoded><pubDate>Sun, 13 Aug 2023 19:50:20 +1000</pubDate></item><item><title><![CDATA[Transformers Expressible in Simple Logic]]></title><link>https://www.nownextlater.ai/Insights/post/Transformers-Expressible-in-Simple-Logic</link><description><![CDATA[A new study has shown that transformers can be expressed in a simple logic formalism. This finding challenges the perception that transformers are inscrutable black boxes and suggests avenues for interpreting how they work.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_vTZ0UtxrTVKUpnRBcGFGwQ" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_OK0gUtTtTG-p2y_MzdYVpA" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_JUny_qLdSUqolf8oNtk3FQ" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_-muaciReSB2TX7jNYu8tFQ" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_-muaciReSB2TX7jNYu8tFQ"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>A new study from New York University and the Allen Institute for AI has shown that large language models called transformers can be expressed in a simple logic formalism. This finding challenges the perception that transformers are inscrutable black boxes and suggests avenues for interpreting how they work.</p><p><br></p><p>Transformers are a type of neural network behind major AI achievements like chatbots and language translation. They are trained on massive datasets to generate human-like text. Despite their impressive capabilities, how transformers arrive at their outputs has remained poorly understood.</p><p><br></p><p>The researchers proved transformers can be translated into symbolic logic sentences that replicate their function. Specifically, they showed transformers fit within a logic called first-order logic with majority quantifiers. This logic allows logical sentences with familiar constructs like &quot;AND&quot;, &quot;OR&quot;, and &quot;IF-THEN&quot;, as well as majority quantifiers that check if a condition holds for over half of the elements.</p><p><br></p><p>While real-world transformers are complex neural networks, this study theoretically shows their reasoning can be captured by simple logical expressions. For instance, the logic could recognize patterns like &quot;three As followed by three Bs&quot;, which transformers are known to identify.</p><p><br></p><p>The findings disprove the notion that transformers are inscrutable black boxes. Instead, they suggest transformers implement a form of reasoning not radically different from familiar logical formalisms. The possibility of expressing transformers in interpretable logic could enable explaining how they arrive at outputs, like detecting biases.</p><p><br></p><p>For business leaders deploying AI, this research opens possibilities for making transformers more transparent and accountable. It provides a path toward debugging models to avoid failures or bias. The ability to translate transformers into logical sentences could allow systematically checking if undesirable reasoning patterns occur.</p><p><br></p><p>Overall, this theoretical advance challenges prevailing views of transformers as hopelessly opaque. It demonstrates their thinking can be characterized in understandable logic, unlocking new ways for technologists to interpret these increasingly critical AI models. The research brings transformers closer to human-level reasoning by showing their outputs are not ineffable, but rather can be explained through logic.</p><p><br></p><p>Sources:</p><div style="color:inherit;"><div><div><div><div><p><a href="https://arxiv.org/abs/2210.02671" title="A Logic for Expressing Log-Precision Transformers " rel="">A Logic for Expressing Log-Precision Transformers </a><br></p><p></p><div style="color:inherit;"><div><div><div><div><p>William Merrill and Ashish Sabharwal </p><div style="color:inherit;"><div><div><div><span style="font-size:10pt;font-weight:500;"><br></span></div>
</div></div></div></div></div></div></div></div><p></p></div></div></div></div></div><p></p></div><p></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Sun, 13 Aug 2023 19:50:20 +1000</pubDate></item><item><title><![CDATA[DisentQA: Catching Knowledge Gaps and Avoiding Misleading Users]]></title><link>https://www.nownextlater.ai/Insights/post/enabling-ai-to-untangle-different-knowledge-sources</link><description><![CDATA[Building QA Systems that catch knowledge gaps and avoid misleading users.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_ewF7pMN9Q_eczUOQpCYtUA" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_XdQfIANyTi-5Z3w2LSGv-A" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_tGWqJgjLSlyldj1XkXMcGw" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_KipIDvLOVMb6oIC8bF9TkA" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_KipIDvLOVMb6oIC8bF9TkA"] .zpimage-container figure img { width: 500px ; height: 486.01px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_KipIDvLOVMb6oIC8bF9TkA"] .zpimage-container figure img { width:500px ; height:486.01px ; } } @media (max-width: 767px) { [data-element-id="elm_KipIDvLOVMb6oIC8bF9TkA"] .zpimage-container figure img { width:500px ; height:486.01px ; } } [data-element-id="elm_KipIDvLOVMb6oIC8bF9TkA"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-12%20at%209.09.37%20am.png" width="500" height="486.01" loading="lazy" size="medium" alt="Example outputs from our disentangled QA model on the Natural Questions dataset. " data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_TNbKqQ17TP256B60EqRP7w" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_TNbKqQ17TP256B60EqRP7w"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><div style="color:inherit;"><p>Imagine you ask your phone &quot;Who wrote the song Hello by Adele?&quot; and it gives you an incorrect answer, insisting the song is by Taylor Swift. This shows artificial intelligence sometimes confuses its own training knowledge with external facts.</p><p><br></p><p>Researchers want to fix this issue to make AI assistants more helpful and honest. Their solution: <span style="color:inherit;">Build QA Systems that catch knowledge gaps and avoid misleading users by </span>teaching the system to provide two responses:</p><ol><li>The factual answer based on given information (e.g. Adele)</li><li>What it privately recalls from its memory (e.g. Taylor Swift)</li></ol><p><br></p><p>This highlights any mismatches between its training knowledge and external data. It's like when we say &quot;Hmm, I thought X, but the website says Y.&quot;</p><p><br></p><p>The team trained the AI model by creating quizzes with tricky examples:</p><ul><li>Swapping names in passages to elicit different responses from the context vs. the model's recollection</li><li>Removing passages altogether so the system must say &quot;I don't know&quot;</li></ul><p><br></p><p>After this special training, the model reliably distinguished its own knowledge from given facts. This improved its accuracy and truthfulness.</p><p><br></p><p>Say you ask about a movie release date. The system can now respond:</p><p><span style="font-style:italic;">&quot;The article says July 2022. But I thought it was December 2022.&quot;</span></p><p><br></p><p>This catches any knowledge gaps and avoids misleading users.</p><p><br></p><p>While not perfect, it's major progress toward AI that collaborates in a transparent, helpful manner. The benefits for businesses are clear:</p><ul><li>Avoid frustrated users with incorrect responses</li><li>Build trust by exposing limitations upfront</li><li>Reduce risk from applying flawed knowledge</li><li>Clarify when external data should override internal beliefs</li></ul><p><br></p><p>By recognizing and sharing when its knowledge is incomplete, the AI becomes a more reliable and honest partner. This research brings us closer to truly cooperative human-AI interaction.</p><p><br></p><p>Sources:</p><p><span style="color:inherit;"><a href="https://arxiv.org/pdf/2211.05655.pdf" title="DisentQA: Disentangling Parametric and Contextual Knowledge with Counterfactual Question Answering" rel="">DisentQA: Disentangling Parametric and Contextual Knowledge with Counterfactual Question Answering</a></span></p><p></p></div>
</div><p></p></div></div></div></div></div></div></div> ]]></content:encoded><pubDate>Sat, 12 Aug 2023 09:22:46 +1000</pubDate></item><item><title><![CDATA[The Future of AI Language Models: Making Them More Interpretable and Controllable]]></title><link>https://www.nownextlater.ai/Insights/post/the-future-of-ai-language-models-making-them-more-interpretable-and-controllable</link><description><![CDATA[Backpack models have an internal structure that is more interpretable and controllable compared to existing models like BERT and GPT-3.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_40edV6uaR2CF3GzLJhrmLg" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_bHHV-t--Q8KpHCq-rXWWWQ" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_Pf8sgIpQR7eY5B9Et2TEjg" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_4N6BObUT3TqlygIca0xvtQ" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_4N6BObUT3TqlygIca0xvtQ"] .zpimage-container figure img { width: 1090px ; height: 294.30px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_4N6BObUT3TqlygIca0xvtQ"] .zpimage-container figure img { width:723px ; height:195.21px ; } } @media (max-width: 767px) { [data-element-id="elm_4N6BObUT3TqlygIca0xvtQ"] .zpimage-container figure img { width:415px ; height:112.05px ; } } [data-element-id="elm_4N6BObUT3TqlygIca0xvtQ"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-fit zpimage-tablet-fallback-fit zpimage-mobile-fallback-fit hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-08%20at%207.20.24%20pm.png" width="415" height="112.05" loading="lazy" size="fit" alt="TheeffectontheconditionalprobabilitydistributionofaBackpackLMontheprefixwhenthenurse walked into the room" data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_4k10nHrXTz-4dumsQW3w8A" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_4k10nHrXTz-4dumsQW3w8A"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Artificial intelligence has made great strides in recent years, especially in natural language processing. Systems like ChatGPT and Claude can now hold impressively human-like conversations. However, a major limitation of these AI language models is that they operate like a black box - their internal workings are complex and opaque.</p><p><br></p><p>Researchers at Stanford have proposed a new AI architecture called Backpack that aims to fix this problem. Backpack models have an internal structure that is more interpretable and controllable compared to existing models like BERT and GPT-3.</p><p><br></p><p>Here's an analogy to understand how Backpack works:</p><p><br></p><p>Think of words as Lego blocks. Each block can be connected to other blocks in many ways to build something. Existing AI models are like throwing all the Lego pieces together in a pile - there are endless ways to combine them, but you can't understand or control the resulting structure.</p><p><br></p><p>A Backpack model is more like having clearly labeled Lego pieces in different bags. For each word, there are &quot;sense vectors&quot; that represent its different meanings and uses. When the model sees a word in a sentence, it decides which sense vectors to pull out of the bag to understand and predict that usage.</p><p><br></p><p>This structure offers two key benefits:</p><ol style="margin-left:40px;"><li>Interpretability: We can inspect the different sense vectors for a word and understand what aspects of meaning they represent. This is like looking inside the bags to see the different kinds of Lego pieces.</li><li>Control: We can directly edit the sense vectors to change the model's behavior. For example, reducing a gender-biased sense vector for the word &quot;nurse&quot; can reduce sexist outputs. This is like removing certain Lego pieces from a bag to change what can be built with it.</li></ol><p><br></p><p>In initial tests, Backpack models matched the performance of existing models like GPT-2 while offering far more transparency. Researchers were able to do things like swap associations (so &quot;MacBook&quot; predicts &quot;HP&quot; instead of &quot;Apple&quot;) and reduce gender bias in occupations.</p><p><br></p><p>The inventors stress that Backpack is still early stage research. The approach needs to be scaled up and tested across different languages and applications. But it represents an exciting step towards AI systems that are not black boxes. Instead of blindly trusting model outputs, users can interpret why it behaves in certain ways and directly edit its knowledge.</p><p><br></p><p>As AI becomes more powerful and ubiquitous in products and services, retaining human agency is crucial. Approaches like Backpack could make future AI not only smarter but easier to understand and actively improve. Business leaders should track developments in interpretable AI closely, as it is an important competitive differentiator down the line.</p><p><br></p><p>Source:</p><p><a href="https://arxiv.org/abs/2305.16765" title="arxiv" rel="">arxiv</a></p><p></p><p><br></p></div><p></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 07:59:55 +1000</pubDate></item><item><title><![CDATA[Tracking Political Bias from Data to Models to Decisions]]></title><link>https://www.nownextlater.ai/Insights/post/tracking-political-bias-from-data-to-models-to-decisions</link><description><![CDATA[Measuring the political leaning of various pretrained LMs.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_CATDepYUTZ6u_AaEaiMDmw" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_TQE0pWSXSsmNWQMJz0xU5w" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_vkv1E3dlTz2JTwVrqkIAvg" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_sXssLby6kfc7e84vxKKd2g" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_sXssLby6kfc7e84vxKKd2g"] .zpimage-container figure img { width: 1090px ; height: 606.49px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_sXssLby6kfc7e84vxKKd2g"] .zpimage-container figure img { width:723px ; height:402.28px ; } } @media (max-width: 767px) { [data-element-id="elm_sXssLby6kfc7e84vxKKd2g"] .zpimage-container figure img { width:415px ; height:230.91px ; } } [data-element-id="elm_sXssLby6kfc7e84vxKKd2g"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-fit zpimage-tablet-fallback-fit zpimage-mobile-fallback-fit hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-08%20at%206.30.01%20pm.png" width="415" height="230.91" loading="lazy" size="fit" alt="Measuring the political leaning of various pretrained LMs." data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_dXL84-vaTjup8bh7r_IhLw" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_dXL84-vaTjup8bh7r_IhLw"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><p>Recent advances in AI, especially large language models like GPT-4 and Claude 2, have unlocked new capabilities in generating text and speech. However, these models are still &quot;sponges&quot; that absorb patterns, including potential societal biases, from their training data. A new study from the University of Washington digs into an important question - can political bias in the data propagate to the models and affect downstream decisions? Their findings highlight risks that businesses should be aware of when deploying these technologies.</p><p><br></p><div style="color:inherit;"><p>The researchers focused on political leanings across two dimensions - social values (liberal vs conservative) and economic values (left vs right). First, they evaluated the inherent biases of 14 major language models by analyzing their responses to statements from a standard political compass test. The models occupied a range of positions across the political spectrum, with BERT variants being more conservative and GPT models more libertarian.</p><p><br></p><p>Next, they examined if further pretraining these models on partisan news and social media corpora leads them to shift their political stances. The results show that left-leaning data induces liberal shifts, while right-leaning data causes conservative shifts. However, the overall shifts were small, suggesting inherent biases persist.</p><p><br></p><p>Finally, they tested if these biased models perform differently on social-impact tasks like hate speech and misinformation detection. While overall performance was similar, models exhibited double standards - left-leaning ones were more sensitive to offensive speech targeting minorities but overlooked attacks on dominant groups. Right-leaning models showed the opposite.</p><p><br></p><p>For business leaders, these findings highlight risks of unintended bias and unfairness creeping into AI systems built on large language models. While some bias is inevitable given the training data, being aware of its extent and impact can help inform ethical AI practices. Companies should proactively probe for biases, use diverse evaluation datasets, and leverage different perspectives through ensemble approaches.</p><p><br></p><p>Tracking bias from data to models to decisions is vital for ensuring AI transparency and accountability. <br></p><p><br></p><p>Source:</p><p><a href="https://arxiv.org/pdf/2305.08283.pdf" title="Arxiv" rel="">Arxiv</a><br></p><p></p></div><p></p></div>
</div><div data-element-id="elm__u7614OwkSLZagtLzOCBGw" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm__u7614OwkSLZagtLzOCBGw"] .zpimage-container figure img { width: 500px ; height: 500.00px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm__u7614OwkSLZagtLzOCBGw"] .zpimage-container figure img { width:500px ; height:500.00px ; } } @media (max-width: 767px) { [data-element-id="elm__u7614OwkSLZagtLzOCBGw"] .zpimage-container figure img { width:500px ; height:500.00px ; } } [data-element-id="elm__u7614OwkSLZagtLzOCBGw"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium "><figure role="none" class="zpimage-data-ref"><a class="zpimage-anchor" href="/responsible-ai-in-the-age-of-generative-models-ai-governance-ethics-and-risk-management" target="" rel=""><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Navy%20and%20Blue%20Modern%20We%20Provide%20Business%20Solutions%20Facebook%20Ad%20-1200%20x%201200%20px-.png" width="500" height="500.00" loading="lazy" size="medium"/></picture></a></figure></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 07:57:52 +1000</pubDate></item><item><title><![CDATA[Understanding How AI Generates Images from Text]]></title><link>https://www.nownextlater.ai/Insights/post/understanding-how-ai-generates-images-from-text</link><description><![CDATA[In a paper titled "What the DAAM: Interpreting Stable Diffusion Using Cross Attention", researchers propose a method called DAAM (Diffusion Attentive Attribution Maps) to analyze how words in a prompt influence different parts of the generated image.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_S_ThR09CRDuVWQrYiHfN9A" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_UjY4eUtLR56tAebAvz_5YQ" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_f6aZA38_Q2a4WrwIYDMrSg" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"> [data-element-id="elm_f6aZA38_Q2a4WrwIYDMrSg"].zpelem-col{ border-radius:1px; } </style><div data-element-id="elm_3h_Z1k4B-_HRIofT4vhyyg" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_3h_Z1k4B-_HRIofT4vhyyg"] .zpimage-container figure img { width: 942px ; height: 345.60px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_3h_Z1k4B-_HRIofT4vhyyg"] .zpimage-container figure img { width:723px ; height:265.25px ; } } @media (max-width: 767px) { [data-element-id="elm_3h_Z1k4B-_HRIofT4vhyyg"] .zpimage-container figure img { width:415px ; height:152.25px ; } } [data-element-id="elm_3h_Z1k4B-_HRIofT4vhyyg"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-fit zpimage-tablet-fallback-fit zpimage-mobile-fallback-fit hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-08%20at%204.56.27%20pm.png" width="415" height="152.25" loading="lazy" size="fit" alt="Example generations and DAAM heat maps from COCO for each interpretable part-of-speech." data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_fySW7qlHSyO1GNVUxb6iNw" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_fySW7qlHSyO1GNVUxb6iNw"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Recent advances in AI image generation have led to impressive results, with systems like Stable Diffusion able to create highly realistic images from simple text prompts. But while these models can generate remarkably detailed pictures, they remain something of a black box. We don't have much insight into how they actually convert the text into pixel representations.</p><p><br></p><p>New research from computer scientists is helping peel back the curtain on these AI systems. In a paper titled &quot;What the DAAM: Interpreting Stable Diffusion Using Cross Attention&quot;, researchers propose a method called DAAM (Diffusion Attentive Attribution Maps) to analyze how words in a prompt influence different parts of the generated image.</p><p><br></p><p>DAAM creates heat maps showing which pixels are most related to each word in the text prompt. For example, for a prompt like &quot;a blue bird flying&quot;, the word &quot;blue&quot; would highlight the blue parts of the bird, &quot;bird&quot; would highlight the full bird, and &quot;flying&quot; would highlight the motion blurred wings and body.</p><p><br></p><p>By aggregating attention scores between text and image patches across the AI model's layers, DAAM produces interpretable maps linking words to visual features. The researchers validated DAAM by testing how well it can perform noun segmentation, a common computer vision benchmark <span style="color:inherit;">where the goal is to identify the regions in an image that correspond to noun objects</span>. DAAM achieved competitive scores, despite having no explicit training.</p><p><br></p><p>Experiments with DAAM revealed new insights about these generative AI systems:</p><ul><li>The relationships between words in a prompt translate to visual relationships in the image. For example, verbs like &quot;flying&quot; encapsulate their subjects like &quot;bird&quot;.</li><li>Using similar words like &quot;giraffe&quot; and &quot;zebra&quot; leads to worse image generation, likely because their features get entangled. The DAAM maps heavily overlap for such co-hyponyms. <span style="color:inherit;">Essentially, co-hyponyms are words that share the same superclass or category but are not synonyms. </span><span style="color:inherit;">They have a sort of &quot;sibling&quot; status as distinct members of their parent group. </span><span style="color:inherit;">Giraffe and zebra are co-hyponyms. They both belong to the hypernym &quot;wild animals&quot;.</span></li><li>Descriptive adjectives like &quot;blue&quot; attend too broadly across the whole image, suggesting objects are entangled with their surroundings. Changing the adjective modifies the entire scene.</li></ul><p><br></p><p>For business leaders, research like DAAM is important because it improves the explainability of AI systems. As generative models become more ubiquitous, understanding how they operate will help identify limitations and better assess risks. Models that entangle features more could potentially suffer from bias or produce unrealistic outputs.</p><p><br></p><p>DAAM also demonstrates how attention mechanisms in AI models can be repurposed for interpretation, without retraining the models from scratch. This allows transparent analysis without compromising performance.</p><p><br></p><p>Overall, DAAM represents an impactful step toward explainable AI in generative models. Demystifying these systems will be key as businesses increasingly look to utilize powerful generative AI capabilities in their products and processes. Interpretability helps ensure these technologies are trustworthy and dependable.</p><p><br></p><p>Sources:</p><p><a href="https://arxiv.org/pdf/2210.04885.pdf" title="arxiv" rel="">arxiv</a><br></p><p></p></div></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Tue, 08 Aug 2023 16:59:39 +1000</pubDate></item></channel></rss>