<?xml version="1.0" encoding="UTF-8" ?><!-- generator=Zoho Sites --><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><atom:link href="https://www.nownextlater.ai/Insights/tag/interpretability/feed" rel="self" type="application/rss+xml"/><title>Now Next Later AI - Blog #Interpretability</title><description>Now Next Later AI - Blog #Interpretability</description><link>https://www.nownextlater.ai/Insights/tag/interpretability</link><lastBuildDate>Wed, 26 Nov 2025 21:41:00 +1100</lastBuildDate><generator>http://zoho.com/sites/</generator><item><title><![CDATA[Why AI is Harder Than We Think: Key Takeaways for Business Leaders]]></title><link>https://www.nownextlater.ai/Insights/post/Why-AI-is-Harder-Than-We-Think-Key-Takeaways-for-Business-Leaders</link><description><![CDATA[A new study has shown that transformers can be expressed in a simple logic formalism. This finding challenges the perception that transformers are inscrutable black boxes and suggests avenues for interpreting how they work.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_sfUGQRrPRMOb-zqryhN6kA" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_aI0rE4RHRAGxFiFOSq2e8g" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_KQc6QKC-TUS9k5ADouakpw" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_lLX8ye2SQte64RLyZASIwA" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_lLX8ye2SQte64RLyZASIwA"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Artificial intelligence (AI) has seen remarkable advances in recent years, from self-driving cars to defeating humans at games like chess and Go. However, leading AI expert Melanie Mitchell argues that achieving truly human-like &quot;general&quot; AI is much harder than many experts predict. Here are the key takeaways from her paper for business leaders:</p><ul><li>Narrow AI versus general AI: Success in specialized applications like chess or image recognition does not necessarily translate to broader capabilities. We cannot assume today's AI systems are on a steady path to human-level intelligence.</li><li>&quot;Easy&quot; things are hard: Basic human skills like perceiving the world and carrying on a conversation have proven very difficult to replicate in machines. Conversely, AI can excel at things that are quite hard for humans.</li><li>Wishful vocabulary: Terms like &quot;learn,&quot; &quot;understand,&quot; and &quot;think&quot; are often applied to today's AI, but these systems do not have the same underlying capabilities as humans. Using human vocabulary can misleadingly imply advanced intelligence.</li><li>Intelligence is embodied: Human intelligence relies heavily on our experiences and interactions using our entire bodies, not just abstract reasoning. Attempts to achieve human-level intelligence must consider embracing human-like bodies and environments.</li><li>Common sense is key: To operate successfully in the real world, AI needs the vast background knowledge humans accumulate about how the world works. We still do not understand how to enable machines to acquire this &quot;common sense.&quot;</li></ul><p><br></p><p>Key implications:</p><ul><li>Avoid overconfidence about timelines for achieving human-level AI based on hype and narrow successes. True general intelligence likely remains far off.</li><li>Focus investment on applications of existing AI capabilities, not attempts to replicate human thinking. Manage expectations of near-term outcomes.</li><li>Monitor advances in embodied AI and research on common sense reasoning as indicators of progress toward general AI.</li><li>Ensure AI systems have transparent workings, clear objectives, and human oversight. The orthogonality thesis that intelligence can be coupled to any goals does not hold for human-like general intelligence.</li></ul><p><br></p><p>The path to human-level AI is long with much still unknown. By avoiding unfounded assumptions and acknowledging the challenges ahead, business leaders can make wise strategic decisions about how to apply AI technology today and anticipate what may come tomorrow.</p><p><br></p><p>Sources:</p><div style="color:inherit;"><div><div><div><div><div><p><span style="font-size:14px;"><a href="https://arxiv.org/abs/2104.12871" title="Why AI is Harder Than We Think " rel="">Why AI is Harder Than We Think </a></span></p><p></p><span style="font-size:14px;"></span><p><span style="font-size:14px;">Melanie Mitchell </span></p></div></div>
</div></div></div></div><p></p></div><p></p></div></div></div></div></div></div></div>
 ]]></content:encoded><pubDate>Sun, 13 Aug 2023 19:50:20 +1000</pubDate></item><item><title><![CDATA[Transformers Expressible in Simple Logic]]></title><link>https://www.nownextlater.ai/Insights/post/Transformers-Expressible-in-Simple-Logic</link><description><![CDATA[A new study has shown that transformers can be expressed in a simple logic formalism. This finding challenges the perception that transformers are inscrutable black boxes and suggests avenues for interpreting how they work.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_vTZ0UtxrTVKUpnRBcGFGwQ" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_OK0gUtTtTG-p2y_MzdYVpA" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_JUny_qLdSUqolf8oNtk3FQ" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_-muaciReSB2TX7jNYu8tFQ" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_-muaciReSB2TX7jNYu8tFQ"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>A new study from New York University and the Allen Institute for AI has shown that large language models called transformers can be expressed in a simple logic formalism. This finding challenges the perception that transformers are inscrutable black boxes and suggests avenues for interpreting how they work.</p><p><br></p><p>Transformers are a type of neural network behind major AI achievements like chatbots and language translation. They are trained on massive datasets to generate human-like text. Despite their impressive capabilities, how transformers arrive at their outputs has remained poorly understood.</p><p><br></p><p>The researchers proved transformers can be translated into symbolic logic sentences that replicate their function. Specifically, they showed transformers fit within a logic called first-order logic with majority quantifiers. This logic allows logical sentences with familiar constructs like &quot;AND&quot;, &quot;OR&quot;, and &quot;IF-THEN&quot;, as well as majority quantifiers that check if a condition holds for over half of the elements.</p><p><br></p><p>While real-world transformers are complex neural networks, this study theoretically shows their reasoning can be captured by simple logical expressions. For instance, the logic could recognize patterns like &quot;three As followed by three Bs&quot;, which transformers are known to identify.</p><p><br></p><p>The findings disprove the notion that transformers are inscrutable black boxes. Instead, they suggest transformers implement a form of reasoning not radically different from familiar logical formalisms. The possibility of expressing transformers in interpretable logic could enable explaining how they arrive at outputs, like detecting biases.</p><p><br></p><p>For business leaders deploying AI, this research opens possibilities for making transformers more transparent and accountable. It provides a path toward debugging models to avoid failures or bias. The ability to translate transformers into logical sentences could allow systematically checking if undesirable reasoning patterns occur.</p><p><br></p><p>Overall, this theoretical advance challenges prevailing views of transformers as hopelessly opaque. It demonstrates their thinking can be characterized in understandable logic, unlocking new ways for technologists to interpret these increasingly critical AI models. The research brings transformers closer to human-level reasoning by showing their outputs are not ineffable, but rather can be explained through logic.</p><p><br></p><p>Sources:</p><div style="color:inherit;"><div><div><div><div><p><a href="https://arxiv.org/abs/2210.02671" title="A Logic for Expressing Log-Precision Transformers " rel="">A Logic for Expressing Log-Precision Transformers </a><br></p><p></p><div style="color:inherit;"><div><div><div><div><p>William Merrill and Ashish Sabharwal </p><div style="color:inherit;"><div><div><div><span style="font-size:10pt;font-weight:500;"><br></span></div>
</div></div></div></div></div></div></div></div><p></p></div></div></div></div></div><p></p></div><p></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Sun, 13 Aug 2023 19:50:20 +1000</pubDate></item><item><title><![CDATA[DisentQA: Catching Knowledge Gaps and Avoiding Misleading Users]]></title><link>https://www.nownextlater.ai/Insights/post/enabling-ai-to-untangle-different-knowledge-sources</link><description><![CDATA[Building QA Systems that catch knowledge gaps and avoid misleading users.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_ewF7pMN9Q_eczUOQpCYtUA" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_XdQfIANyTi-5Z3w2LSGv-A" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_tGWqJgjLSlyldj1XkXMcGw" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_KipIDvLOVMb6oIC8bF9TkA" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_KipIDvLOVMb6oIC8bF9TkA"] .zpimage-container figure img { width: 500px ; height: 486.01px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_KipIDvLOVMb6oIC8bF9TkA"] .zpimage-container figure img { width:500px ; height:486.01px ; } } @media (max-width: 767px) { [data-element-id="elm_KipIDvLOVMb6oIC8bF9TkA"] .zpimage-container figure img { width:500px ; height:486.01px ; } } [data-element-id="elm_KipIDvLOVMb6oIC8bF9TkA"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-12%20at%209.09.37%20am.png" width="500" height="486.01" loading="lazy" size="medium" alt="Example outputs from our disentangled QA model on the Natural Questions dataset. " data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_TNbKqQ17TP256B60EqRP7w" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_TNbKqQ17TP256B60EqRP7w"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><div style="color:inherit;"><p>Imagine you ask your phone &quot;Who wrote the song Hello by Adele?&quot; and it gives you an incorrect answer, insisting the song is by Taylor Swift. This shows artificial intelligence sometimes confuses its own training knowledge with external facts.</p><p><br></p><p>Researchers want to fix this issue to make AI assistants more helpful and honest. Their solution: <span style="color:inherit;">Build QA Systems that catch knowledge gaps and avoid misleading users by </span>teaching the system to provide two responses:</p><ol><li>The factual answer based on given information (e.g. Adele)</li><li>What it privately recalls from its memory (e.g. Taylor Swift)</li></ol><p><br></p><p>This highlights any mismatches between its training knowledge and external data. It's like when we say &quot;Hmm, I thought X, but the website says Y.&quot;</p><p><br></p><p>The team trained the AI model by creating quizzes with tricky examples:</p><ul><li>Swapping names in passages to elicit different responses from the context vs. the model's recollection</li><li>Removing passages altogether so the system must say &quot;I don't know&quot;</li></ul><p><br></p><p>After this special training, the model reliably distinguished its own knowledge from given facts. This improved its accuracy and truthfulness.</p><p><br></p><p>Say you ask about a movie release date. The system can now respond:</p><p><span style="font-style:italic;">&quot;The article says July 2022. But I thought it was December 2022.&quot;</span></p><p><br></p><p>This catches any knowledge gaps and avoids misleading users.</p><p><br></p><p>While not perfect, it's major progress toward AI that collaborates in a transparent, helpful manner. The benefits for businesses are clear:</p><ul><li>Avoid frustrated users with incorrect responses</li><li>Build trust by exposing limitations upfront</li><li>Reduce risk from applying flawed knowledge</li><li>Clarify when external data should override internal beliefs</li></ul><p><br></p><p>By recognizing and sharing when its knowledge is incomplete, the AI becomes a more reliable and honest partner. This research brings us closer to truly cooperative human-AI interaction.</p><p><br></p><p>Sources:</p><p><span style="color:inherit;"><a href="https://arxiv.org/pdf/2211.05655.pdf" title="DisentQA: Disentangling Parametric and Contextual Knowledge with Counterfactual Question Answering" rel="">DisentQA: Disentangling Parametric and Contextual Knowledge with Counterfactual Question Answering</a></span></p><p></p></div>
</div><p></p></div></div></div></div></div></div></div> ]]></content:encoded><pubDate>Sat, 12 Aug 2023 09:22:46 +1000</pubDate></item><item><title><![CDATA[Peeking Inside the Black Box: Uncovering What AI Models Know About Books]]></title><link>https://www.nownextlater.ai/Insights/post/peeking-inside-the-black-box-uncovering-what-ai-models-know-about-books</link><description><![CDATA[New research from the University of California, Berkeley sheds light on one slice of these models' knowledge: which books they have "read" and memorized. The study uncovers systematic biases in what texts AI systems know most about.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_gs7G1b-XTpGiZydeBC0a0Q" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_DOD6ivG7QmSpfUB8fVxIYQ" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_LiwzJORqQmSQI3gbPKBwuA" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_K_koNQXna4P8T3ZlwBatnw" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_K_koNQXna4P8T3ZlwBatnw"] .zpimage-container figure img { width: 1090px ; height: 671.48px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_K_koNQXna4P8T3ZlwBatnw"] .zpimage-container figure img { width:723px ; height:445.39px ; } } @media (max-width: 767px) { [data-element-id="elm_K_koNQXna4P8T3ZlwBatnw"] .zpimage-container figure img { width:415px ; height:255.65px ; } } [data-element-id="elm_K_koNQXna4P8T3ZlwBatnw"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-fit zpimage-tablet-fallback-fit zpimage-mobile-fallback-fit hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-09%20at%206.48.43%20pm.png" width="415" height="255.65" loading="lazy" size="fit" alt="Top 20 books by GPT-4 name cloze accuracy" data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_a3xA-CbcQJepcHHoPuycYQ" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_a3xA-CbcQJepcHHoPuycYQ"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Artificial intelligence systems like ChatGPT and GPT-4 have demonstrated impressive language skills, holding fluent conversations and answering questions on virtually any topic. But their inner workings remain largely opaque to users. These systems are &quot;black boxes&quot; - we know little about what knowledge they actually contain.</p><p><br></p><p>New research from the University of California, Berkeley sheds light on one slice of these models' knowledge: which books they have &quot;read&quot; and memorized. The study uncovers systematic biases in what texts AI systems know most about, with implications for how we should evaluate them.</p><p>The researchers focused specifically on works of fiction. They selected a sample of 571 English novels published between 1749 and 2020, containing literary classics along with contemporary bestsellers and award winners. The sample spanned mystery, romance, and science fiction genres as well as global Anglophone and African American literature.</p><p><br></p><p>For each book, the team extracted short passages of 40-60 words containing a single character name - but with the name removed. For instance, a passage from Pride and Prejudice might read &quot;______ entered the room and greeted her hosts warmly.&quot; Humans cannot guess the missing name from such brief context. But does the AI system know the name from having read the full book?</p><p><br></p><p>The researchers tested two systems, ChatGPT and GPT-4, by giving each passage and asking what single-word name belongs in the blank. The accuracy of each AI model on this challenging &quot;cloze&quot; task revealed what books it likely memorized.</p><p><br></p><p>The results illuminated clear biases. Both systems strongly favor science fiction and fantasy works like Lord of the Rings and Harry Potter over other genres. They excel at classic literature like Alice in Wonderland and Pride and Prejudice but fare poorly on modern award-winning diverse books. In short, they are more knowledgeable about popular texts.</p><p><br></p><p>What explains this imbalance? The researchers found it closely mirrors what's most duplicated across the internet. There is a strong correlation between AI accuracy on a book and the number of verbatim passages found through Google, Bing, and other sources. The models appear to &quot;know&quot; books in proportion to their web popularity.</p><p><br></p><p>This reliance on the internet has consequences. The study showed AI systems perform better at predicting a book's publication date and summarizing its passages when they have memorized the book. In other words, their reasoning is tied to memorization - causing disparities between popular versus niche texts.</p><p><br></p><p>These insights matter because AI systems like ChatGPT are increasingly used for applications like analyzing literature and human culture. If their knowledge comes largely from duplicated web text, focused on popular sci-fi and fantasy, how well can we trust their judgments about less mainstream books? Their skewed knowledge could propagate biases into downstream decisions.</p><p><br></p><p>The findings illustrate the challenges of opaque &quot;black box&quot; AI systems whose training data is secret. OpenAI, which created ChatGPT and GPT-4, has not revealed what texts were used to train them. This leaves us unable to fully assess their knowledge gaps.</p><p><br></p><p>The researchers argue we should instead push for more transparent, open-source AI systems whose training data is public knowledge. This allows us to better understand their strengths and weaknesses - illuminated through research like this study.</p><p><br></p><p>As AI models grow more capable and ubiquitous, it becomes only more important to peek inside their black boxes. Understanding what knowledge they contain helps ensure we build and apply them responsibly. Analyses of what systems like ChatGPT &quot;know&quot; about books mark an important step toward making AI more intelligible as it continues permeating our lives.</p><p><br></p><p>Sources:</p><p><a href="https://arxiv.org/pdf/2305.00118.pdf" title="arxiv" rel="">arxiv</a><br></p><p></p></div><p></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:03:22 +1000</pubDate></item><item><title><![CDATA[The Future of AI Language Models: Making Them More Interpretable and Controllable]]></title><link>https://www.nownextlater.ai/Insights/post/the-future-of-ai-language-models-making-them-more-interpretable-and-controllable</link><description><![CDATA[Backpack models have an internal structure that is more interpretable and controllable compared to existing models like BERT and GPT-3.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_40edV6uaR2CF3GzLJhrmLg" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_bHHV-t--Q8KpHCq-rXWWWQ" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_Pf8sgIpQR7eY5B9Et2TEjg" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_4N6BObUT3TqlygIca0xvtQ" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_4N6BObUT3TqlygIca0xvtQ"] .zpimage-container figure img { width: 1090px ; height: 294.30px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_4N6BObUT3TqlygIca0xvtQ"] .zpimage-container figure img { width:723px ; height:195.21px ; } } @media (max-width: 767px) { [data-element-id="elm_4N6BObUT3TqlygIca0xvtQ"] .zpimage-container figure img { width:415px ; height:112.05px ; } } [data-element-id="elm_4N6BObUT3TqlygIca0xvtQ"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-fit zpimage-tablet-fallback-fit zpimage-mobile-fallback-fit hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-08%20at%207.20.24%20pm.png" width="415" height="112.05" loading="lazy" size="fit" alt="TheeffectontheconditionalprobabilitydistributionofaBackpackLMontheprefixwhenthenurse walked into the room" data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_4k10nHrXTz-4dumsQW3w8A" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_4k10nHrXTz-4dumsQW3w8A"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Artificial intelligence has made great strides in recent years, especially in natural language processing. Systems like ChatGPT and Claude can now hold impressively human-like conversations. However, a major limitation of these AI language models is that they operate like a black box - their internal workings are complex and opaque.</p><p><br></p><p>Researchers at Stanford have proposed a new AI architecture called Backpack that aims to fix this problem. Backpack models have an internal structure that is more interpretable and controllable compared to existing models like BERT and GPT-3.</p><p><br></p><p>Here's an analogy to understand how Backpack works:</p><p><br></p><p>Think of words as Lego blocks. Each block can be connected to other blocks in many ways to build something. Existing AI models are like throwing all the Lego pieces together in a pile - there are endless ways to combine them, but you can't understand or control the resulting structure.</p><p><br></p><p>A Backpack model is more like having clearly labeled Lego pieces in different bags. For each word, there are &quot;sense vectors&quot; that represent its different meanings and uses. When the model sees a word in a sentence, it decides which sense vectors to pull out of the bag to understand and predict that usage.</p><p><br></p><p>This structure offers two key benefits:</p><ol style="margin-left:40px;"><li>Interpretability: We can inspect the different sense vectors for a word and understand what aspects of meaning they represent. This is like looking inside the bags to see the different kinds of Lego pieces.</li><li>Control: We can directly edit the sense vectors to change the model's behavior. For example, reducing a gender-biased sense vector for the word &quot;nurse&quot; can reduce sexist outputs. This is like removing certain Lego pieces from a bag to change what can be built with it.</li></ol><p><br></p><p>In initial tests, Backpack models matched the performance of existing models like GPT-2 while offering far more transparency. Researchers were able to do things like swap associations (so &quot;MacBook&quot; predicts &quot;HP&quot; instead of &quot;Apple&quot;) and reduce gender bias in occupations.</p><p><br></p><p>The inventors stress that Backpack is still early stage research. The approach needs to be scaled up and tested across different languages and applications. But it represents an exciting step towards AI systems that are not black boxes. Instead of blindly trusting model outputs, users can interpret why it behaves in certain ways and directly edit its knowledge.</p><p><br></p><p>As AI becomes more powerful and ubiquitous in products and services, retaining human agency is crucial. Approaches like Backpack could make future AI not only smarter but easier to understand and actively improve. Business leaders should track developments in interpretable AI closely, as it is an important competitive differentiator down the line.</p><p><br></p><p>Source:</p><p><a href="https://arxiv.org/abs/2305.16765" title="arxiv" rel="">arxiv</a></p><p></p><p><br></p></div><p></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 07:59:55 +1000</pubDate></item></channel></rss>