<?xml version="1.0" encoding="UTF-8" ?><!-- generator=Zoho Sites --><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><atom:link href="https://www.nownextlater.ai/Insights/tag/aistorybrain/feed" rel="self" type="application/rss+xml"/><title>Now Next Later AI - Blog ##AIStoryBrain</title><description>Now Next Later AI - Blog ##AIStoryBrain</description><link>https://www.nownextlater.ai/Insights/tag/aistorybrain</link><lastBuildDate>Wed, 26 Nov 2025 21:34:40 +1100</lastBuildDate><generator>http://zoho.com/sites/</generator><item><title><![CDATA[The Promise of Frozen Language Models]]></title><link>https://www.nownextlater.ai/Insights/post/the-promise-of-frozen-language-models</link><description><![CDATA[In their research paper, AI21 Labs demonstrates that frozen LLMs have untapped potential that can match or exceed fine-tuning approaches, without the downsides.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_pZ5uovp4RkaCvppLft55-A" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_y9ufJ0AHQj-hwYuzvPKprA" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_k5ot811CSPmqowZ_GCcmVw" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_oqcnPNxOp2fdkHf66r8uDw" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_oqcnPNxOp2fdkHf66r8uDw"] .zpimage-container figure img { width: 800px ; height: 600.00px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_oqcnPNxOp2fdkHf66r8uDw"] .zpimage-container figure img { width:500px ; height:375.00px ; } } @media (max-width: 767px) { [data-element-id="elm_oqcnPNxOp2fdkHf66r8uDw"] .zpimage-container figure img { width:500px ; height:375.00px ; } } [data-element-id="elm_oqcnPNxOp2fdkHf66r8uDw"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-large zpimage-tablet-fallback-large zpimage-mobile-fallback-large hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/aaron-burden-1FTQOGziGY4-unsplash-1.jpg" width="500" height="375.00" loading="lazy" size="large" alt="Photo by Aaron Burden on Unsplash" data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_Oj77Wq2UQg2-9wI6vhegkQ" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_Oj77Wq2UQg2-9wI6vhegkQ"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><p>In recent years, artificial intelligence has taken great leaps forward thanks to large language models (LLMs) - AI systems trained on massive amounts of text data that can understand language and generate human-like text. Companies like Google, Microsoft, and startups like OpenAI and Anthropic have invested heavily in developing ever-larger LLMs with billions or even trillions of parameters.</p><p><br></p><div style="color:inherit;"><p>However, once these giant LLMs are trained, companies face a dilemma - whether to &quot;fine-tune&quot; the model by further training it on specific tasks, or keep the model &quot;frozen&quot; without any changes. Fine-tuning allows the LLM to specialize and achieve state-of-the-art performance on specialized tasks. But it comes at a high cost - computationally expensive retraining, reduced versatility, and forgetting of previous capabilities.</p><p><br></p><p>In their research paper, AI21 Labs demonstrates that frozen LLMs have untapped potential that can match or exceed fine-tuning approaches, without these downsides. They present three new techniques to effectively &quot;stand on the shoulders&quot; of frozen giants:</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;">1. Input-Dependent Prompt Tuning</span></p><p><br></p><div style="color:inherit;"><div style="color:inherit;"><div style="color:inherit;"><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;">Large language models are adept at understanding natural language, but they don't automatically know how to perform specific tasks like answering questions or summarizing text.However, their capabilities can be unlocked using prompt tuning.</span></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;"><br></span></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;">The key idea behind prompt tuning is that providing the right prompt text before the input steers the language model towards the desired task.It's like giving the model instructions on how to process the upcoming input.</span></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;"><br></span></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;">For example, if we want the language model to answer questions based on a passage of text, we can prepend the input with a prompt like:</span></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;">&quot;<span style="font-style:italic;">Answer the following question based only on the passage below:</span>&quot;</span></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;">[Text Passage]</span></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;">[Question]</span></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;"><br></span></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;">This tunes the model to approach the upcoming input as a question answering task.The prompt acts like an adapter, steering the versatile model to useful behaviors without any training or fine-tuning.</span></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;"><br></span></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;">So prompt tuning just means optimizing the wording of these instruction prompts for each task to get the best performance from the frozen language model.It's like learning how to most effectively communicate with and direct the model.</span></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;"><br></span></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;">The key innovation from AI21 Labs was making prompt tuning input-dependent. Rather than using one static prompt per task, they trained a small neural network to generate custom prompts tailored to each specific input.</span></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;"><br></span></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;">This input-dependent prompting allowed a single frozen language model to master over 100 diverse tasks, from question answering to summarization to sentiment analysis,&nbsp;matching extensive fine-tuning without degradation.</span></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;"><br></span></p><p style="font-weight:400;text-indent:0px;"><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:14px;">The prompts serve as lightweight yet powerful steering instructions that can specialize a frozen model on the fly based on the input.It's like having a dynamic adapter that configures the model differently for each unique situation.</span></p></div></div></div><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;">2. Huge Frozen Readers for Question Answering</span></p><p><br></p><div style="color:inherit;"><p>In open-domain question answering, the AI system must answer questions by finding relevant information from a massive collection of text passages, like Wikipedia.</p><p><br></p><p>Typically, these systems use a smaller &quot;reader&quot; model to read through the relevant passages and figure out the answer. That's because even the largest language models can only process a limited amount of text at once.</p><p><br></p><p>But smaller reader models have less knowledge and reasoning ability than giant language models with billions or trillions of parameters. So they don't fully unlock the potential of these frozen giants.</p><p><br></p><p>AI21 Labs tackled this by adding a &quot;re-ranking&quot; stage to condense the most important information from the passages into a condensed form that fits into the giant frozen language model.</p><p><br></p><p>This allowed their 17 billion parameter model to read enough of the relevant context to match specialized reader models that were extensively fine-tuned for question answering.</p><p><br></p><p>In essence, the smaller re-ranking model acts like a search engine, retrieving and condensing the most useful knowledge to fit the limitations of the frozen giant.</p><p><br></p><p>This gives the huge frozen model access to all the relevant information it needs to apply its powerful reasoning abilities. The giants' knowledge and capabilities can be tapped without fine-tuning that risks degrading other skills.</p><p><br></p><p>It demonstrates how frozen language models have untapped potential that can be unlocked with the right surrounding components, like the re-ranking stage here. Their true capabilities can be accessed without resorting to extensive fine-tuning.</p><p><br></p></div><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;">3. Recursive Application of a Single LLM</span></p><p><br></p><div style="color:inherit;"><p>Typically, large language models are used to process an input query just once before generating an output response. The model reads the input, does its internal reasoning, and returns a single output.</p><p><br></p><p>But AI21 Labs found that recursively applying the model on its own outputs can actually improve performance. Essentially, the model refines and enhances its initial output by processing it again.</p><p><br></p><p>It's like having the model double-check its own work and refine its initial response. Humans often re-read what we initially wrote to improve the wording and fix errors. Recursively applying language models does something similar, but in an automated way.</p><p><br></p><p>To implement this, AI21 built a small 2-layer neural network &quot;connector&quot; that feeds the language model's output back into its input.</p><p><br></p><p>So the model first processes the original query as normal. But then the connector passes the model's initial output back into it as the new input. This triggers it to refine and enhance that initial output.</p><p><br></p><p>In tests for question answering, just two recursive passes through a 7 billion parameter model allowed it to match the performance of a much larger 17 billion parameter model.</p><p><br></p><p>Essentially, it nearly doubled the capabilities of the smaller model by re-applying it recursively. This shows how recursive application unlocks additional performance without requiring even larger pretrained models.</p><p><br></p><p>The connector module creates a feedback loop, allowing the model to re-process its own output and correct errors or improve phrasing, much like a human would. This technique amplifies the capabilities of a given model without expensive retraining or fine-tuning.</p></div><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;">Business Implications</span></p><p><br></p><p>These techniques enable building capable AI systems on top of a single, frozen pretrained LLM instead of an array of specialized fine-tuned models. This offers tangible business benefits:</p><p><br></p><ul><li><span style="font-family:&quot;Josefin Sans&quot;, sans-serif;">Cost Savings</span> - Avoiding expensive training of multiple large models cuts costs. Just maintaining and serving one frozen LLM backbone provides economies of scale.</li><li><span style="font-family:&quot;Josefin Sans&quot;, sans-serif;">Simplicity</span> - Relying on prompting and other external components is far simpler than intricately fine-tuning models. Less specialized engineering effort is required.</li><li><span style="font-family:&quot;Josefin Sans&quot;, sans-serif;">Flexibility </span>- New capabilities can be added without interfering with existing ones. Fine-tuning risks degradation on previous tasks.</li><li><span style="font-family:&quot;Josefin Sans&quot;, sans-serif;">Efficiency</span> - Recursive passing allows improving performance on-demand by re-applying the LLM only when beneficial. Bigger pretrained models must be applied to all inputs.</li></ul><p><br></p><p>While fine-tuning revolutionized AI, endless model growth is impractical. Frozen language models present an alluring path forward - unlocking their full potential with the right neural &quot;plug-ins&quot; provides a scalable approach to building production AI systems.</p></div><div><br></div><div><br></div><div>Source:</div><div><div style="color:inherit;"><div><div><div><div><p><a href="https://arxiv.org/pdf/2204.10019.pdf" title="STANDING ON THE SHOULDERS OF GIANT FROZEN LANGUAGE MODELS " rel="">S</a><a href="https://arxiv.org/pdf/2204.10019.pdf" title="STANDING ON THE SHOULDERS OF GIANT FROZEN LANGUAGE MODELS " rel="">TANDING ON THE </a><a href="https://arxiv.org/pdf/2204.10019.pdf" title="STANDING ON THE SHOULDERS OF GIANT FROZEN LANGUAGE MODELS " rel="">S</a><a href="https://arxiv.org/pdf/2204.10019.pdf" title="STANDING ON THE SHOULDERS OF GIANT FROZEN LANGUAGE MODELS " rel="">HOULDERS OF </a><a href="https://arxiv.org/pdf/2204.10019.pdf" title="STANDING ON THE SHOULDERS OF GIANT FROZEN LANGUAGE MODELS " rel="">G</a><a href="https://arxiv.org/pdf/2204.10019.pdf" title="STANDING ON THE SHOULDERS OF GIANT FROZEN LANGUAGE MODELS " rel="">IANT </a><a href="https://arxiv.org/pdf/2204.10019.pdf" title="STANDING ON THE SHOULDERS OF GIANT FROZEN LANGUAGE MODELS " rel="">F</a><a href="https://arxiv.org/pdf/2204.10019.pdf" title="STANDING ON THE SHOULDERS OF GIANT FROZEN LANGUAGE MODELS " rel="">ROZEN </a><a href="https://arxiv.org/pdf/2204.10019.pdf" title="STANDING ON THE SHOULDERS OF GIANT FROZEN LANGUAGE MODELS " rel="">L</a><a href="https://arxiv.org/pdf/2204.10019.pdf" title="STANDING ON THE SHOULDERS OF GIANT FROZEN LANGUAGE MODELS " rel="">ANGUAGE </a><a href="https://arxiv.org/pdf/2204.10019.pdf" title="STANDING ON THE SHOULDERS OF GIANT FROZEN LANGUAGE MODELS " rel="">M</a><a href="https://arxiv.org/pdf/2204.10019.pdf" title="STANDING ON THE SHOULDERS OF GIANT FROZEN LANGUAGE MODELS " rel="">ODELS </a></p><p></p></div>
</div></div></div></div></div></div></div><div data-element-id="elm_9FN7BtAIh0F1sKkt6uGxEg" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_9FN7BtAIh0F1sKkt6uGxEg"] .zpimage-container figure img { width: 500px ; height: 500.00px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_9FN7BtAIh0F1sKkt6uGxEg"] .zpimage-container figure img { width:500px ; height:500.00px ; } } @media (max-width: 767px) { [data-element-id="elm_9FN7BtAIh0F1sKkt6uGxEg"] .zpimage-container figure img { width:500px ; height:500.00px ; } } [data-element-id="elm_9FN7BtAIh0F1sKkt6uGxEg"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium "><figure role="none" class="zpimage-data-ref"><a class="zpimage-anchor" href="/responsible-ai-in-the-age-of-generative-models-ai-governance-ethics-and-risk-management" target="" rel=""><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Navy%20and%20Blue%20Modern%20We%20Provide%20Business%20Solutions%20Facebook%20Ad%20-1200%20x%201200%20px-.png" width="500" height="500.00" loading="lazy" size="medium"/></picture></a></figure></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Fri, 11 Aug 2023 10:04:48 +1000</pubDate></item><item><title><![CDATA[Making AI More Useful and Reliable with Modular Systems: MRKL]]></title><link>https://www.nownextlater.ai/Insights/post/making-ai-more-useful-and-reliable-with-modular-systems-mrkl</link><description><![CDATA[LLMs have some serious limitations that constrain their usefulness for real-world applications. To overcome these limitations, AI researchers have proposed a new type of AI system architecture called Modular Reasoning, Knowledge and Language (MRKL).]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_r4wXq1G1Rjii1H0uUl9mjg" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_omUWrDQnQ1qO0KkRlkHxhQ" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_75IN8_ocSBO7BPhEfIXiZA" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_vVhNHqQD1NhKqb_Xbyx6AQ" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_vVhNHqQD1NhKqb_Xbyx6AQ"] .zpimage-container figure img { width: 500px ; height: 600.42px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_vVhNHqQD1NhKqb_Xbyx6AQ"] .zpimage-container figure img { width:500px ; height:600.42px ; } } @media (max-width: 767px) { [data-element-id="elm_vVhNHqQD1NhKqb_Xbyx6AQ"] .zpimage-container figure img { width:500px ; height:600.42px ; } } [data-element-id="elm_vVhNHqQD1NhKqb_Xbyx6AQ"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-10%20at%209.09.12%20pm.png" width="500" height="600.42" loading="lazy" size="medium" alt="MRKL Systems" data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_N4IffEM4Qnep0M0r7j--OA" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_N4IffEM4Qnep0M0r7j--OA"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><p style="text-align:left;">In recent years, large language models (LLMs) like GPT-4 have shown impressive capabilities in generating human-like text and engaging in natural language conversations. However, LLMs also have some serious limitations that constrain their usefulness for real-world applications. To overcome these limitations, AI researchers have proposed a new type of AI system architecture called Modular Reasoning, Knowledge and Language (MRKL). <br></p><p style="text-align:left;"><br></p><div style="color:inherit;"><p><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">The Limitations of Large Language Models</span></p><p></p><p><br></p><p>While LLMs can produce remarkably fluent text, they often generate factual errors, nonsensical statements, and inconsistent responses. This happens because LLMs do not actually have any real understanding of the world or ability to reason - they just recognize patterns in the massive datasets they are trained on. As a result, LLMs lack:</p><ul style="margin-left:40px;"><li>Access to current, real-time information that is constantly changing, like stock prices or weather data. The pre-trained models only know what was in their training data.</li><li>Access to proprietary data like customer records that exist in a company's databases. LLMs cannot connect to external databases.</li><li>Ability to perform symbolic reasoning and math. They struggle with simple arithmetic and logical deductions.</li><li>Ability to learn major new capabilities without catastrophic forgetting. Fine-tuning LLMs on new datasets leads to losing their original skills.</li></ul><p><br></p><p>These problems severely limit the reliability and usefulness of LLMs for practical business applications. Companies cannot deploy unreliable AI assistants that generate false information or nonsensical outputs.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Introducing Modular AI Systems</span></p><p></p><p><br></p><p>To address the weaknesses of monolithic LLMs, AI researchers have proposed breaking AI systems into modules with different capabilities that can work together:</p><ul style="margin-left:40px;"><li>Neural modules based on LLMs that handle natural language</li><li>Symbolic modules that perform logical reasoning and math</li><li>Access to external knowledge bases like databases and APIs</li></ul><p><br></p><p>This is the idea behind Modular Reasoning, Knowledge and Language (MRKL) architectures. MRKL systems have a router module that analyzes incoming questions and routes them to the most appropriate module - either the core LLM, a symbolic module, or an external database.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Benefits of the Modular Approach</span></p><p></p><p><br></p><p>Modular AI architectures provide important benefits compared to monolithic LLMs:</p><ul style="margin-left:40px;"><li>Reliability through redundancy. If the core LLM fails, questions can be routed to more reliable modules.</li><li>Easy extensibility by adding modules without retraining the whole system.</li><li>Explainability, since it's clear which module produced an answer.</li><li>Up-to-date real-time data from external APIs and databases.</li><li>Secure access to proprietary data sources.</li><li>Improved reasoning abilities by combining neural networks and symbolic modules.</li><li>Avoidance of catastrophic forgetting. New skills don't override old ones.</li></ul><p><br></p><p>This modularity and hybrid approach allows AI systems to leverage the strengths of different techniques while minimizing their weaknesses.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Challenges of Integrating Neural and Symbolic AI</span></p><p></p><p><br></p><p>A key challenge in building modular AI systems is integrating the neural LLM components with the symbolic reasoning modules. These two types of AI rely on completely different processing techniques - neural nets versus discrete logical operations.</p><p><br></p><p>Researchers have found that even extracting basic math problems from text for input to a calculator module requires specialized training to reach high accuracy. For example, the query &quot;I lost one ball&quot; needs to be recognized as a subtraction problem: X - 1 = ?.</p><p><br></p><p>By using large datasets of mathematically annotated text, modular AI systems can be trained to extract appropriate reasoning tasks with over 99% accuracy. But significant research is still required to handle more complex reasoning across modules. Integrating neural networks and symbolic systems remains an active area of investigation.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">The Future of Modular AI</span></p><p></p><p><br></p><p>Modular architectures represent an exciting evolution in AI that combines the strengths of different techniques. Companies like Anthropic and AI21 Labs are actively developing modular AI platforms to provide businesses with safer and more usable AI assistants. While challenges remain, the future appears bright for this hybrid approach to artificial intelligence.</p><p><br></p><p>Source:</p><p><span style="color:inherit;"><a href="https://arxiv.org/pdf/2205.00445.pdf" title="MRKL Systems" rel="">MRKL Systems</a></span></p><p></p></div></div>
</div><div data-element-id="elm_GZHJUMh2KTh65_KlSMcEGA" data-element-type="codeSnippet" class="zpelement zpelem-codesnippet "><div class="zpsnippet-container"><div class="video-container"><iframe src="https://www.youtube.com/embed/kSsRV1pSrhs?modestbranding=1&rel=0&cc_load_policy=1&iv_load_policy=3&controls=0&disablekb=1" width="560" height="315" title="Huge Language Models and Neuro-Symbolic AI - Prof. Yoav Shoham" loading=“lazy” frameborder="0" allow="fullscreen"></iframe></div>
</div></div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 21:21:30 +1000</pubDate></item><item><title><![CDATA[Enhancing AI with Symbolic Thinking]]></title><link>https://www.nownextlater.ai/Insights/post/enhancing-ai-with-symbolic-thinking</link><description><![CDATA[Researchers are exploring how to combine LLMs with neurosymbolic methods that incorporate logical reasoning and structure.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_92MOykrxQKud1bMa7iJXXQ" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_U54X55u7TYirdWrR1rLrvA" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_x5U3oKbOR3K_vjQPTrwGfg" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_5XLnBfPx_RXLzPyqbU7ksA" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_5XLnBfPx_RXLzPyqbU7ksA"] .zpimage-container figure img { width: 500px ; height: 710.69px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_5XLnBfPx_RXLzPyqbU7ksA"] .zpimage-container figure img { width:500px ; height:710.69px ; } } @media (max-width: 767px) { [data-element-id="elm_5XLnBfPx_RXLzPyqbU7ksA"] .zpimage-container figure img { width:500px ; height:710.69px ; } } [data-element-id="elm_5XLnBfPx_RXLzPyqbU7ksA"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-10%20at%206.30.11%20pm.png" width="500" height="710.69" loading="lazy" size="medium" alt="Given facts, rules, and a question all ex- pressed in natural language, ProofWriter answers the question and generates a proof of the answer." data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_s3YK1MdKTt-gsY1xAsJt_w" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_s3YK1MdKTt-gsY1xAsJt_w"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><div style="color:inherit;"><p>The rapid progress of artificial intelligence over the past decade owes much to a class of algorithms called transformer neural networks. Transformers gave rise to large language models (LLMs) like GPT-4 or Claude 2 that display impressive natural language abilities.</p><p><br></p><p>But as AI becomes more integrated into business processes, sole reliance on data-driven machine learning approaches like transformers may prove limiting. Researchers are exploring how to combine these powerful statistical models with neurosymbolic methods that incorporate logical reasoning and structure.</p><p><br></p><p>The result could be AI systems that blend raw pattern recognition power with human-like compositional generalization and interpretability.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">The Rise of Large Language Models</span></p><p><br></p><p>Much of the current excitement around AI stems from the advances of LLMs over the past few years. Models like GPT-3, PaLM, and Google's LaMDA have shown the ability to generate human-like text, answer questions, and accomplish tasks from basic prompts.</p><p><br></p><p>LLMs owe their abilities to a neural network architecture called transformers. Transformers process text more holistically than previous recurrent neural networks. They capture long-range dependencies in language by attending to all words in a context.</p><p><br></p><p>Training transformers on massive text corpora like the internet produces universal language models. With enough data and compute, these models learn statistical representations that prove surprisingly versatile for language tasks.</p><p><br></p><p>Finetuning techniques allow specializing LLMs to specific applications by updating the models on task data. For example, a finetuned GPT-3 model can be adapted into a conversational chatbot or a code completion tool.</p><p><br></p><p>The broad capabilities of LLMs along with their ease of use via prompting led to widespread adoption. Startups like Anthropic and Cohere are commercializing LLMs for business use cases. Apps built on LLMs range from automating customer support to generating content to synthesizing code.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Limits of Language Models</span></p><p></p><p><br></p><p>But for all their progress, LLMs still suffer from key limitations. Most notably, they display limited compositional generalization outside the distribution of their training data. For example, a LLM trained on English text will struggle with novel sentence structures or made-up words.</p><p><br></p><p>Humans seamlessly compose known concepts into new combinations thanks to our intuitive understanding of language syntax and meaning. Neural networks have no such innate symbolic reasoning capabilities.</p><p><br></p><p>LLMs are also black boxes. They can generate plausible and useful text or code but offer no interpretable justification for their outputs. Lack of interpretability makes it hard to audit models or identify causes of failures.</p><p><br></p><p>Finally, the massive scale of data and compute required to train LLMs makes them environmentally costly. Requiring less data and smaller models would allow much wider deployment of AI technology.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Integrating Symbolic Representations</span></p><p></p><p><br></p><div style="color:inherit;"><p>To overcome the limits of language models, researchers are finding ways to incorporate more logical reasoning. The aim is to complement the statistical learning with capabilities closer to human understanding.</p><p><br></p><p>One approach injects structured knowledge representations into the training process. For example, some methods jointly train the language model with a knowledge graph. <span style="color:inherit;">Knowledge graphs are data structures that represent facts as networks of entities and relationships. They encode real-world knowledge in a machine-readable graph format with nodes for entities like people and edges for relationships like &quot;employed at&quot;. This allows computers to automatically reason over millions of interconnected facts. Knowledge graphs help power many AI applications today including search, recommendations, and question answering. </span>The knowledge graph acts like a symbolic memory bank to improve reasoning.</p><p><br></p><p>Other techniques draw inspiration from classic logic programming languages like Prolog. These languages represent knowledge as human-readable rules. By integrating them into the training, the aim is to bake in more systematic symbolic thinking.</p><p><br></p><p>Researchers are also finding ways to refine and check language model outputs using logical constraints. For instance, one idea runs the text through separate logic rules as an extra plausibility filter beyond the statistical patterns.</p><p><br></p><p>In each case, the goal is to guide, restrict, and enhance the pattern-finding abilities of language models with more deliberate symbolic reasoning. Just like humans blend intuitive thinking with logic, the hope is to achieve AI systems that integrate learned statistical correlations with structured symbolic representations.</p><p><br></p><p>The end result could be models that display more generalized reasoning abilities, while also producing outputs we can audit, validate, and explain.</p></div><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Towards Hybrid Intelligence</span></p><p></p><p><br></p><p>Ultimately, the aim is achieving hybrid systems that integrate the complementary strengths of neural and symbolic AI. Some <span style="color:inherit;">researchers</span> argue that intelligence emerges from the interplay of two mechanisms:</p><ul><li>Correlation-based pattern recognition that is data-driven and associative.</li><li>Model-based compositional generalization relying on structured representations and explicit rules.</li></ul><p><br></p><p>Large transformer networks excel at the former while neurosymbolic methods specialize in the latter. Combining these two modes of reasoning could thus give rise to more human-like artificial intelligence.</p><p><br></p><p>The business implications of such hybrid AI systems are far-reaching. Logical components would allow verifying conclusions, checking ethical compliance, and generating step-by-step explanations. Incorporating domain constraints would reduce data needs and may lead to safer and less environmentally costly systems.</p><p><br></p><p>At the same time, retaining differentiable components preserves versatility, allows critiquing and updating symbolic knowledge, and facilitates integrating with downstream machine learning tasks.</p><p><br></p><p>Realizing this vision of integrated reasoning poses research challenges. Tradeoffs exist between symbolic interpretability and neural flexibility. Multi-component systems risk bottlenecks limiting end-to-end learning. Architectures that blur gradients across reasoning layers may be needed.</p><p><br></p><p>Nonetheless, the potential payoff for deployable, ethical, and broadly capable AI merits investment into these hybrid systems. Given the enthusiasm around LLMs today, injecting connections to symbolic reasoning could be a crucial next step in fulfilling their promise while mitigating risks.</p><p><br></p><p><span style="color:inherit;">Blending logical rule-based reasoning with modern neural networks could create more capable and reliable AI systems. This combination of human-like symbolic thinking and data-driven pattern recognition represents an exciting path forward. The result may be AI that better aligns with human intelligence in terms of adaptability, efficiency, and trustworthiness. Integrating the strengths of both of these approaches could lead to more advanced and human-compatible AI.</span></p><p><span style="color:inherit;"><br></span></p><p><span style="color:inherit;">Sources:</span></p><div style="color:inherit;"><a href="https://arxiv.org/abs/2205.11916" title="Constraining large language models with logic." rel="">Constraining large language models with logic</a><br></div><div style="color:inherit;"><a href="https://arxiv.org/abs/2302.07819" title="Neurologic decoding improves logical consistency of text generated by large language models." rel="">Neurologic decoding improves logical consistency of text generated by large language models</a></div><div style="color:inherit;"><a href="https://arxiv.org/abs/2305.13179" title="Teaching transformers to systematically reason with differentiable logic." rel="">Teaching transformers to systematically reason with differentiable logic</a></div><div style="color:inherit;"><a href="https://arxiv.org/abs/2012.13048" title="ProofWriter: Generating Implications, Proofs, and Abductive Statements over Natural Language." rel="">ProofWriter: Generating Implications, Proofs, and Abductive Statements over Natural Language</a></div><div style="color:inherit;"><a href="https://arxiv.org/abs/1909.03193" title="KG-BERT: BERT for knowledge graph completion." rel="">KG-BERT: BERT for knowledge graph completion</a></div><div style="color:inherit;"><a href="https://arxiv.org/abs/2305.13179" title="Neuro-symbolic concept learner: Discovering objects and their properties." rel="">Neuro-symbolic concept learner: Discovering objects and their properties</a><br></div><p></p></div></div><p></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 18:34:26 +1000</pubDate></item><item><title><![CDATA[Testing AI's Ability to Understand Language in Context]]></title><link>https://www.nownextlater.ai/Insights/post/testing-ai-s-ability-to-understand-language-in-context</link><description><![CDATA[Researchers have developed a benchmark called the LAMBADA dataset to rigorously test how well AI models can leverage broader discourse context when predicting an upcoming word.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_xF4t6QesR8uxc92FhVi5Gw" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_8bhpItgQSkqzqqtLxxL5eQ" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_yTn8c-kASd-8tBJB9x60Aw" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"> [data-element-id="elm_yTn8c-kASd-8tBJB9x60Aw"].zpelem-col{ border-radius:1px; } </style><div data-element-id="elm_Mrls-pd6TVySli4Sre_gpQ" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_Mrls-pd6TVySli4Sre_gpQ"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Artificial intelligence has made great strides in natural language processing in recent years. Systems can now translate text, answer questions, and generate coherent paragraphs on demand. However, most AI still struggles with true language understanding that requires integrating information across long texts.</p><p><br></p><p><span style="color:inherit;">Back in 2016, </span>to address this limitation, researchers developed a benchmark called the LAMBADA dataset to rigorously test how well AI models can leverage broader discourse context when predicting an upcoming word.</p><p><br></p><p>LAMBADA contains over 10,000 passages extracted from fiction books, with the last word blanked out in each passage. When humans are given the full passage as context, they can easily guess the missing word. However, if humans only see the final sentence containing the blank, it becomes virtually impossible to predict the missing word.</p><p><br></p><p>For example, the sentence &quot;Do you honestly think that I would want you to have a ?&quot; on its own has many plausible words that could fill in the blank. But when given the full passage about a couple discussing pregnancy concerns beforehand, it becomes clear from the context that the missing word is &quot;miscarriage.&quot;</p><p><br></p><p>The researchers tested a wide range of AI systems on LAMBADA, including statistical n-gram models as well as advanced neural network architectures like LSTMs. Back then, all the models performed extremely poorly, with 0% to 7% accuracy in predicting the missing word. The models often relied on simple techniques like picking a random proper noun from the passage. Even methods designed to track broader context failed to match human performance. LAMBADA continues to be used today too test new projects such as <a href="https://blog.novelai.net/a-new-model-clio-is-coming-to-opus-ef4e2457c601" title="Novel AI" rel="">Novel AI</a>, and this time Models are performing with over 70% accuracy.<br></p><p></p><p><br></p><p>Truly intelligent systems will need to integrate information across long passages and reason about that context to understand language the way people do.</p><p><br></p><p>While AI chatbots and virtual assistants are improving customer service and other applications, they cannot yet achieve the sophistication of human context processing. Benchmarks like LAMBADA push innovators to develop the next generation of AI that skillfully uses context instead of relying on surface-level statistical patterns.</p><p><br></p><p>Just as IQ tests expanded to gauge different types of intelligence beyond a single number, benchmarks like LAMBADA are important for building well-rounded language AI systems. Advancing contextual language understanding will enable more fluent, trustworthy interfaces between people and machines. Whether in customer service or product development, AI that masters using context could unlock new levels of human-computer interaction.</p><p><br></p><p>Sources:</p><p><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:16px;"><a href="https://www.researchgate.net/publication/306093716_The_LAMBADA_dataset_Word_prediction_requiring_a_broad_discourse_context" title="The LAMBADA dataset: Word prediction requiring a broad discourse context" rel="">The LAMBADA dataset: Word prediction requiring a broad discourse context</a></span></p><p></p><p></p></div>
<p></p></div></div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:08:00 +1000</pubDate></item><item><title><![CDATA[Filling in the Blanks: AI Learns to Suggest Missing Pieces of Stories]]></title><link>https://www.nownextlater.ai/Insights/post/filling-in-the-blanks-ai-learns-to-suggest-missing-pieces-of-stories</link><description><![CDATA[AI research from 2019 explored how to automatically generate reasonable suggestions for missing sections of text.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_SY8i5KsBTYSBUTnW4F6Vog" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_owVqUUNJTL6eYSWPLHXg-Q" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_B4yMguWGR2e1JyIJ2453AA" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_smgWUcsE8U2E7JFjJXRSIg" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_smgWUcsE8U2E7JFjJXRSIg"] .zpimage-container figure img { width: 500px ; height: 407.45px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_smgWUcsE8U2E7JFjJXRSIg"] .zpimage-container figure img { width:500px ; height:407.45px ; } } @media (max-width: 767px) { [data-element-id="elm_smgWUcsE8U2E7JFjJXRSIg"] .zpimage-container figure img { width:500px ; height:407.45px ; } } [data-element-id="elm_smgWUcsE8U2E7JFjJXRSIg"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-09%20at%2011.51.47%20pm.png" width="500" height="407.45" loading="lazy" size="medium" alt="In the one stage baseline, the missing span is predicted given the context and the target length." data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_moUMIMAonYIEMJV_e5K92g" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_moUMIMAonYIEMJV_e5K92g"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><div style="color:inherit;"><p>Stories unfold step-by-step, but writers sometimes get stuck on how to connect one part to the next. AI research from 2019 explored how to automatically generate reasonable suggestions for missing sections of text. This &quot;story infilling&quot; aimed to assist creative writing by proposing ideas that align with the existing story while still surprising the author.</p><p><br></p><p>The researchers found that standard AI language models at the time struggled to balance coherence with novelty when filling in gaps. The generated text ended up too boring or too random. To address this limitation, they designed a two-step hierarchical system:</p><ul><li>First, the AI randomly selected a few rare, interesting words that could plausibly fit into the storyline based on the context. For a medieval fantasy passage, it might suggest words like &quot;dragon,&quot; &quot;princess,&quot; or &quot;castle.&quot; The system focused on rare words since they provide more information to guide the rest of the text.</li><li>Second, the system generated full sentences conditioned on those interesting words, searching likely combinations that form coherent text. Leveraging the rare words prevented repetitive suggestions, while allowing the model to focus on fluency and coherence.</li></ul><p><br></p><p>The researchers tested story infilling on passages from children's tales with missing sections of 15-30 words. Human evaluators preferred the hierarchical model's suggestions over non-hierarchical methods, which sacrificed diversity or quality.</p></div></div></div>
</div><div data-element-id="elm_smtSU1vMQJ-NiRm8qmX3eQ" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_smtSU1vMQJ-NiRm8qmX3eQ"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;">While an early attempt, the study shows promise for AI-assisted writing tools. The approach mirrors a writer's workflow - first deciding on key ideas, then piecing together suitable wording. Similar techniques may enable more human-like narrative understanding and creativity.<p><br></p><p>The field has greatly advanced since 2019 with models like Claude and GPT-4. Yet even powerful AI still struggles with high-level plot and character consistency. Explicitly decomposing generation into steps of planning and drafting, as humans do, is one way to address these challenges. While AI cannot replace human creativity, structured models could soon provide useful brainstorming and revision tools for real authors.</p><p><br></p><p>Sources:</p><p><span style="color:inherit;"><a href="https://www.seas.upenn.edu/%7Eccb/publications/story-infilling.pdf" title="Unsupervised Hierarchical Story Infilling" rel="">Unsupervised Hierarchical Story Infilling</a></span></p><p></p><br></div></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:07:37 +1000</pubDate></item><item><title><![CDATA[Behind the Scenes of Storytelling: Using AI to Plan and Structure Narratives]]></title><link>https://www.nownextlater.ai/Insights/post/behind-the-scenes-of-storytelling-using-ai-to-plan-and-structure-narratives</link><description><![CDATA[In 2019, researchers explored how artificial intelligence could use hierarchical models to improve computer-generated stories.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_zGmGX_ExR0OL3dEJ-qE92A" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_Sgf28QJsRkyTptjDigI1Zg" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_2nW8rLDxSvSle-fxQu_DRw" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"> [data-element-id="elm_2nW8rLDxSvSle-fxQu_DRw"].zpelem-col{ border-radius:1px; } </style><div data-element-id="elm_CrdU5M3JEZRY4kdPjaU4rQ" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_CrdU5M3JEZRY4kdPjaU4rQ"] .zpimage-container figure img { width: 500px ; height: 662.94px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_CrdU5M3JEZRY4kdPjaU4rQ"] .zpimage-container figure img { width:500px ; height:662.94px ; } } @media (max-width: 767px) { [data-element-id="elm_CrdU5M3JEZRY4kdPjaU4rQ"] .zpimage-container figure img { width:500px ; height:662.94px ; } } [data-element-id="elm_CrdU5M3JEZRY4kdPjaU4rQ"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-tablet-align-center zpimage-mobile-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-09%20at%2011.31.59%20pm.png" width="500" height="662.94" loading="lazy" size="medium" alt="Generating entity references for different genres" data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_A7S0QEDFQbSp0Oya0DQ8KA" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_A7S0QEDFQbSp0Oya0DQ8KA"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><p></p><p></p><div style="color:inherit;"><p></p><p>Storytelling seems almost magical. Writers conjure up entire worlds from their imaginations. But even master storytellers rely on plans and outlines to craft complex, coherent narratives spanning hundreds of words. In 2019, researchers explored how artificial intelligence could similarly use hierarchical models to improve computer-generated stories.</p><p><br/></p><p>Up until then, most AI systems created stories simply word-by-word from left to right. While fine for short texts, this method struggled with long-term plot and character consistency. The researchers proposed &quot;coarse-to-fine&quot; techniques to first generate story outlines, then build surface-level details conditioned on the outline.</p><p><br/></p><p>Their approach involved three steps: modeling the sequence of actions using verbs and arguments, generating story sentences with placeholder entities like &quot;ent0&quot;, and finally rewriting the placeholders with specific references. This mirrored how human writers first sketch a plot's arc, then go back to flesh out settings and characters.</p><p><br/></p><p>By creating more structured drafts, the AI models improved event diversity and entity consistency compared to previous approaches. The placeholder entities also made it easier to track characters, replacing different mentions with the same token. The researchers found that human judges strongly preferred stories created with hierarchical planning versus direct generation.</p><p><br/></p><p>While an early attempt, this work showed the promise of mimicking writing strategies like outlining and revising. The field has advanced rapidly since 2019 as models like GPT-4 or Claude 2 now generate amazingly fluent text. But behind the scenes, AI still struggles with plot and people - areas where hierarchical techniques could help. The research highlights the value of breaking narration into more human-like steps. A technique currently being explored by several&nbsp;<span style="color:inherit;">AI-assisted writing</span> startups such as <a href="https://novelai.net/" title="Novel AI" rel="">Novel AI</a> and <a href="https://www.sudowrite.com/" title="Sudowrite" rel="">Sudowrite</a>.<br/></p><p></p><p></p><p><br/></p><p>Just as outlines aid human storytellers, explicit planning and revisions may allow AI to better learn from experience. More structured generation spaces let models focus on specific challenges like action sequences before full text. While AI has seen stunning progress, people remain the masters of storycraft. Studying the narrative strategies of writers may guide systems to become more helpful to writers.<br/></p><p><br/></p><p>Source:</p><p><span style="color:inherit;"><a href="https://arxiv.org/pdf/1902.01109.pdf" title="Strategies for Structuring Story Generation" rel="">Strategies for Structuring Story Generation</a></span></p><p></p></div></div>
</div><div data-element-id="elm_FPPf8SQKIdBhofTA6erWbQ" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_FPPf8SQKIdBhofTA6erWbQ"] .zpimage-container figure img { width: 1090px ; height: 773.22px ; } } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-tablet-align-center zpimage-mobile-align-center zpimage-size-fit zpimage-tablet-fallback-fit zpimage-mobile-fallback-fit "><figure role="none" class="zpimage-data-ref"><a class="zpimage-anchor" href="https://www.reel-intelligence.org/" target="" title="Reel intelligence" rel=""><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Winner%20TV%20Pilot%20Screenplay%20-2-.png" size="fit" alt="Reel Intelligence"/></picture></a></figure></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:07:19 +1000</pubDate></item><item><title><![CDATA[Reading Between the Lines: Using Math to Uncover Hidden Patterns in Books]]></title><link>https://www.nownextlater.ai/Insights/post/reading-between-the-lines-using-math-to-uncover-hidden-patterns-in-books</link><description><![CDATA[Books may seem like straightforward stories, but researchers are finding fascinating mathematical patterns hidden in the text.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_BN-GUz9yTH6qOGKp8XYI2Q" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_6P1Nxs5lTRaNx8x9Z_sLXA" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_XhJZ4QBFRB6SJ6gQWdY-bQ" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_wDEEkImbYnmEBHGpW6Ludw" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_wDEEkImbYnmEBHGpW6Ludw"] .zpimage-container figure img { width: 500px ; height: 525.08px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_wDEEkImbYnmEBHGpW6Ludw"] .zpimage-container figure img { width:500px ; height:525.08px ; } } @media (max-width: 767px) { [data-element-id="elm_wDEEkImbYnmEBHGpW6Ludw"] .zpimage-container figure img { width:500px ; height:525.08px ; } } [data-element-id="elm_wDEEkImbYnmEBHGpW6Ludw"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-09%20at%2010.18.41%20pm.png" width="500" height="525.08" loading="lazy" size="medium" alt="An ‘ousiogram’ (Dodds et al., 2021) displaying power and danger scores for a subset of 14,499 unique words appearing in Terry Pratchett’s 41-book Discworld series." data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_1minUEv_RVGhhrISEzPBbA" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_1minUEv_RVGhhrISEzPBbA"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Books may seem like straightforward stories, but researchers are finding mathematical patterns hidden in the text. By tracking how words are used over the course of a book in minute detail, they can reveal new insights into plot, emotion, and structure that are not visible to the naked eye.</p><p><br></p><p>The researchers started by scoring a large number of words based on their emotional meaning. For example, positive words like &quot;love&quot; scored higher while negative words like &quot;war&quot; scored lower. They used a framework called &quot;ousiometrics&quot; which boils down emotions to two key dimensions: power and danger. Power relates to agency, confidence, and positivity. Danger relates to emotional uncertainty, negativity, and aggression.</p><p><br></p><p>They then took thousands of books and broke them down into short segments of 50 words each. For each segment, they calculated the average power and danger scores based on the words present. This turned each book into a rolling wave of numbers, with peaks representing more emotional sections and valleys as more neutral parts.</p><p><br></p><p>Short books generally showed a steady wave pattern while long books had more fluctuations in emotion over the course of the text. Surprisingly, when they zoomed in on long books they found the fluctuating highs and lows had a consistent length of a few thousand words. This matches the typical length of chapters in published fiction.</p><p><br></p><p>To study the patterns further, the researchers used a technique called empirical mode decomposition that breaks down fluctuations in data into distinct components, much like musical notes make up chords. The text segments were also compared to &quot;shuffled&quot; versions of the books with random word order. The real books differed from the random versions after a certain decomposition level, indicating that the fluctuations were not random but reflected an underlying structure.</p><p><br></p><p>These findings suggest longer books have a wave-like shape that is closer to collections of short stories or chapters. The emotional ups and downs of the text cycle on a scale of thousands of words, perhaps reflecting how long the human brain can comfortably process a complex narrative before needing a reset. Shorter books lacked these larger fluctuations.</p><p><br></p><p>While we intuitively understand how passages evoke certain moods, the researchers were able to quantify the pacing of emotional highs and lows mathematically. Their work helps confirm the existence of nested patterns in writing - punctuation gives phrases, paragraphs offer local structure, chapters provide mid-level segments, and over the full book arcs emerge.</p><p><br></p><p>So the next time you open a book, think about the hidden rhythms inside that subtly influence your experience. The feelings evoked in the story may follow mathematical waves as you steadily progress from cover to cover. This emerging field opens up new ways of appreciating the art and science of expert storytelling.</p><p><br></p><p>Sources:</p><p><a href="https://www.nature.com/articles/s41599-023-01680-4" title="A decomposition of book structure through ousiometric fluctuations in cumulative word-time" rel="">A decomposition of book structure through ousiometric fluctuations in cumulative word-time</a></p><p></p></div><p></p></div>
</div><div data-element-id="elm_BZslHjh1L1NAYTd780h3LQ" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_BZslHjh1L1NAYTd780h3LQ"] .zpimage-container figure img { width: 800px ; height: 344.00px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_BZslHjh1L1NAYTd780h3LQ"] .zpimage-container figure img { width:500px ; height:215.00px ; } } @media (max-width: 767px) { [data-element-id="elm_BZslHjh1L1NAYTd780h3LQ"] .zpimage-container figure img { width:500px ; height:215.00px ; } } [data-element-id="elm_BZslHjh1L1NAYTd780h3LQ"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-large zpimage-tablet-fallback-large zpimage-mobile-fallback-large "><figure role="none" class="zpimage-data-ref"><a class="zpimage-anchor" href="/aibooks" target="" rel=""><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Untitled%20design%20-4-.png" width="500" height="215.00" loading="lazy" size="large"/></picture></a></figure></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:06:53 +1000</pubDate></item><item><title><![CDATA[Storywrangler: Tracking Culture and Events through Twitter's Lens]]></title><link>https://www.nownextlater.ai/Insights/post/storywrangler-tracking-culture-and-events-through-twitter-s-lens</link><description><![CDATA[Researchers developed a tool called Storywrangler that leveraged Twitter data to create an "instrument for understanding our world through the lens of social media."]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_HszOwGyHS3u2ZbcwO98WUQ" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_NUndbqp2QH-raYQCD0h2-A" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_TG5dekBlReurU5nBQ7MK4A" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_HdtaSh_XqLf8A0UJCjxcLg" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_HdtaSh_XqLf8A0UJCjxcLg"] .zpimage-container figure img { width: 500px ; height: 386.22px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_HdtaSh_XqLf8A0UJCjxcLg"] .zpimage-container figure img { width:500px ; height:386.22px ; } } @media (max-width: 767px) { [data-element-id="elm_HdtaSh_XqLf8A0UJCjxcLg"] .zpimage-container figure img { width:500px ; height:386.22px ; } } [data-element-id="elm_HdtaSh_XqLf8A0UJCjxcLg"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-09%20at%2010.00.19%20pm.png" width="500" height="386.22" loading="lazy" size="medium" alt="Screenshot of the Storywrangler site showing example Twitter n-gram time series for the first half of 2020." data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_ooBK-RQ6Sx-jwMEa5QUoDQ" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_ooBK-RQ6Sx-jwMEa5QUoDQ"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><p class="whitespace-pre-wrap">Social media platforms like Twitter offered an unprecedented window into the real-time thoughts, conversations, and interests of millions of people. Researchers developed a tool called Storywrangler that leveraged Twitter data to create an &quot;instrument for understanding our world through the lens of social media.&quot;</p><p class="whitespace-pre-wrap"><br></p><p class="whitespace-pre-wrap">Storywrangler analyzed over 100 billion tweets dating back to 2008 to detect trends in word usage over time. It broke down tweets into &quot;n-grams&quot; - sequences of one, two, or three words - and tracked how the usage frequencies of these n-grams changed on a daily basis across different languages.</p><p class="whitespace-pre-wrap">This massive database allowed researchers to see how real-world events, from natural disasters to political movements, were reflected in the narratives that unfolded on Twitter. For example, Storywrangler revealed surging interest in climate-related terms during major storms and wildfires. And it captured the rapid rise and fall of hashtags associated with social justice protests. Beyond reacting to news, Twitter also mirrored more subtle cultural shifts, like the waxing and waning popularity of celebrities or diets.</p><p class="whitespace-pre-wrap"><br></p><p class="whitespace-pre-wrap">Storywrangler went beyond tracking raw frequencies - it also quantified how widely information spread on social media through shares and reposts. This helped distinguish niche conversations from truly viral ideas. The researchers used &quot;contagiograms&quot; to visualize both the popularity and amplification of n-grams over time.</p><p class="whitespace-pre-wrap"><br></p><p class="whitespace-pre-wrap">There were certainly limitations to the Twitter lens. The platform's user base skewed young, urban, and affluent compared to the general population. Bots and organized campaigns could artificially inflate interest in certain topics. And the meanings of words themselves evolved across the years.</p><p class="whitespace-pre-wrap"><br></p><p class="whitespace-pre-wrap">But used carefully, Storywrangler offered an unparalleled window into the collective consciousness - recording not just major news events but also the mundane daily conversations of millions worldwide. It aimed to complement more traditional data sources like books and news archives. The researchers hoped Storywrangler would enable more data-driven computational social science to understand our fast-changing, digitally-connected world.</p><p class="whitespace-pre-wrap"><br></p><p class="whitespace-pre-wrap">Source:</p><p class="whitespace-pre-wrap"><span style="color:inherit;"><a href="https://arxiv.org/pdf/2007.12988.pdf" title="Storywrangler: A massive exploratorium for sociolinguistic, cultural, socioeconomic, and political timelines using Twitter" rel="">Storywrangler: A massive exploratorium for sociolinguistic, cultural, socioeconomic, and political timelines using Twitter</a></span></p><p class="whitespace-pre-wrap"></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:05:59 +1000</pubDate></item><item><title><![CDATA[The Six Basic Emotional Story Arcs, According to Science]]></title><link>https://www.nownextlater.ai/Insights/post/the-six-basic-emotional-story-arcs-according-to-science</link><description><![CDATA[A 2016 study analyzed over a thousand stories to uncover the basic emotional arcs that form the building blocks of narratives.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_iJroV7Y1S-q_HYpadHOoFw" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_TJ-lGOB6RGqRjii5YBqo0w" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_Ux_5S_7-SbyGACCOmwa23A" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"> [data-element-id="elm_Ux_5S_7-SbyGACCOmwa23A"].zpelem-col{ border-radius:1px; } </style><div data-element-id="elm_rsNFUQqs7_GnBrqvomrYQw" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_rsNFUQqs7_GnBrqvomrYQw"] .zpimage-container figure img { width: 500px ; height: 465.71px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_rsNFUQqs7_GnBrqvomrYQw"] .zpimage-container figure img { width:500px ; height:465.71px ; } } @media (max-width: 767px) { [data-element-id="elm_rsNFUQqs7_GnBrqvomrYQw"] .zpimage-container figure img { width:500px ; height:465.71px ; } } [data-element-id="elm_rsNFUQqs7_GnBrqvomrYQw"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-tablet-align-center zpimage-mobile-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-09%20at%209.47.14%20pm.png" width="500" height="465.71" loading="lazy" size="medium" alt=" Schematic of how we compute emotional arcs" data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_o26FvbMXTDWN2cv7eQG6WQ" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_o26FvbMXTDWN2cv7eQG6WQ"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><p class="whitespace-pre-wrap">Stories have gripped humans for ages by evoking powerful emotions. Even years ago, researchers wondered - are there fundamental patterns underlying how tales tug our heartstrings? A 2016 study analyzed over a thousand stories to uncover the basic emotional arcs that form the building blocks of narratives.</p><p class="whitespace-pre-wrap"><br/></p><p class="whitespace-pre-wrap">Using digital methods, researchers at the University of Vermont quantified the moment-to-moment emotional trajectories of stories from the Project Gutenberg collection. They tracked sentiment throughout each book using a rolling average approach. This generated an &quot;emotional arc&quot; capturing how positivity rises and falls across the narrative.</p><p class="whitespace-pre-wrap"><br/></p><p class="whitespace-pre-wrap">Through data science techniques like matrix decomposition, clustering, and neural network mapping, six core emotional arcs emerged:</p><ol class="list-decimal pl-8 space-y-2"><li class="whitespace-normal">Rags to riches (rise)</li><li class="whitespace-normal">Tragedy or riches to rags (fall)</li><li class="whitespace-normal">Man in a hole (fall then rise)</li><li class="whitespace-normal">Icarus (rise then fall)</li><li class="whitespace-normal">Cinderella (rise, fall, then rise)</li><li class="whitespace-normal">Oedipus (fall, rise, then fall again)</li></ol><p class="whitespace-pre-wrap"><br/></p><p class="whitespace-pre-wrap">You can likely picture classic stories matching these shapes. Cinderella follows a rags-to-riches-to-rags-to-riches pattern. Oedipus the King exhibits a tragic fall, brief rise, then another fall. Each arc formally captures intuitions storytellers have traded on for ages.</p><p class="whitespace-pre-wrap"><br/></p><p class="whitespace-pre-wrap">Notably, the study found arcs with a fall-rise shape (man in a hole, Cinderella) were collectively most common, at around 30% of tales. The next most prevalent were tragedies at around 32%, then oedipal fall-rise-falls at 31%. Purely rising arcs were rarest at just 5% of stories.</p><p class="whitespace-pre-wrap"><br/></p><p class="whitespace-pre-wrap">This indicates emotional rollercoasters captivate us more than simple rises. Tragedy has proven perennially popular, despite leaving readers sad. Stories that plunge protagonists into despair before rising contain greater drama.</p><p class="whitespace-pre-wrap"><br/></p><p class="whitespace-pre-wrap">The analysis also uncovered that more complex arcs with multiple peaks and valleys enjoyed greater success by one metric: website downloads. Stories whose shape matched the Icarus, Oedipus, and double &quot;man in a hole&quot; arcs saw far more downloads than simpler arcs.</p><p class="whitespace-pre-wrap"><br/></p><p class="whitespace-pre-wrap">This hints the emotional journey impacts how narratives spread. The anguish of tragedy may make such tales powerfully shareable. Multi-phasic arcs may also hook readers through twists and turns.</p><p class="whitespace-pre-wrap"><br/></p><p class="whitespace-pre-wrap">Of course, downloads provide only a rough measure of success. Other factors like marketing and fame contribute. Still, the findings suggest crafting an evocative emotional trajectory helps stories resonate.</p><p class="whitespace-pre-wrap"><br/></p><p class="whitespace-pre-wrap">The study demonstrates how data science can unearth hidden patterns in the arts. Formalizing intuitions about entertainment with empirical evidence remains novel, intriguing territory.</p><p class="whitespace-pre-wrap"><br/></p><p class="whitespace-pre-wrap">Quantitative analysis today is even more powerful thanks to advances in machine learning and cultural analytics. New techniques promise insights unavailable to past generations of literary scholars.</p><p class="whitespace-pre-wrap"><br/></p><p class="whitespace-pre-wrap">Combining these digital humanities approaches with insight and traditional criticism will likely bear the richest fruits. As machines grow skilled at classifying sentiment and archetypes, what new discoveries await about the stories humans compulsively tell? What makes a narrative emotionally compelling transcends any one discipline.</p><p></p><p><br/></p><p><br/></p><p>Sources:</p><p><span style="color:inherit;"><a href="https://arxiv.org/pdf/1606.07772.pdf" title="The emotional arcs of stories are dominated by six basic shapes" rel="">The emotional arcs of stories are dominated by six basic shapes</a></span></p><p></p></div>
</div><div data-element-id="elm_a7jWJr3scwvRtbDULvojyA" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_a7jWJr3scwvRtbDULvojyA"] .zpimage-container figure img { width: 1090px ; height: 773.22px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_a7jWJr3scwvRtbDULvojyA"] .zpimage-container figure img { width:500px ; height:215.00px ; } } @media (max-width: 767px) { [data-element-id="elm_a7jWJr3scwvRtbDULvojyA"] .zpimage-container figure img { width:500px ; height:215.00px ; } } [data-element-id="elm_a7jWJr3scwvRtbDULvojyA"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-tablet-align-center zpimage-mobile-align-center zpimage-size-fit zpimage-tablet-fallback-large zpimage-mobile-fallback-large "><figure role="none" class="zpimage-data-ref"><a class="zpimage-anchor" href="https://www.reel-intelligence.org/" target="" title="Reel Intelligence" rel=""><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Winner%20TV%20Pilot%20Screenplay%20-2-.png" width="500" height="215.00" loading="lazy" size="fit"/></picture></a></figure></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:05:37 +1000</pubDate></item><item><title><![CDATA[Teaching AI to Tell Better Tales by Integrating External Knowledge]]></title><link>https://www.nownextlater.ai/Insights/post/Teaching-AI-to-Tell-Better-Tales-by-Integrating-External-Knowledge</link><description><![CDATA[New research explores how integrating structured knowledge into AI systems can enhance storytelling abilities.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_-2LoK4FLSj6kE0maiRXg6A" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_2uZedO7XTwmjrBCrAZQSKw" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_Tz4iZq4gSd6jgmc03jak_g" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_2YA5oGBzkoYbOgntrUGhXA" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_2YA5oGBzkoYbOgntrUGhXA"] .zpimage-container figure img { width: 500px ; height: 389.03px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_2YA5oGBzkoYbOgntrUGhXA"] .zpimage-container figure img { width:500px ; height:389.03px ; } } @media (max-width: 767px) { [data-element-id="elm_2YA5oGBzkoYbOgntrUGhXA"] .zpimage-container figure img { width:500px ; height:389.03px ; } } [data-element-id="elm_2YA5oGBzkoYbOgntrUGhXA"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-09%20at%209.41.19%20pm.png" width="500" height="389.03" loading="lazy" size="medium" alt="Three layers of narratological concepts about story: fabula, plot, and discourse." data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_kjuSNjtuTimsXg6Qn73jCw" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_kjuSNjtuTimsXg6Qn73jCw"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Storytelling comes naturally to humans. But for machines, spinning an engaging narrative remains an elusive goal. While AI can generate remarkably fluent text, its tales often lack coherence or get repetitive. New research explores how integrating structured knowledge into AI systems can enhance storytelling abilities.</p><p><br></p><p>When reading a story, we draw on general knowledge about how events logically unfold and characters plausibly act. We track complex plot threads and fill gaps using common sense. Machines lack this innate understanding we take for granted. Their stories can become nonsensical or contradictory.</p><p>To tackle this, researchers are providing AI systems explicit knowledge in structured formats. This external knowledge acts like a guide, keeping machine-generated plots on track. It also helps avoid stale repetitions by expanding the ideas available to pull from.</p><p><br></p><p>Several common limitations plague today's AI storytellers:</p><ul><li>Lack of long-term coherence. Without a sense of overall narrative arc, they ramble aimlessly.</li><li>Insufficient grounding in real-world facts. Stories come off vague rather than richly descriptive.</li><li>Repetition. They loop the same words and phrases like a broken record.</li><li>Hallucination. They fabricate events that don't logically follow.</li></ul><p><br></p><p>Integrating knowledge resources like ConceptNet, which contains common sense facts about the world, alleviates these issues. The knowledge functions like an annotated outline, steering the plot. It also provides a memory bank of concepts to reference, varying the content.</p><p><br></p><p>But effectively harnessing external knowledge remains challenging. Two main strategies have emerged:</p><ol><li>Injecting knowledge directly into the AI system's training process, like teaching a human author.</li><li>Using knowledge as an external guiding reference during story generation.</li></ol><p><br></p><p>Each approach has trade-offs. Weighting structured resources too strongly can pollute the system's original language skills. But using knowledge merely as a loose guide can fail to correct nonsensical narration.</p><p><br></p><p>Striking the right balance is an active research problem. Scientists are also expanding the knowledge available to AI storytellers with new databases. Most systems today use generic common sense facts. But resources detailing specific people, places, and events could enable more detailed, vivid storytelling.</p><p><br></p><p>Automating evaluation also poses difficulties. No single &quot;correct&quot; story exists for a given prompt. Automatic metrics struggle to account for creativity and interest - aspects requiring human judgment. More robust evaluation is critical to gauge progress.</p><p><br></p><p>Despite hurdles, knowledge-infused narration clearly improves coherence, factual grounding, and variation. AI authors with a knowledge boost spin far more convincing yarns. The research provides a roadmap for machines to better mimic core elements of human storytelling.</p><p><br></p><p>Rather than viewing imagination and structure as at odds, they are complementary. Master storytellers combine free-flowing creativity with purposeful intent. By fusing extensive knowledge with unrestrained generation, machines inch closer toward unlocking that balancing act.</p><p><br></p><p>Sources:</p><p><span style="color:inherit;"><a href="https://arxiv.org/pdf/2212.04634.pdf" title="Open-world Story Generation with Structured Knowledge Enhancement: A Comprehensive Survey" rel="">Open-world Story Generation with Structured Knowledge Enhancement: A Comprehensive Survey</a></span></p><p></p></div><p></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:05:08 +1000</pubDate></item></channel></rss>