<?xml version="1.0" encoding="UTF-8" ?><!-- generator=Zoho Sites --><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><atom:link href="https://www.nownextlater.ai/Insights/ai-story-brain/feed" rel="self" type="application/rss+xml"/><title>Now Next Later AI - Blog , AI Story Brain</title><description>Now Next Later AI - Blog , AI Story Brain</description><link>https://www.nownextlater.ai/Insights/ai-story-brain</link><lastBuildDate>Wed, 26 Nov 2025 21:22:36 +1100</lastBuildDate><generator>http://zoho.com/sites/</generator><item><title><![CDATA[What is Tokenization? Let's Explore, Using Novel AI's New Tokenizer as a Use Case]]></title><link>https://www.nownextlater.ai/Insights/post/does-novel-ai-s-new-tokenizer-really-boost-storytelling</link><description><![CDATA[What is Tokenization? Let's Explore, Using Novel AI's New Tokenizer as a Use Case]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_j92KpyjVQguMwwTuLUblCw" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_y3ppOIQmQPatJLZsEqQmHA" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_rPQvraI_QoWt9m8irgPCow" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_SotbxYsxTU7FCItxtgBEhg" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_SotbxYsxTU7FCItxtgBEhg"] .zpimage-container figure img { width: 500px ; height: 492.01px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_SotbxYsxTU7FCItxtgBEhg"] .zpimage-container figure img { width:500px ; height:492.01px ; } } @media (max-width: 767px) { [data-element-id="elm_SotbxYsxTU7FCItxtgBEhg"] .zpimage-container figure img { width:500px ; height:492.01px ; } } [data-element-id="elm_SotbxYsxTU7FCItxtgBEhg"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-10%20at%2011.38.41%20am.png" width="500" height="492.01" loading="lazy" size="medium" alt="Novel AI's Tokenizer" data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_M6sQVlnHS5-6esy1xFZy5w" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_M6sQVlnHS5-6esy1xFZy5w"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><div style="color:inherit;"><p style="margin-bottom:12pt;"><span style="font-size:10.5pt;font-weight:400;">Tokenization is a foundational step in natural language processing (NLP) and machine learning.&nbsp;</span></p><p style="margin-bottom:12pt;"><span style="font-size:10.5pt;font-weight:400;">Large Language Models are big statistical calculators that work with numbers, not words. Tokenisation converts the words into numbers, with each number representing a position in a dictionary of all the possible words.</span></p><p style="margin-bottom:12pt;"><span style="font-size:10.5pt;font-weight:400;">Tokenization breaks down a piece of text into smaller units, called &quot;tokens.&quot; These tokens can represent whole words, parts of words, or even multiple words in some languages. For instance, the sentence &quot;ChatGPT is fun!&quot; might be broken down into tokens like [&quot;Chat&quot;, &quot;G&quot;, &quot;PT&quot;, &quot; is&quot;, &quot; fun&quot;, &quot;!&quot;].</span></p><p style="margin-bottom:12pt;"><span style="font-size:10.5pt;font-weight:400;">You can choose from multiple tokenization methods, but it's crucial to consistently use the same method during both training and text generation.</span></p><h1 style="margin-bottom:12pt;"><span style="font-size:20pt;font-weight:400;">Why is this important for large language models?</span></h1><ol><li style="font-size:10.5pt;font-weight:400;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:700;">Understanding Context:</span><span style="font-size:10.5pt;font-weight:400;"> By breaking text into tokens, the model can process and understand the context around each token. It's like looking at each puzzle piece and understanding where it might fit in the bigger picture.</span></p></li><li style="font-size:10.5pt;font-weight:400;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:700;">Efficiency: </span><span style="font-size:10.5pt;font-weight:400;">Language models have a limit to how many tokens they can process at once. By tokenizing text, they can manage and process information more efficiently.</span></p></li><li style="font-size:10.5pt;font-weight:400;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:700;">Flexibility:</span><span style="font-size:10.5pt;font-weight:400;"> Different languages have different structures. Tokenization allows these models to be flexible and work with multiple languages. For example, in English, spaces often separate words, but in languages like Chinese, words are often clustered together without spaces. Tokenization helps the model handle such variations.</span></p></li><li style="font-size:10.5pt;font-weight:400;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:700;">Training: </span><span style="font-size:10.5pt;font-weight:400;">When training these models on vast amounts of text, tokenization ensures that the model learns from consistent and standardized units of text.</span></p></li></ol><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">In essence, for large language models, tokenization is a foundational step that allows them to read, understand, and generate human-like text across various languages and contexts.</span></p><h1 style="margin-bottom:10pt;"><span style="font-size:20pt;font-weight:400;">Trade-offs</span></h1><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Different tokenization strategies come with their own sets of trade-offs. Let's delve into some of these key trade-offs:</span></p><ol><li style="font-size:10.5pt;font-weight:700;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:700;">Granularity:</span></p></li></ol><ul><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Fine-grained (e.g., character-level):</span></p></li><ul><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Pros: Can handle any word or term, even if it's never seen it before. It's very flexible and can be language-agnostic.</span></p></li><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Cons: Requires more tokens to represent a text, which can be computationally expensive and may not capture semantic meanings as effectively.</span></p></li></ul><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Coarse-grained (e.g., word-level):</span></p></li><ul><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Pros: Can capture semantic meanings more directly and is often more efficient in terms of the number of tokens used.</span></p></li><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Cons: Struggles with out-of-vocabulary words and might not be as flexible across different languages.</span></p></li></ul></ul><ol start="2"><li style="font-size:10.5pt;font-weight:700;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:700;">Language Dependence:</span></p></li></ol><ul><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Language-specific tokenizers:</span></p></li><ul><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Pros: Optimized for a particular language, capturing its nuances and structures effectively.</span></p></li><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Cons: Not versatile. A separate tokenizer would be needed for each language, which isn't scalable for multilingual models.</span></p></li></ul><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Language-agnostic tokenizers:</span></p></li><ul><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Pros: Can be used across multiple languages, making them ideal for multilingual models.</span></p></li><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Cons: Might not capture the intricacies of each individual language as effectively as a language-specific tokenizer.</span></p></li></ul></ul><ol start="3"><li style="font-size:10.5pt;font-weight:700;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:700;">Fixed vs. Dynamic Vocabulary:</span></p></li></ol><ul><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Fixed Vocabulary:</span></p></li><ul><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Pros: Deterministic and consistent in its tokenization. Easier to manage and deploy.</span></p></li><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Cons: Struggles with out-of-vocabulary terms and might become outdated as language evolves.</span></p></li></ul><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Dynamic (or adaptive) Vocabulary:</span></p></li><ul><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Pros: Can adjust to new terms or slang, making it more flexible and up-to-date.</span></p></li><li style="font-size:11pt;font-weight:400;margin-left:36pt;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:400;">Cons: More complex to implement and might introduce inconsistencies over time.</span></p></li></ul></ul><ol start="4"><li style="font-size:10.5pt;font-weight:400;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:700;">Efficiency vs. Coverage:</span><span style="font-size:10.5pt;font-weight:400;"> Some tokenizers aim for maximum coverage, ensuring they can handle any text thrown at them. Others prioritize efficiency, using the fewest tokens possible to represent a text. There's a balance to strike here: more coverage can mean more computational overhead, while prioritizing efficiency might mean sacrificing the ability to handle rare terms.</span></p></li><li style="font-size:10.5pt;font-weight:400;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:700;">Complexity and Overhead: </span><span style="font-size:10.5pt;font-weight:400;">Advanced tokenization methods, like Byte-Pair Encoding (BPE) or SentencePiece, can handle a wide range of text types and languages. However, they introduce additional computational and implementation overhead compared to simpler methods.</span></p></li><li style="font-size:10.5pt;font-weight:400;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:700;">Consistency:</span><span style="font-size:10.5pt;font-weight:400;"> Some tokenization methods might tokenize the same text differently based on context, leading to potential inconsistencies. While this can be beneficial in capturing nuanced meanings, it can also introduce unpredictability in the model's behavior.</span></p></li></ol><p style="margin-bottom:12pt;"><span style="font-size:10.5pt;font-weight:400;">Choosing a tokenizer for a large language model involves weighing these trade-offs based on the specific goals and constraints of the project. Whether the priority is multilingual support, computational efficiency, or capturing linguistic nuances, understanding these trade-offs is crucial in making an informed decision.</span></p><h1 style="margin-bottom:12pt;"><span style="font-size:20pt;font-weight:400;">Novel AI's Tokenizer to </span><span style="font-size:20pt;font-weight:400;font-style:italic;">&quot;enable stronger storytelling capabilities&quot;</span></h1><p style="margin-bottom:12pt;"><span style="font-size:11pt;font-weight:400;">Novel AI recently developed a custom tokenizer for their AI models. The </span><a href="https://github.com/NovelAI/novelai-tokenizer"><span style="font-size:11pt;font-weight:400;text-decoration:underline;">GitHub project</span></a><span style="font-size:11pt;font-weight:400;"> detailing their process provides an inside look at engineering tradeoffs like vocabulary size, compression ratio, and handling numerals. The focus of this project is to build a tokenizer that enables stronger storytelling capabilities. Let's explore!</span></p><p style="margin-bottom:12pt;"><span style="font-size:11pt;font-weight:400;">On the surface, the new tokenizer offers clear advantages:</span></p><ol><li style="font-size:11pt;font-weight:400;"><p><span style="font-size:11pt;font-weight:700;">Improved Granularity and Flexibility:</span><span style="font-size:11pt;font-weight:400;"> Unlike traditional tokenizers, this one offers a balance between </span><span style="font-size:11pt;font-weight:400;text-decoration:underline;background-color:rgb(2, 184, 187);">word-level and subword-level tokenization</span><span style="font-size:11pt;font-weight:400;">. By breaking down words into meaningful fragments, it can better understand and generate nuanced text. This is especially crucial for storytelling where context, nuance, and subtlety matter.</span></p></li></ol><br><ol start="2"><li style="font-size:11pt;font-weight:400;"><p><span style="font-size:11pt;font-weight:700;">Compression Ratio: </span><span style="font-size:11pt;font-weight:400;">A higher compression ratio means the model can process and understand larger chunks of text at once. </span><span style="font-size:11pt;font-weight:400;text-decoration:underline;background-color:rgb(2, 184, 187);">This is vital for maintaining context in long narratives or when referencing earlier parts of a story</span><span style="font-size:11pt;font-weight:400;">. By achieving a 7-19% higher compression ratio than the LLaMa tokenizer on significant parts of the English dataset, it's evident that the tokenizer is more efficient. This efficiency can translate to richer and more coherent narratives, especially in longer stories.</span></p></li></ol><br><ol start="3"><li style="font-size:11pt;font-weight:400;"><p><span style="font-size:11pt;font-weight:700;">Adaptability and Evolution: </span><span style="font-size:11pt;font-weight:400;">The iterative approach to tokenizer training, with multiple runs and rebalancing, ensures that the tokenizer is optimized for the specific nuances of your dataset. This adaptability is key for evolving storytelling styles and trends.</span></p></li></ol><br><p><span style="font-size:10.5pt;font-weight:400;">Some other pros are mentioned. These are more specific to the Novel AI project:</span></p><br><ol start="4"><li style="font-size:10.5pt;font-weight:400;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:700;">Multilingual Capabilities:</span><span style="font-size:10.5pt;font-weight:400;"> By accommodating both English and Japanese from the start, the tokenizer is designed for bilingual storytelling. This means it can seamlessly switch between languages or even blend them, offering richer narratives and reaching a broader audience.</span></p></li><li style="font-size:10.5pt;font-weight:400;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:700;">Efficient Handling of Unicode Characters: </span><span style="font-size:10.5pt;font-weight:400;">The ability to natively tokenize Unicode characters, especially emojis, allows for more expressive storytelling. Emojis, in modern communication, can convey emotions, context, and tone, making them valuable in narratives. But they are less relevant to novel writing.</span></p></li><li style="font-size:10.5pt;font-weight:400;"><p style="margin-bottom:10pt;"><span style="font-size:10.5pt;font-weight:700;">Numeric Understanding: </span><span style="font-size:10.5pt;font-weight:400;">Tokenizing numbers digit by digit enhances the model's capability to understand and manipulate numeric values. This is crucial for stories that involve dates, quantities, or any numerical context.</span></p></li></ol><br><p style="margin-bottom:12pt;"><span style="font-size:11pt;font-weight:700;">Disadvantages:</span></p><ol><li style="font-size:11pt;font-weight:400;"><p style="margin-bottom:10pt;"><span style="font-size:11pt;font-weight:700;">Complexity and Maintenance:</span><span style="font-size:11pt;font-weight:400;"> Training the tokenizer added development time and complexity.</span></p></li><li style="font-size:11pt;font-weight:400;"><p style="margin-bottom:10pt;"><span style="font-size:11pt;font-weight:700;">BPE vs. Unigram: </span><span style="font-size:11pt;font-weight:400;">The decision to choose BPE over Unigram was based on compression ratio. While BPE might offer better compression, Unigram might provide more natural word segmentations. <span style="background-color:rgb(2, 184, 187);">The storytelling quality might be affected if the tokenizer doesn't segment words in a way that's intuitive to human readers.</span></span></p></li><li style="font-size:11pt;font-weight:400;"><p style="margin-bottom:10pt;"><span style="font-size:11pt;font-weight:700;">Multilingual Limitations: </span><span style="font-size:11pt;font-weight:400;">While accommodating both English and Japanese is a strength, it might also be a limitation. The tokenizer might be overly specialized for these two languages, potentially making it less effective for other languages or multilingual contexts beyond English and Japanese.</span></p></li><li style="font-size:11pt;font-weight:400;"><p style="margin-bottom:10pt;"><span style="font-size:11pt;font-weight:700;">Vocabulary Size: </span><span style="font-size:11pt;font-weight:400;">The decision to use a vocabulary size of 65535 tokens, while efficient from a computational standpoint, might introduce limitations. Is this size sufficient to capture the nuances of both English and Japanese, especially given the richness of the Japanese writing system?</span></p></li><li style="font-size:11pt;font-weight:400;"><p style="margin-bottom:10pt;"><span style="font-size:11pt;font-weight:700;">Numeric Tokenization:</span><span style="font-size:11pt;font-weight:400;"> Tokenizing numbers digit by digit can indeed improve the model's understanding of numeric values. However, it might also make the model less adept at recognizing larger numerical patterns or relationships between numbers.</span></p></li><li style="font-size:11pt;font-weight:400;"><p style="margin-bottom:10pt;"><span style="font-size:11pt;font-weight:700;">Handling of Unicode Characters:</span><span style="font-size:11pt;font-weight:400;"> While the ability to natively tokenize Unicode characters is a strength, there's a potential for overfitting or misinterpretation. Emojis and other Unicode characters can have different meanings in different contexts or cultures. Relying heavily on them might lead to misunderstandings in generated narratives.</span></p></li></ol></div></div><p></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 11:43:18 +1000</pubDate></item><item><title><![CDATA[Testing AI's Ability to Understand Language in Context]]></title><link>https://www.nownextlater.ai/Insights/post/testing-ai-s-ability-to-understand-language-in-context</link><description><![CDATA[Researchers have developed a benchmark called the LAMBADA dataset to rigorously test how well AI models can leverage broader discourse context when predicting an upcoming word.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_xF4t6QesR8uxc92FhVi5Gw" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_8bhpItgQSkqzqqtLxxL5eQ" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_yTn8c-kASd-8tBJB9x60Aw" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"> [data-element-id="elm_yTn8c-kASd-8tBJB9x60Aw"].zpelem-col{ border-radius:1px; } </style><div data-element-id="elm_Mrls-pd6TVySli4Sre_gpQ" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_Mrls-pd6TVySli4Sre_gpQ"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Artificial intelligence has made great strides in natural language processing in recent years. Systems can now translate text, answer questions, and generate coherent paragraphs on demand. However, most AI still struggles with true language understanding that requires integrating information across long texts.</p><p><br></p><p><span style="color:inherit;">Back in 2016, </span>to address this limitation, researchers developed a benchmark called the LAMBADA dataset to rigorously test how well AI models can leverage broader discourse context when predicting an upcoming word.</p><p><br></p><p>LAMBADA contains over 10,000 passages extracted from fiction books, with the last word blanked out in each passage. When humans are given the full passage as context, they can easily guess the missing word. However, if humans only see the final sentence containing the blank, it becomes virtually impossible to predict the missing word.</p><p><br></p><p>For example, the sentence &quot;Do you honestly think that I would want you to have a ?&quot; on its own has many plausible words that could fill in the blank. But when given the full passage about a couple discussing pregnancy concerns beforehand, it becomes clear from the context that the missing word is &quot;miscarriage.&quot;</p><p><br></p><p>The researchers tested a wide range of AI systems on LAMBADA, including statistical n-gram models as well as advanced neural network architectures like LSTMs. Back then, all the models performed extremely poorly, with 0% to 7% accuracy in predicting the missing word. The models often relied on simple techniques like picking a random proper noun from the passage. Even methods designed to track broader context failed to match human performance. LAMBADA continues to be used today too test new projects such as <a href="https://blog.novelai.net/a-new-model-clio-is-coming-to-opus-ef4e2457c601" title="Novel AI" rel="">Novel AI</a>, and this time Models are performing with over 70% accuracy.<br></p><p></p><p><br></p><p>Truly intelligent systems will need to integrate information across long passages and reason about that context to understand language the way people do.</p><p><br></p><p>While AI chatbots and virtual assistants are improving customer service and other applications, they cannot yet achieve the sophistication of human context processing. Benchmarks like LAMBADA push innovators to develop the next generation of AI that skillfully uses context instead of relying on surface-level statistical patterns.</p><p><br></p><p>Just as IQ tests expanded to gauge different types of intelligence beyond a single number, benchmarks like LAMBADA are important for building well-rounded language AI systems. Advancing contextual language understanding will enable more fluent, trustworthy interfaces between people and machines. Whether in customer service or product development, AI that masters using context could unlock new levels of human-computer interaction.</p><p><br></p><p>Sources:</p><p><span style="font-family:&quot;Questrial&quot;, sans-serif;font-size:16px;"><a href="https://www.researchgate.net/publication/306093716_The_LAMBADA_dataset_Word_prediction_requiring_a_broad_discourse_context" title="The LAMBADA dataset: Word prediction requiring a broad discourse context" rel="">The LAMBADA dataset: Word prediction requiring a broad discourse context</a></span></p><p></p><p></p></div>
<p></p></div></div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:08:00 +1000</pubDate></item><item><title><![CDATA[Filling in the Blanks: AI Learns to Suggest Missing Pieces of Stories]]></title><link>https://www.nownextlater.ai/Insights/post/filling-in-the-blanks-ai-learns-to-suggest-missing-pieces-of-stories</link><description><![CDATA[AI research from 2019 explored how to automatically generate reasonable suggestions for missing sections of text.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_SY8i5KsBTYSBUTnW4F6Vog" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_owVqUUNJTL6eYSWPLHXg-Q" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_B4yMguWGR2e1JyIJ2453AA" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_smgWUcsE8U2E7JFjJXRSIg" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_smgWUcsE8U2E7JFjJXRSIg"] .zpimage-container figure img { width: 500px ; height: 407.45px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_smgWUcsE8U2E7JFjJXRSIg"] .zpimage-container figure img { width:500px ; height:407.45px ; } } @media (max-width: 767px) { [data-element-id="elm_smgWUcsE8U2E7JFjJXRSIg"] .zpimage-container figure img { width:500px ; height:407.45px ; } } [data-element-id="elm_smgWUcsE8U2E7JFjJXRSIg"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-09%20at%2011.51.47%20pm.png" width="500" height="407.45" loading="lazy" size="medium" alt="In the one stage baseline, the missing span is predicted given the context and the target length." data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_moUMIMAonYIEMJV_e5K92g" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_moUMIMAonYIEMJV_e5K92g"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><div style="color:inherit;"><p>Stories unfold step-by-step, but writers sometimes get stuck on how to connect one part to the next. AI research from 2019 explored how to automatically generate reasonable suggestions for missing sections of text. This &quot;story infilling&quot; aimed to assist creative writing by proposing ideas that align with the existing story while still surprising the author.</p><p><br></p><p>The researchers found that standard AI language models at the time struggled to balance coherence with novelty when filling in gaps. The generated text ended up too boring or too random. To address this limitation, they designed a two-step hierarchical system:</p><ul><li>First, the AI randomly selected a few rare, interesting words that could plausibly fit into the storyline based on the context. For a medieval fantasy passage, it might suggest words like &quot;dragon,&quot; &quot;princess,&quot; or &quot;castle.&quot; The system focused on rare words since they provide more information to guide the rest of the text.</li><li>Second, the system generated full sentences conditioned on those interesting words, searching likely combinations that form coherent text. Leveraging the rare words prevented repetitive suggestions, while allowing the model to focus on fluency and coherence.</li></ul><p><br></p><p>The researchers tested story infilling on passages from children's tales with missing sections of 15-30 words. Human evaluators preferred the hierarchical model's suggestions over non-hierarchical methods, which sacrificed diversity or quality.</p></div></div></div>
</div><div data-element-id="elm_smtSU1vMQJ-NiRm8qmX3eQ" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_smtSU1vMQJ-NiRm8qmX3eQ"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;">While an early attempt, the study shows promise for AI-assisted writing tools. The approach mirrors a writer's workflow - first deciding on key ideas, then piecing together suitable wording. Similar techniques may enable more human-like narrative understanding and creativity.<p><br></p><p>The field has greatly advanced since 2019 with models like Claude and GPT-4. Yet even powerful AI still struggles with high-level plot and character consistency. Explicitly decomposing generation into steps of planning and drafting, as humans do, is one way to address these challenges. While AI cannot replace human creativity, structured models could soon provide useful brainstorming and revision tools for real authors.</p><p><br></p><p>Sources:</p><p><span style="color:inherit;"><a href="https://www.seas.upenn.edu/%7Eccb/publications/story-infilling.pdf" title="Unsupervised Hierarchical Story Infilling" rel="">Unsupervised Hierarchical Story Infilling</a></span></p><p></p><br></div></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:07:37 +1000</pubDate></item><item><title><![CDATA[Behind the Scenes of Storytelling: Using AI to Plan and Structure Narratives]]></title><link>https://www.nownextlater.ai/Insights/post/behind-the-scenes-of-storytelling-using-ai-to-plan-and-structure-narratives</link><description><![CDATA[In 2019, researchers explored how artificial intelligence could use hierarchical models to improve computer-generated stories.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_zGmGX_ExR0OL3dEJ-qE92A" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_Sgf28QJsRkyTptjDigI1Zg" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_2nW8rLDxSvSle-fxQu_DRw" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"> [data-element-id="elm_2nW8rLDxSvSle-fxQu_DRw"].zpelem-col{ border-radius:1px; } </style><div data-element-id="elm_CrdU5M3JEZRY4kdPjaU4rQ" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_CrdU5M3JEZRY4kdPjaU4rQ"] .zpimage-container figure img { width: 500px ; height: 662.94px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_CrdU5M3JEZRY4kdPjaU4rQ"] .zpimage-container figure img { width:500px ; height:662.94px ; } } @media (max-width: 767px) { [data-element-id="elm_CrdU5M3JEZRY4kdPjaU4rQ"] .zpimage-container figure img { width:500px ; height:662.94px ; } } [data-element-id="elm_CrdU5M3JEZRY4kdPjaU4rQ"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-tablet-align-center zpimage-mobile-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-09%20at%2011.31.59%20pm.png" width="500" height="662.94" loading="lazy" size="medium" alt="Generating entity references for different genres" data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_A7S0QEDFQbSp0Oya0DQ8KA" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_A7S0QEDFQbSp0Oya0DQ8KA"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><p></p><p></p><div style="color:inherit;"><p></p><p>Storytelling seems almost magical. Writers conjure up entire worlds from their imaginations. But even master storytellers rely on plans and outlines to craft complex, coherent narratives spanning hundreds of words. In 2019, researchers explored how artificial intelligence could similarly use hierarchical models to improve computer-generated stories.</p><p><br/></p><p>Up until then, most AI systems created stories simply word-by-word from left to right. While fine for short texts, this method struggled with long-term plot and character consistency. The researchers proposed &quot;coarse-to-fine&quot; techniques to first generate story outlines, then build surface-level details conditioned on the outline.</p><p><br/></p><p>Their approach involved three steps: modeling the sequence of actions using verbs and arguments, generating story sentences with placeholder entities like &quot;ent0&quot;, and finally rewriting the placeholders with specific references. This mirrored how human writers first sketch a plot's arc, then go back to flesh out settings and characters.</p><p><br/></p><p>By creating more structured drafts, the AI models improved event diversity and entity consistency compared to previous approaches. The placeholder entities also made it easier to track characters, replacing different mentions with the same token. The researchers found that human judges strongly preferred stories created with hierarchical planning versus direct generation.</p><p><br/></p><p>While an early attempt, this work showed the promise of mimicking writing strategies like outlining and revising. The field has advanced rapidly since 2019 as models like GPT-4 or Claude 2 now generate amazingly fluent text. But behind the scenes, AI still struggles with plot and people - areas where hierarchical techniques could help. The research highlights the value of breaking narration into more human-like steps. A technique currently being explored by several&nbsp;<span style="color:inherit;">AI-assisted writing</span> startups such as <a href="https://novelai.net/" title="Novel AI" rel="">Novel AI</a> and <a href="https://www.sudowrite.com/" title="Sudowrite" rel="">Sudowrite</a>.<br/></p><p></p><p></p><p><br/></p><p>Just as outlines aid human storytellers, explicit planning and revisions may allow AI to better learn from experience. More structured generation spaces let models focus on specific challenges like action sequences before full text. While AI has seen stunning progress, people remain the masters of storycraft. Studying the narrative strategies of writers may guide systems to become more helpful to writers.<br/></p><p><br/></p><p>Source:</p><p><span style="color:inherit;"><a href="https://arxiv.org/pdf/1902.01109.pdf" title="Strategies for Structuring Story Generation" rel="">Strategies for Structuring Story Generation</a></span></p><p></p></div></div>
</div><div data-element-id="elm_FPPf8SQKIdBhofTA6erWbQ" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_FPPf8SQKIdBhofTA6erWbQ"] .zpimage-container figure img { width: 1090px ; height: 773.22px ; } } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-tablet-align-center zpimage-mobile-align-center zpimage-size-fit zpimage-tablet-fallback-fit zpimage-mobile-fallback-fit "><figure role="none" class="zpimage-data-ref"><a class="zpimage-anchor" href="https://www.reel-intelligence.org/" target="" title="Reel intelligence" rel=""><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Winner%20TV%20Pilot%20Screenplay%20-2-.png" size="fit" alt="Reel Intelligence"/></picture></a></figure></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:07:19 +1000</pubDate></item><item><title><![CDATA[Reading Between the Lines: Using Math to Uncover Hidden Patterns in Books]]></title><link>https://www.nownextlater.ai/Insights/post/reading-between-the-lines-using-math-to-uncover-hidden-patterns-in-books</link><description><![CDATA[Books may seem like straightforward stories, but researchers are finding fascinating mathematical patterns hidden in the text.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_BN-GUz9yTH6qOGKp8XYI2Q" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_6P1Nxs5lTRaNx8x9Z_sLXA" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_XhJZ4QBFRB6SJ6gQWdY-bQ" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_wDEEkImbYnmEBHGpW6Ludw" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_wDEEkImbYnmEBHGpW6Ludw"] .zpimage-container figure img { width: 500px ; height: 525.08px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_wDEEkImbYnmEBHGpW6Ludw"] .zpimage-container figure img { width:500px ; height:525.08px ; } } @media (max-width: 767px) { [data-element-id="elm_wDEEkImbYnmEBHGpW6Ludw"] .zpimage-container figure img { width:500px ; height:525.08px ; } } [data-element-id="elm_wDEEkImbYnmEBHGpW6Ludw"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-09%20at%2010.18.41%20pm.png" width="500" height="525.08" loading="lazy" size="medium" alt="An ‘ousiogram’ (Dodds et al., 2021) displaying power and danger scores for a subset of 14,499 unique words appearing in Terry Pratchett’s 41-book Discworld series." data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_1minUEv_RVGhhrISEzPBbA" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_1minUEv_RVGhhrISEzPBbA"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Books may seem like straightforward stories, but researchers are finding mathematical patterns hidden in the text. By tracking how words are used over the course of a book in minute detail, they can reveal new insights into plot, emotion, and structure that are not visible to the naked eye.</p><p><br></p><p>The researchers started by scoring a large number of words based on their emotional meaning. For example, positive words like &quot;love&quot; scored higher while negative words like &quot;war&quot; scored lower. They used a framework called &quot;ousiometrics&quot; which boils down emotions to two key dimensions: power and danger. Power relates to agency, confidence, and positivity. Danger relates to emotional uncertainty, negativity, and aggression.</p><p><br></p><p>They then took thousands of books and broke them down into short segments of 50 words each. For each segment, they calculated the average power and danger scores based on the words present. This turned each book into a rolling wave of numbers, with peaks representing more emotional sections and valleys as more neutral parts.</p><p><br></p><p>Short books generally showed a steady wave pattern while long books had more fluctuations in emotion over the course of the text. Surprisingly, when they zoomed in on long books they found the fluctuating highs and lows had a consistent length of a few thousand words. This matches the typical length of chapters in published fiction.</p><p><br></p><p>To study the patterns further, the researchers used a technique called empirical mode decomposition that breaks down fluctuations in data into distinct components, much like musical notes make up chords. The text segments were also compared to &quot;shuffled&quot; versions of the books with random word order. The real books differed from the random versions after a certain decomposition level, indicating that the fluctuations were not random but reflected an underlying structure.</p><p><br></p><p>These findings suggest longer books have a wave-like shape that is closer to collections of short stories or chapters. The emotional ups and downs of the text cycle on a scale of thousands of words, perhaps reflecting how long the human brain can comfortably process a complex narrative before needing a reset. Shorter books lacked these larger fluctuations.</p><p><br></p><p>While we intuitively understand how passages evoke certain moods, the researchers were able to quantify the pacing of emotional highs and lows mathematically. Their work helps confirm the existence of nested patterns in writing - punctuation gives phrases, paragraphs offer local structure, chapters provide mid-level segments, and over the full book arcs emerge.</p><p><br></p><p>So the next time you open a book, think about the hidden rhythms inside that subtly influence your experience. The feelings evoked in the story may follow mathematical waves as you steadily progress from cover to cover. This emerging field opens up new ways of appreciating the art and science of expert storytelling.</p><p><br></p><p>Sources:</p><p><a href="https://www.nature.com/articles/s41599-023-01680-4" title="A decomposition of book structure through ousiometric fluctuations in cumulative word-time" rel="">A decomposition of book structure through ousiometric fluctuations in cumulative word-time</a></p><p></p></div><p></p></div>
</div><div data-element-id="elm_BZslHjh1L1NAYTd780h3LQ" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_BZslHjh1L1NAYTd780h3LQ"] .zpimage-container figure img { width: 800px ; height: 344.00px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_BZslHjh1L1NAYTd780h3LQ"] .zpimage-container figure img { width:500px ; height:215.00px ; } } @media (max-width: 767px) { [data-element-id="elm_BZslHjh1L1NAYTd780h3LQ"] .zpimage-container figure img { width:500px ; height:215.00px ; } } [data-element-id="elm_BZslHjh1L1NAYTd780h3LQ"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-large zpimage-tablet-fallback-large zpimage-mobile-fallback-large "><figure role="none" class="zpimage-data-ref"><a class="zpimage-anchor" href="/aibooks" target="" rel=""><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Untitled%20design%20-4-.png" width="500" height="215.00" loading="lazy" size="large"/></picture></a></figure></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:06:53 +1000</pubDate></item><item><title><![CDATA[Storywrangler: Tracking Culture and Events through Twitter's Lens]]></title><link>https://www.nownextlater.ai/Insights/post/storywrangler-tracking-culture-and-events-through-twitter-s-lens</link><description><![CDATA[Researchers developed a tool called Storywrangler that leveraged Twitter data to create an "instrument for understanding our world through the lens of social media."]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_HszOwGyHS3u2ZbcwO98WUQ" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_NUndbqp2QH-raYQCD0h2-A" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_TG5dekBlReurU5nBQ7MK4A" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_HdtaSh_XqLf8A0UJCjxcLg" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_HdtaSh_XqLf8A0UJCjxcLg"] .zpimage-container figure img { width: 500px ; height: 386.22px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_HdtaSh_XqLf8A0UJCjxcLg"] .zpimage-container figure img { width:500px ; height:386.22px ; } } @media (max-width: 767px) { [data-element-id="elm_HdtaSh_XqLf8A0UJCjxcLg"] .zpimage-container figure img { width:500px ; height:386.22px ; } } [data-element-id="elm_HdtaSh_XqLf8A0UJCjxcLg"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-09%20at%2010.00.19%20pm.png" width="500" height="386.22" loading="lazy" size="medium" alt="Screenshot of the Storywrangler site showing example Twitter n-gram time series for the first half of 2020." data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_ooBK-RQ6Sx-jwMEa5QUoDQ" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_ooBK-RQ6Sx-jwMEa5QUoDQ"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><p class="whitespace-pre-wrap">Social media platforms like Twitter offered an unprecedented window into the real-time thoughts, conversations, and interests of millions of people. Researchers developed a tool called Storywrangler that leveraged Twitter data to create an &quot;instrument for understanding our world through the lens of social media.&quot;</p><p class="whitespace-pre-wrap"><br></p><p class="whitespace-pre-wrap">Storywrangler analyzed over 100 billion tweets dating back to 2008 to detect trends in word usage over time. It broke down tweets into &quot;n-grams&quot; - sequences of one, two, or three words - and tracked how the usage frequencies of these n-grams changed on a daily basis across different languages.</p><p class="whitespace-pre-wrap">This massive database allowed researchers to see how real-world events, from natural disasters to political movements, were reflected in the narratives that unfolded on Twitter. For example, Storywrangler revealed surging interest in climate-related terms during major storms and wildfires. And it captured the rapid rise and fall of hashtags associated with social justice protests. Beyond reacting to news, Twitter also mirrored more subtle cultural shifts, like the waxing and waning popularity of celebrities or diets.</p><p class="whitespace-pre-wrap"><br></p><p class="whitespace-pre-wrap">Storywrangler went beyond tracking raw frequencies - it also quantified how widely information spread on social media through shares and reposts. This helped distinguish niche conversations from truly viral ideas. The researchers used &quot;contagiograms&quot; to visualize both the popularity and amplification of n-grams over time.</p><p class="whitespace-pre-wrap"><br></p><p class="whitespace-pre-wrap">There were certainly limitations to the Twitter lens. The platform's user base skewed young, urban, and affluent compared to the general population. Bots and organized campaigns could artificially inflate interest in certain topics. And the meanings of words themselves evolved across the years.</p><p class="whitespace-pre-wrap"><br></p><p class="whitespace-pre-wrap">But used carefully, Storywrangler offered an unparalleled window into the collective consciousness - recording not just major news events but also the mundane daily conversations of millions worldwide. It aimed to complement more traditional data sources like books and news archives. The researchers hoped Storywrangler would enable more data-driven computational social science to understand our fast-changing, digitally-connected world.</p><p class="whitespace-pre-wrap"><br></p><p class="whitespace-pre-wrap">Source:</p><p class="whitespace-pre-wrap"><span style="color:inherit;"><a href="https://arxiv.org/pdf/2007.12988.pdf" title="Storywrangler: A massive exploratorium for sociolinguistic, cultural, socioeconomic, and political timelines using Twitter" rel="">Storywrangler: A massive exploratorium for sociolinguistic, cultural, socioeconomic, and political timelines using Twitter</a></span></p><p class="whitespace-pre-wrap"></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:05:59 +1000</pubDate></item><item><title><![CDATA[Teaching AI to Tell Better Tales by Integrating External Knowledge]]></title><link>https://www.nownextlater.ai/Insights/post/Teaching-AI-to-Tell-Better-Tales-by-Integrating-External-Knowledge</link><description><![CDATA[New research explores how integrating structured knowledge into AI systems can enhance storytelling abilities.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_-2LoK4FLSj6kE0maiRXg6A" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_2uZedO7XTwmjrBCrAZQSKw" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_Tz4iZq4gSd6jgmc03jak_g" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_2YA5oGBzkoYbOgntrUGhXA" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_2YA5oGBzkoYbOgntrUGhXA"] .zpimage-container figure img { width: 500px ; height: 389.03px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_2YA5oGBzkoYbOgntrUGhXA"] .zpimage-container figure img { width:500px ; height:389.03px ; } } @media (max-width: 767px) { [data-element-id="elm_2YA5oGBzkoYbOgntrUGhXA"] .zpimage-container figure img { width:500px ; height:389.03px ; } } [data-element-id="elm_2YA5oGBzkoYbOgntrUGhXA"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-09%20at%209.41.19%20pm.png" width="500" height="389.03" loading="lazy" size="medium" alt="Three layers of narratological concepts about story: fabula, plot, and discourse." data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_kjuSNjtuTimsXg6Qn73jCw" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_kjuSNjtuTimsXg6Qn73jCw"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Storytelling comes naturally to humans. But for machines, spinning an engaging narrative remains an elusive goal. While AI can generate remarkably fluent text, its tales often lack coherence or get repetitive. New research explores how integrating structured knowledge into AI systems can enhance storytelling abilities.</p><p><br></p><p>When reading a story, we draw on general knowledge about how events logically unfold and characters plausibly act. We track complex plot threads and fill gaps using common sense. Machines lack this innate understanding we take for granted. Their stories can become nonsensical or contradictory.</p><p>To tackle this, researchers are providing AI systems explicit knowledge in structured formats. This external knowledge acts like a guide, keeping machine-generated plots on track. It also helps avoid stale repetitions by expanding the ideas available to pull from.</p><p><br></p><p>Several common limitations plague today's AI storytellers:</p><ul><li>Lack of long-term coherence. Without a sense of overall narrative arc, they ramble aimlessly.</li><li>Insufficient grounding in real-world facts. Stories come off vague rather than richly descriptive.</li><li>Repetition. They loop the same words and phrases like a broken record.</li><li>Hallucination. They fabricate events that don't logically follow.</li></ul><p><br></p><p>Integrating knowledge resources like ConceptNet, which contains common sense facts about the world, alleviates these issues. The knowledge functions like an annotated outline, steering the plot. It also provides a memory bank of concepts to reference, varying the content.</p><p><br></p><p>But effectively harnessing external knowledge remains challenging. Two main strategies have emerged:</p><ol><li>Injecting knowledge directly into the AI system's training process, like teaching a human author.</li><li>Using knowledge as an external guiding reference during story generation.</li></ol><p><br></p><p>Each approach has trade-offs. Weighting structured resources too strongly can pollute the system's original language skills. But using knowledge merely as a loose guide can fail to correct nonsensical narration.</p><p><br></p><p>Striking the right balance is an active research problem. Scientists are also expanding the knowledge available to AI storytellers with new databases. Most systems today use generic common sense facts. But resources detailing specific people, places, and events could enable more detailed, vivid storytelling.</p><p><br></p><p>Automating evaluation also poses difficulties. No single &quot;correct&quot; story exists for a given prompt. Automatic metrics struggle to account for creativity and interest - aspects requiring human judgment. More robust evaluation is critical to gauge progress.</p><p><br></p><p>Despite hurdles, knowledge-infused narration clearly improves coherence, factual grounding, and variation. AI authors with a knowledge boost spin far more convincing yarns. The research provides a roadmap for machines to better mimic core elements of human storytelling.</p><p><br></p><p>Rather than viewing imagination and structure as at odds, they are complementary. Master storytellers combine free-flowing creativity with purposeful intent. By fusing extensive knowledge with unrestrained generation, machines inch closer toward unlocking that balancing act.</p><p><br></p><p>Sources:</p><p><span style="color:inherit;"><a href="https://arxiv.org/pdf/2212.04634.pdf" title="Open-world Story Generation with Structured Knowledge Enhancement: A Comprehensive Survey" rel="">Open-world Story Generation with Structured Knowledge Enhancement: A Comprehensive Survey</a></span></p><p></p></div><p></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:05:08 +1000</pubDate></item><item><title><![CDATA[Teaching AI to Craft Coherent Stories]]></title><link>https://www.nownextlater.ai/Insights/post/Teaching-AI-to-Craft-Coherent-Stories</link><description><![CDATA[New research from Stanford University demonstrates how "emotion maps" could improve story generation.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_5pdLFLcdSimmqoTYE0K6Kg" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_A0wHFCauRW2Gv_TkglOtkg" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_9ir451-oSWK-HnmcTFIUCQ" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_mQp_YFlv2tErqnpf5Ur11A" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_mQp_YFlv2tErqnpf5Ur11A"] .zpimage-container figure img { width: 800px ; height: 418.90px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_mQp_YFlv2tErqnpf5Ur11A"] .zpimage-container figure img { width:500px ; height:261.81px ; } } @media (max-width: 767px) { [data-element-id="elm_mQp_YFlv2tErqnpf5Ur11A"] .zpimage-container figure img { width:500px ; height:261.81px ; } } [data-element-id="elm_mQp_YFlv2tErqnpf5Ur11A"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-tablet-align-center zpimage-mobile-align-center zpimage-size-large zpimage-tablet-fallback-large zpimage-mobile-fallback-large hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-09%20at%209.15.49%20pm.png" width="500" height="261.81" loading="lazy" size="large" alt="Average emotion map characteristics for stories tagged &quot;Depression&quot; in the new dataset. Note the strong prevalence of negative emotion." data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_VYcBjLcoQYWexCsYimf8yQ" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_VYcBjLcoQYWexCsYimf8yQ"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Storytelling comes naturally to humans, but is exceptionally difficult for artificial intelligence. Machine learning models that generate remarkably fluent text still struggle to craft narrative arcs spanning paragraphs or pages. New research from Stanford University demonstrates how &quot;emotion maps&quot; could improve story generation.</p><p><br/></p><p>The key challenge is imbuing AI systems with a high-level understanding of plot and long-range dependencies - core elements of compelling stories. Without such top-down guidance, machine-written tales easily become repetitive and disjointed.</p><p><br/></p><p>The Stanford project explores a technique called hierarchical generation. This involves first creating a short premise or prompt, then expanding that outline into a full story. The premise acts like an anchor, guiding the system to remain on topic and logically progress the narrative.</p><p><br/></p><p>But how can we represent a good premise for AI? The researchers move beyond using text, instead generating &quot;emotion maps.&quot; These maps contain a series of numerical scores representing different emotive attributes. Each score tracks how positive or negative, sad or joyful consecutive sections of the story feel.</p><p><br/></p><p>For instance, a map may start very positive, then become sadder, and end on a more hopeful note. Feeding these maps as prompts produces stories that logically follow the intended emotional arc. The numbers offer a bird's-eye view of the narrative's affective flow.</p><p><br/></p><p>Remarkably, this numerically-conditioned approach achieved comparable results on two standard story datasets as previous efforts using text prompts. The generated stories displayed coherent grammar and punctuation, sensibly reacting to the emotion map's ups and downs.</p><p><br/></p><p>To better understand the relationship between maps and stories, the researchers introduced a new metric called Average Emotional Similarity. This quantifies how closely a story's actual emotion aligns with its prompt map. Initial results demonstrate some correlation, confirming the maps exert influence on the tone of generated text.</p><p><br/></p><p>There are several advantages to conditioning story generation on simplified cues rather than verbose outlines. Maps neatly capture narrative essence in a compact, rapidly computed form. Reducing hand-authoring effort also enables building larger training datasets.</p><p><br/></p><p>However, many challenges remain. Emotionless stories often flummox the system, producing bizarre outputs. Repetition and hallucination still crop up, demonstrating the need for greater plot coherency. And evaluating story quality continues to prove difficult without extensive human judgement.</p><p><br/></p><p>Nonetheless, this research highlights the potential of hierarchical methods to imbue AI storytelling with greater purpose. The raw material exists in today's pretrained language models - machines that have &quot;read&quot; vast amounts of text. We must guide them toward higher reasoning about concepts like theme, characters, and dramatic structure.</p><p><br/></p><p>Interactive tools could empower human authors to easily craft emotion maps, generating stories tailored to their creative vision. Teachers might build maps to help students practice writing logically paced narratives. Therapists could use emotive cadences to gently evoke memories or feelings from patients.</p><p><br/></p><p>The capacity for machines to conjure compelling tales could transform how we communicate ideas and experiences. But achieving this dream will hinge on passing down our innate sense for what makes a story worth telling. With innovations like emotion maps lighting the way, artificial authors inch toward unlocking our imagination.</p><p><br/></p><p>source:</p><div style="color:inherit;"><div><div><div><p><a href="https://www.google.com/url?sa=t&amp;rct=j&amp;q=&amp;esrc=s&amp;source=web&amp;cd=&amp;ved=2ahUKEwjX-M24ss-AAxVETGwGHQoICNcQFnoECBkQAQ&amp;url=https%3A%2F%2Fweb.stanford.edu%2Fclass%2Farchive%2Fcs%2Fcs224n%2Fcs224n.1214%2Freports%2Ffinal_reports%2Freport104.pdf&amp;usg=AOvVaw2Iwkk04_oaGHGAXlr-uqtB&amp;opi=89978449" title="Hierarchical, Feature-Based Text Generation " rel="">Hierarchical, </a><a href="https://www.google.com/url?sa=t&amp;rct=j&amp;q=&amp;esrc=s&amp;source=web&amp;cd=&amp;ved=2ahUKEwjX-M24ss-AAxVETGwGHQoICNcQFnoECBkQAQ&amp;url=https%3A%2F%2Fweb.stanford.edu%2Fclass%2Farchive%2Fcs%2Fcs224n%2Fcs224n.1214%2Freports%2Ffinal_reports%2Freport104.pdf&amp;usg=AOvVaw2Iwkk04_oaGHGAXlr-uqtB&amp;opi=89978449" title="Hierarchical, Feature-Based Text Generation " rel="">Feature-Based </a><a href="https://www.google.com/url?sa=t&amp;rct=j&amp;q=&amp;esrc=s&amp;source=web&amp;cd=&amp;ved=2ahUKEwjX-M24ss-AAxVETGwGHQoICNcQFnoECBkQAQ&amp;url=https%3A%2F%2Fweb.stanford.edu%2Fclass%2Farchive%2Fcs%2Fcs224n%2Fcs224n.1214%2Freports%2Ffinal_reports%2Freport104.pdf&amp;usg=AOvVaw2Iwkk04_oaGHGAXlr-uqtB&amp;opi=89978449" title="Hierarchical, Feature-Based Text Generation " rel="">Text </a><a href="https://www.google.com/url?sa=t&amp;rct=j&amp;q=&amp;esrc=s&amp;source=web&amp;cd=&amp;ved=2ahUKEwjX-M24ss-AAxVETGwGHQoICNcQFnoECBkQAQ&amp;url=https%3A%2F%2Fweb.stanford.edu%2Fclass%2Farchive%2Fcs%2Fcs224n%2Fcs224n.1214%2Freports%2Ffinal_reports%2Freport104.pdf&amp;usg=AOvVaw2Iwkk04_oaGHGAXlr-uqtB&amp;opi=89978449" title="Hierarchical, Feature-Based Text Generation " rel="">Generation </a><br/></p><p></p><div style="color:inherit;"><div>Caitlin Hogan Department of Computer Science Stanford University </div>
</div><p></p></div></div></div></div><p></p></div><p></p></div></div><div data-element-id="elm_atkWncPAs8-XRGwZmMJBuw" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_atkWncPAs8-XRGwZmMJBuw"] .zpimage-container figure img { width: 1090px ; height: 773.22px ; } } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-tablet-align-center zpimage-mobile-align-center zpimage-size-fit zpimage-tablet-fallback-fit zpimage-mobile-fallback-fit "><figure role="none" class="zpimage-data-ref"><a class="zpimage-anchor" href="https://www.reel-intelligence.org/" target="" title="Reel Intelligence" rel=""><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Winner%20TV%20Pilot%20Screenplay%20-2-.png" size="fit"/></picture></a></figure></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:04:45 +1000</pubDate></item><item><title><![CDATA[Teaching AI to Tell Coherent Stories]]></title><link>https://www.nownextlater.ai/Insights/post/teaching-ai-to-tell-coherent-stories</link><description><![CDATA[Back in 2018, researchers from Facebook AI developed a new method to improve story generation through hierarchical modeling. Their approach mimics how people plan out narratives.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_tQLbaWaBQj2087vOICMB-w" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_dTbUlqiJQ8arZPfnMA9eOw" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_0gLbapfvR8-xevlg0lLJjg" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_MU_X1Z_dT7W6u5b5FF-fxA" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_MU_X1Z_dT7W6u5b5FF-fxA"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Back in 2018, researchers from Facebook AI developed a new method to improve story generation through hierarchical modeling. Their approach mimics how people plan out narratives. While significant developments have occurred in language generation, it is worth exploring this technique as several current projects leverage these techniques.<br></p><p><br></p><p>The key innovation is generating a short premise first, then expanding that premise into a full story. Take the premise &quot;A knight goes on a quest to save the kingdom.&quot; From this high-level summary, a system can flesh out details - the specific characters, events, and dialogue - while staying focused on the overarching plot.</p><p><br></p><p>This technique helps in two ways. First, the premise acts like an outline, guiding the story generation process. Second, conditioning the story on the premise makes it easier for the AI to stay on topic. Without such grounding, AI systems tend to lose coherence as they generate text word-by-word.</p><p>To train and test hierarchical story generation, the researchers built a new dataset using the r/WritingPrompts subreddit. This online community shares story premises, or prompts, that inspire other users to write original tales. Drawing on over 300,000 prompt-story pairs, the dataset captures diverse genres and narrative styles.</p><p><br></p><p>The researchers' AI system first generates a short prompt, similar to a human providing a premise. It then passes this prompt to a second model that expands it into a full story. Both steps use sequence-to-sequence neural networks, which translate an input sequence into target text.</p><p><br></p><p>To improve story coherence, the researchers introduced two key innovations. <br></p><div style="color:inherit;"><ul><li>First, they trained a standard sequence-to-sequence neural network model on the story generation task. This model learns to generate fluent stories, but often ignores the premise and fails to maintain consistency with it.</li><li>Next, they trained a second sequence-to-sequence model, but this time provided the hidden state outputs of the first pre-trained model to the second model during training. In other words, the second model learns on top of the representations already learned by the first model. It has access to the pre-trained model's outputs. This encourages the second model to focus specifically on relating the story back to the premise, rather than just improving language modeling in general</li></ul></div><p></p><p><span style="color:inherit;"><br></span></p><p><span style="color:inherit;">By &quot;fusing&quot; the second model with the first pre-trained model in this way, the researchers aim to improve coherence between the premise and final story. The second model builds on top of the first to better maintain relevance.</span></p><p><span style="color:inherit;"><br></span></p><p>Experiments found these advances substantially boosted performance. The AI's stories scored higher in human evaluations for coherence, relevance to the prompt, and overall quality compared to baseline systems. The gated self-attention enabled referring back to any previous part of the story. And model fusion encouraged tighter connections between the premise and story.</p><p><br></p><p>While far from perfect, these results demonstrate AI's increasing capacity for controllable, long-form text generation. The hierarchical approach mimics how people first conceptualize, then craft, narratives. Such human-inspired techniques will be key to teaching machines to tell truly compelling tales spanning paragraphs or pages.</p><p><br></p><p>The researchers highlight several directions for improvement. The premises generated by the AI tend to be generic, lacking the creativity of human prompts. Repetition remains an issue when expanding premises into stories. And problems like dropped pronouns persist.</p><p><br></p><p>Nonetheless, this work moves neural story writing systems in a promising direction. As models strengthen their understanding of narrative cause-and-effect, characters, and more, their power as digital storytellers will grow. Hierarchical modeling that mirrors human planning seems a fitting way to imbue AI with our innate gift for spinning both short yarns and epics.</p><p><br></p><p>Sources:</p><p><a href="https://arxiv.org/pdf/1805.04833.pdf" title="arxiv" rel="">arxiv</a><br></p><p></p></div><p></p></div>
</div><div data-element-id="elm_JC64W1WgifP5ZdRShAK_UQ" data-element-type="codeSnippet" class="zpelement zpelem-codesnippet "><div class="zpsnippet-container"><iframe src="https://player.vimeo.com/video/285801163?h=a817f4f945&byline=0&portrait=0" width="640" height="360" frameborder="0" allow="autoplay; fullscreen; picture-in-picture" allowfullscreen></iframe><p><a href="https://vimeo.com/285801163">Hierarchical Neural Story Generation</a> from <a href="https://vimeo.com/aclweb">ACL</a> on <a href="https://vimeo.com">Vimeo</a>.</p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:03:41 +1000</pubDate></item><item><title><![CDATA[Peeking Inside the Black Box: Uncovering What AI Models Know About Books]]></title><link>https://www.nownextlater.ai/Insights/post/peeking-inside-the-black-box-uncovering-what-ai-models-know-about-books</link><description><![CDATA[New research from the University of California, Berkeley sheds light on one slice of these models' knowledge: which books they have "read" and memorized. The study uncovers systematic biases in what texts AI systems know most about.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_gs7G1b-XTpGiZydeBC0a0Q" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_DOD6ivG7QmSpfUB8fVxIYQ" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_LiwzJORqQmSQI3gbPKBwuA" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_K_koNQXna4P8T3ZlwBatnw" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_K_koNQXna4P8T3ZlwBatnw"] .zpimage-container figure img { width: 1090px ; height: 671.48px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_K_koNQXna4P8T3ZlwBatnw"] .zpimage-container figure img { width:723px ; height:445.39px ; } } @media (max-width: 767px) { [data-element-id="elm_K_koNQXna4P8T3ZlwBatnw"] .zpimage-container figure img { width:415px ; height:255.65px ; } } [data-element-id="elm_K_koNQXna4P8T3ZlwBatnw"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-fit zpimage-tablet-fallback-fit zpimage-mobile-fallback-fit hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-09%20at%206.48.43%20pm.png" width="415" height="255.65" loading="lazy" size="fit" alt="Top 20 books by GPT-4 name cloze accuracy" data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_a3xA-CbcQJepcHHoPuycYQ" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_a3xA-CbcQJepcHHoPuycYQ"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Artificial intelligence systems like ChatGPT and GPT-4 have demonstrated impressive language skills, holding fluent conversations and answering questions on virtually any topic. But their inner workings remain largely opaque to users. These systems are &quot;black boxes&quot; - we know little about what knowledge they actually contain.</p><p><br></p><p>New research from the University of California, Berkeley sheds light on one slice of these models' knowledge: which books they have &quot;read&quot; and memorized. The study uncovers systematic biases in what texts AI systems know most about, with implications for how we should evaluate them.</p><p>The researchers focused specifically on works of fiction. They selected a sample of 571 English novels published between 1749 and 2020, containing literary classics along with contemporary bestsellers and award winners. The sample spanned mystery, romance, and science fiction genres as well as global Anglophone and African American literature.</p><p><br></p><p>For each book, the team extracted short passages of 40-60 words containing a single character name - but with the name removed. For instance, a passage from Pride and Prejudice might read &quot;______ entered the room and greeted her hosts warmly.&quot; Humans cannot guess the missing name from such brief context. But does the AI system know the name from having read the full book?</p><p><br></p><p>The researchers tested two systems, ChatGPT and GPT-4, by giving each passage and asking what single-word name belongs in the blank. The accuracy of each AI model on this challenging &quot;cloze&quot; task revealed what books it likely memorized.</p><p><br></p><p>The results illuminated clear biases. Both systems strongly favor science fiction and fantasy works like Lord of the Rings and Harry Potter over other genres. They excel at classic literature like Alice in Wonderland and Pride and Prejudice but fare poorly on modern award-winning diverse books. In short, they are more knowledgeable about popular texts.</p><p><br></p><p>What explains this imbalance? The researchers found it closely mirrors what's most duplicated across the internet. There is a strong correlation between AI accuracy on a book and the number of verbatim passages found through Google, Bing, and other sources. The models appear to &quot;know&quot; books in proportion to their web popularity.</p><p><br></p><p>This reliance on the internet has consequences. The study showed AI systems perform better at predicting a book's publication date and summarizing its passages when they have memorized the book. In other words, their reasoning is tied to memorization - causing disparities between popular versus niche texts.</p><p><br></p><p>These insights matter because AI systems like ChatGPT are increasingly used for applications like analyzing literature and human culture. If their knowledge comes largely from duplicated web text, focused on popular sci-fi and fantasy, how well can we trust their judgments about less mainstream books? Their skewed knowledge could propagate biases into downstream decisions.</p><p><br></p><p>The findings illustrate the challenges of opaque &quot;black box&quot; AI systems whose training data is secret. OpenAI, which created ChatGPT and GPT-4, has not revealed what texts were used to train them. This leaves us unable to fully assess their knowledge gaps.</p><p><br></p><p>The researchers argue we should instead push for more transparent, open-source AI systems whose training data is public knowledge. This allows us to better understand their strengths and weaknesses - illuminated through research like this study.</p><p><br></p><p>As AI models grow more capable and ubiquitous, it becomes only more important to peek inside their black boxes. Understanding what knowledge they contain helps ensure we build and apply them responsibly. Analyses of what systems like ChatGPT &quot;know&quot; about books mark an important step toward making AI more intelligible as it continues permeating our lives.</p><p><br></p><p>Sources:</p><p><a href="https://arxiv.org/pdf/2305.00118.pdf" title="arxiv" rel="">arxiv</a><br></p><p></p></div><p></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:03:22 +1000</pubDate></item></channel></rss>