<?xml version="1.0" encoding="UTF-8" ?><!-- generator=Zoho Sites --><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><atom:link href="https://www.nownextlater.ai/Insights/tag/chain-of-thought-prompting/feed" rel="self" type="application/rss+xml"/><title>Now Next Later AI - Blog #Chain-of-thought Prompting</title><description>Now Next Later AI - Blog #Chain-of-thought Prompting</description><link>https://www.nownextlater.ai/Insights/tag/chain-of-thought-prompting</link><lastBuildDate>Wed, 26 Nov 2025 21:35:47 +1100</lastBuildDate><generator>http://zoho.com/sites/</generator><item><title><![CDATA[Improving the Reasoning Capabilities of Large Language Models: Chain-of-Thought Prompting]]></title><link>https://www.nownextlater.ai/Insights/post/Improving-the-Reasoning-Capabilities-of-Large-Language-Models</link><description><![CDATA[We will explore recent research aimed at improving the reasoning capabilities of large language models and discuss its implications for business leaders.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_HeUKbAqXR0-jTwSZjpGN-g" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_B8wH2PGNQy66Ywn6a11JnA" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_acCeBoFwT-20cAUaSpmiKA" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_k_vSHOAOrs0r7vanN6uP-A" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_k_vSHOAOrs0r7vanN6uP-A"] .zpimage-container figure img { width: 1090px ; height: 470.74px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_k_vSHOAOrs0r7vanN6uP-A"] .zpimage-container figure img { width:723px ; height:312.25px ; } } @media (max-width: 767px) { [data-element-id="elm_k_vSHOAOrs0r7vanN6uP-A"] .zpimage-container figure img { width:415px ; height:179.23px ; } } [data-element-id="elm_k_vSHOAOrs0r7vanN6uP-A"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-fit zpimage-tablet-fallback-fit zpimage-mobile-fallback-fit hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-09%20at%2010.09.24%20am.png" width="415" height="179.23" loading="lazy" size="fit" alt="Chain-of-thoughtprompting example" data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_dgB2_25ZSEmibdbh9Ipskg" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_dgB2_25ZSEmibdbh9Ipskg"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-center " data-editor="true"><div style="color:inherit;text-align:left;"><span style="color:inherit;"><span><span style="font-size:16px;font-weight:400;text-indent:0px;">Artificial Intelligence has made significant strides in recent years, particularly in the field of Natural Language Processing (NLP). Large language models, such as GPT-4 or Claude 2 , have demonstrated impressive capabilities in generating human-like text. However, these models often struggle with tasks that require reasoning or understanding the context. We will explore recent research aimed at improving the reasoning capabilities of large language models and discuss its implications for business leaders.</span></span></span></div>
<div style="color:inherit;text-align:left;"><span style="color:inherit;"><span><span style="font-size:16px;font-weight:400;text-indent:0px;"><br></span></span></span><h2 style="font-weight:600;text-indent:0px;">The Challenge of Reasoning in AI</h2><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">One of the key challenges in AI is enabling machines to reason like humans. While large language models can generate text that sounds plausible, they often lack the ability to reason or understand the context. This can lead to outputs that are nonsensical or incorrect, despite sounding reasonable.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">For instance, if you ask a language model a complex question or a question that requires understanding of the real world beyond its training data, it may struggle to provide a correct answer. This is because these models are trained on large amounts of text data and generate responses based on patterns they've learned from this data, rather than truly understanding the content.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><h2 style="font-weight:600;text-indent:0px;">Chain-of-Thought Prompting</h2><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">A recent research paper titled &quot;Chain-of-Thought Prompting Elicits Reasoning in Large Language Models&quot; proposes a novel approach to address this issue. The authors introduce a method called &quot;Chain-of-Thought Prompting&quot; (CoTP), which structures the prompts in a way that encourages the model to build upon its previous responses, creating a chain of thought.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">The idea is to ask the model to explain its reasoning, justify its answers, and consider alternative viewpoints. This approach can help to mitigate some of the limitations of current language models, such as their tendency to generate plausible-sounding but incorrect or nonsensical answers.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">The researchers found that this method significantly improved the model's performance on tasks that require reasoning, such as answering complex questions or solving problems. The model was also better able to provide coherent and detailed explanations of its thought process.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><div style="color:inherit;"><p style="font-size:16px;font-weight:400;text-indent:0px;">Let's consider an example:</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;"><strong style="font-weight:600;">Question:</strong> &quot;If I have 10 apples and I give away 3, then I buy a dozen more and give away 5, how many apples do I have left?&quot;</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;"><strong style="font-weight:600;">Incorrect Language Model Response:</strong> &quot;You have 22 apples left.&quot;</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">The correct calculation should be: Start with 10 apples, give away 3 (leaving 7), buy a dozen more (adding 12, for a total of 19), then give away 5. So, you should have 14 apples left.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">Now, let's see how a chain-of-thought prompting approach might help address this:</p><ol><li><p><strong style="font-weight:600;">Prompt:</strong> &quot;You start with 10 apples and give away 3. How many apples do you have now?&quot; <strong style="font-weight:600;">Model Response:</strong> &quot;You have 7 apples.&quot;</p></li><li><p><strong style="font-weight:600;">Prompt:</strong> &quot;You buy a dozen more apples. How many apples do you have now?&quot; <strong style="font-weight:600;">Model Response:</strong> &quot;You have 19 apples.&quot;</p></li><li><p><strong style="font-weight:600;">Prompt:</strong> &quot;You give away 5 of your apples. How many apples do you have left?&quot; <strong style="font-weight:600;">Model Response:</strong> &quot;You have 14 apples left.</p></li></ol><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">In this chain-of-thought approach, each step of the calculation is broken down into a separate prompt, making it easier for the model to handle the complexity of the overall problem. This approach encourages the model to build on its previous responses, leading to more accurate and reasoned responses.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">While you can prompt step-by-step, just asking the model to reason step-by-step might produce the desired outcome. <br></p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p></div>
<p style="font-size:16px;font-weight:400;text-indent:0px;"></p><h2 style="font-weight:600;text-indent:0px;">Enhancing Temporal Reasoning</h2><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">Another research paper titled &quot;Towards Benchmarking and Improving the Temporal Reasoning Capability of Large Language Models&quot; focuses on improving the temporal reasoning capability of large language models. The authors introduce a comprehensive probing dataset to evaluate the temporal reasoning capability of these models. They also propose a novel learning framework to improve this capability, based on temporal span extraction and time-sensitive reinforcement learning.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;"><span style="color:inherit;"><span style="font-size:16px;font-weight:400;text-indent:0px;">It's worth noting that temporal reasoning is a complex task that involves understanding not just the order of events, but also their duration and the intervals between them. This makes it a challenging problem for AI, but also a crucial one, as many real-world tasks require the ability to reason about time.</span></span></p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">Temporal reasoning is crucial for many real-world applications. For example, understanding that an event occurred before or after another event can be critical for making business decisions or understanding historical trends.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><h2 style="font-weight:600;text-indent:0px;">Visual-Language Models and Reasoning</h2><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">In the paper &quot;Enhance Reasoning Ability of Visual-Language Models via Large Language Models&quot;, the authors propose a method called TReE, which transfers the reasoning ability of a large language model to a visual language model in zero-shot scenarios. This approach could be particularly useful for tasks that involve both visual and textual information, such as image captioning or visual question answering.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><h2 style="font-weight:600;text-indent:0px;">Implications for Business Leaders</h2><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">The advancements in improving the reasoning capabilities of large language models have significant implications for business leaders. These models can be used to automate various tasks, such as customer service, content generation, and data analysis. Improved reasoning capabilities can make these models more effective and reliable, leading to better business outcomes.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">Moreover, these models can provide valuable insights by analyzing large amounts of text data, such as customer reviews or social media posts. They can identify trends, sentiments, and key topics, helping business leaders to make informed decisions.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">However, it's important to remember that these models are not perfect. They still lack a true understanding of the world and can make mistakes. Therefore, their outputs should always be reviewed and verified by human experts.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">In conclusion, the research in improving the reasoning capabilities of large language models is a promising step towards making AI systems more intelligent and useful. As these models continue to improve, they will become an increasingly valuable tool for businesses.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><h2 style="font-weight:600;text-indent:0px;">References</h2><ol><li><a href="https://arxiv.org/abs/2201.11903" rel="">Chain-of-Thought Prompting Elicits Reasoning in Large Language Models</a></li><li><a href="https://arxiv.org/abs/2306.08952" rel="">Towards Benchmarking and Improving the Temporal Reasoning Capability of Large Language Models</a></li><li><a href="https://arxiv.org/abs/2305.13267">Enhance Reasoning Ability of Visual-Language Models via Large Language Models</a></li></ol></div>
<p></p></div></div><div data-element-id="elm_2Mp_dnTcMDIfy3sgSR-Ouw" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_2Mp_dnTcMDIfy3sgSR-Ouw"] .zpimage-container figure img { width: 800px ; height: 344.00px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_2Mp_dnTcMDIfy3sgSR-Ouw"] .zpimage-container figure img { width:500px ; height:215.00px ; } } @media (max-width: 767px) { [data-element-id="elm_2Mp_dnTcMDIfy3sgSR-Ouw"] .zpimage-container figure img { width:500px ; height:215.00px ; } } [data-element-id="elm_2Mp_dnTcMDIfy3sgSR-Ouw"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-large zpimage-tablet-fallback-large zpimage-mobile-fallback-large "><figure role="none" class="zpimage-data-ref"><a class="zpimage-anchor" href="/aibooks" target="" rel=""><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Untitled%20design%20-4-.png" width="500" height="215.00" loading="lazy" size="large"/></picture></a></figure></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Tue, 08 Aug 2023 12:16:17 +1000</pubDate></item></channel></rss>