<?xml version="1.0" encoding="UTF-8" ?><!-- generator=Zoho Sites --><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><atom:link href="https://www.nownextlater.ai/Insights/tag/rlhf/feed" rel="self" type="application/rss+xml"/><title>Now Next Later AI - Blog #RLHF</title><description>Now Next Later AI - Blog #RLHF</description><link>https://www.nownextlater.ai/Insights/tag/rlhf</link><lastBuildDate>Wed, 26 Nov 2025 21:23:29 +1100</lastBuildDate><generator>http://zoho.com/sites/</generator><item><title><![CDATA[Training AI with Human Feedback for Better Summaries: RLHF]]></title><link>https://www.nownextlater.ai/Insights/post/training-ai-with-human-feedback-for-better-summaries</link><description><![CDATA[The researchers found that optimizing the AI for direct human preferences significantly boosted performance compared to just training it to mimic reference summaries.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_jmOaxdE_SPuTL-nJNyNFlQ" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_Xhs844GCT62ObVhlNUom6Q" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_opgc9LH9QpWurqowqkqQWw" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_e9Nn4O5N_DbWbjMZmCTT5w" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_e9Nn4O5N_DbWbjMZmCTT5w"] .zpimage-container figure img { width: 1090px ; height: 614.50px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_e9Nn4O5N_DbWbjMZmCTT5w"] .zpimage-container figure img { width:723px ; height:407.60px ; } } @media (max-width: 767px) { [data-element-id="elm_e9Nn4O5N_DbWbjMZmCTT5w"] .zpimage-container figure img { width:415px ; height:233.96px ; } } [data-element-id="elm_e9Nn4O5N_DbWbjMZmCTT5w"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-fit zpimage-tablet-fallback-fit zpimage-mobile-fallback-fit hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-09%20at%205.53.30%20pm.png" width="415" height="233.96" loading="lazy" size="fit" alt="Diagram of our human feedback, reward model training, and policy training procedure" data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_7mp3GMzgRISVcOTV4SkgLA" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_7mp3GMzgRISVcOTV4SkgLA"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Artificial intelligence has become very capable at summarizing text. But current models are still trained indirectly - to mimic human reference summaries, not to directly satisfy users.</p><p><br></p><p>Researchers at OpenAI investigated whether getting regular people to compare AI-generated summaries can produce higher quality models.</p><p><br></p><p>Their technique worked as follows:</p><ul><li>They built a text summarization system based on a neural network language model.</li><li>For each piece of text, the system generated multiple summary options.</li><li>They showed human evaluators pairs of summaries for the same text and had them pick the better one.</li><li>They used these comparisons to train a separate &quot;reward model&quot; to predict which summary humans would prefer.</li><li>They then used this reward model to fine-tune the main summarization system, via a reinforcement learning algorithm. The system learns to generate summaries that score highly on the reward model.</li><li>By repeating this loop of gathering feedback, training the reward model, and tuning the summarizer, the system improves over time.</li></ul><p><br></p><p>The researchers found that optimizing the AI for direct human preferences significantly boosted performance compared to just training it to mimic reference summaries.</p><p><br></p><p>With enough feedback data, the AI's summaries surpassed the quality of the original human-written summaries used for training.</p><p><br></p><p>The key is the tight feedback loop between users and the system - the AI learns dynamically what people see as a high quality summary.</p><p><br></p><p>This technique provides a template for any application where humans can evaluate outputs, like speech or translation. The AI learns to satisfy users, not just match a benchmark.</p><p><br></p><p>For business leaders, it shows the value of measuring how well AI fulfills human needs, beyond just technical metrics. As AI advances, aligning it with people may require creative feedback techniques like this.</p><p><br></p><p>Source:</p><p><a href="https://arxiv.org/pdf/2009.01325.pdf" title="arxiv" rel="">arxiv</a><br></p><p></p></div><p></p></div>
</div><div data-element-id="elm_s0o8K2xNT1Zj1-nBy5RAuQ" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_s0o8K2xNT1Zj1-nBy5RAuQ"] .zpimage-container figure img { width: 800px ; height: 344.00px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_s0o8K2xNT1Zj1-nBy5RAuQ"] .zpimage-container figure img { width:500px ; height:215.00px ; } } @media (max-width: 767px) { [data-element-id="elm_s0o8K2xNT1Zj1-nBy5RAuQ"] .zpimage-container figure img { width:500px ; height:215.00px ; } } [data-element-id="elm_s0o8K2xNT1Zj1-nBy5RAuQ"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-large zpimage-tablet-fallback-large zpimage-mobile-fallback-large "><figure role="none" class="zpimage-data-ref"><a class="zpimage-anchor" href="/aibooks" target="" rel=""><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Untitled%20design%20-4-.png" width="500" height="215.00" loading="lazy" size="large"/></picture></a></figure></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:02:30 +1000</pubDate></item><item><title><![CDATA[What is RLHF: Reinforcement Learning from Human Feedback]]></title><link>https://www.nownextlater.ai/Insights/post/what-is-rlhf-reinforcement-learning-from-human-feedback</link><description><![CDATA[What is RLHF?]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_xpJigRzsT_KcaAE-7xC7zg" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_trBI0mD2SU-g9a70SuPlPA" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_-8DJeSS3T0yPrv_g82sSug" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"> [data-element-id="elm_-8DJeSS3T0yPrv_g82sSug"].zpelem-col{ border-radius:1px; } </style><div data-element-id="elm_QzsV8q9OTJayrFNGCpSsuA" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_QzsV8q9OTJayrFNGCpSsuA"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-center " data-editor="true"><div style="text-align:left;"><br></div><div style="text-align:left;"><div style="text-align:left;"><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">What is RLHF?</span></div></div><div style="color:inherit;text-align:left;"><p><br></p><p>Reinforcement learning from human feedback (RLHF) is an advanced technique to train large AI language models to generate higher quality text outputs. It incorporates direct human feedback into the training loop.</p><p><br></p><p>In simple terms, RLHF works by:</p><ol><li>Starting with a large pre-trained language model that can generate text snippets from prompts</li><li>Systematically collecting data on human preferences by having people compare and rank different AI-generated options from the same prompt. This preference data is used to train a separate &quot;reward model&quot; that will score new AI-generated text.</li><li>Using reinforcement learning algorithms to fine-tune and optimize the original language model to maximize the scores from the reward model when generating text, while constraining the model from changing too radically.</li></ol><p><br></p><p>The key innovation is creating training data from human comparisons of AI outputs, rather than relying on human-labeled datasets. This allows the AI to learn more complex, subjective attributes like creativity, truthfulness, and harmlessness that are difficult to directly program rules for.</p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;"><br></span></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Why does RLHF matter?</span></p><p></p><p><br></p><p>RLHF was a pivotal technique used by AI labs like Anthropic and OpenAI to develop large chatbot models with strong language generation abilities like Claude and ChatGPT.</p><p><br></p><p>It allows AI text generation to better align with human preferences and values, beyond just predicting the next word accurately. This makes the AI more usable for open-ended real world applications.</p><p><br></p><p>However, RLHF requires extensive data, compute resources, and engineering to scale up. There are also core limitations around dependency on ongoing human feedback and balancing performance gains with potential inaccuracies or harms.</p><div><br></div>
<p><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Current state and limitations</span></p><p></p><p><br></p><p>RLHF remains an active area of research with open challenges. There are some open source tools available but no simple universal implementation.</p><p><br></p><p>Key limitations include:</p><ol style="margin-left:40px;"><li>Costly data needs - RLHF requires substantial volumes of human feedback data, which can require extensive paid human labeling.</li><li>Potential for poor generalization - Performance is highly dependent on human annotator quality and biases. The AI has no inherent ground truth, so can learn harmful or factually inaccurate behaviors.</li></ol><p><br></p><p>Leading AI labs view advancing RLHF capabilities and mitigating limitations as important open problems. This includes developing new algorithms, analyzing model dynamics, and techniques to reduce harms.</p><p><br></p><p>RLHF allows cutting edge AI to better align with human values but still has major challenges around data, harms, and implementation complexity. It shows promise for the future with more research.</p></div></div>
</div><div data-element-id="elm_AqcaI1O26g126Akl0vQESA" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_AqcaI1O26g126Akl0vQESA"] .zpimage-container figure img { width: 800px ; height: 344.00px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_AqcaI1O26g126Akl0vQESA"] .zpimage-container figure img { width:500px ; height:215.00px ; } } @media (max-width: 767px) { [data-element-id="elm_AqcaI1O26g126Akl0vQESA"] .zpimage-container figure img { width:500px ; height:215.00px ; } } [data-element-id="elm_AqcaI1O26g126Akl0vQESA"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-large zpimage-tablet-fallback-large zpimage-mobile-fallback-large "><figure role="none" class="zpimage-data-ref"><a class="zpimage-anchor" href="/aibooks" target="" rel=""><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Untitled%20design%20-4-.png" width="500" height="215.00" loading="lazy" size="large" alt="AI Books"/></picture></a></figure></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Tue, 08 Aug 2023 15:19:18 +1000</pubDate></item></channel></rss>