<?xml version="1.0" encoding="UTF-8" ?><!-- generator=Zoho Sites --><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><atom:link href="https://www.nownextlater.ai/Insights/tag/bloomberggpt/feed" rel="self" type="application/rss+xml"/><title>Now Next Later AI - Blog #BloombergGPT</title><description>Now Next Later AI - Blog #BloombergGPT</description><link>https://www.nownextlater.ai/Insights/tag/bloomberggpt</link><lastBuildDate>Wed, 26 Nov 2025 21:35:25 +1100</lastBuildDate><generator>http://zoho.com/sites/</generator><item><title><![CDATA[A New LLM for Finance: BloombergGPT]]></title><link>https://www.nownextlater.ai/Insights/post/A-New-LLM-for-Finance-BloombergGPT</link><description><![CDATA[When designing BloombergGPT, a new large language model optimized for financial data, the researchers at Bloomberg made an important decision around tokenization.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_LgERJvdpQnmGEWGJ0BF4iQ" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_dlLM9Wo0TcqY30xfrQFVJw" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_-qyUhWkLTuK7dJU7MqTs4g" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"> [data-element-id="elm_-qyUhWkLTuK7dJU7MqTs4g"].zpelem-col{ border-radius:1px; } </style><div data-element-id="elm_HuJZ70YsFrNImBOyCB3h1A" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_HuJZ70YsFrNImBOyCB3h1A"] .zpimage-container figure img { width: 500px ; height: 327.53px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_HuJZ70YsFrNImBOyCB3h1A"] .zpimage-container figure img { width:500px ; height:327.53px ; } } @media (max-width: 767px) { [data-element-id="elm_HuJZ70YsFrNImBOyCB3h1A"] .zpimage-container figure img { width:500px ; height:327.53px ; } } [data-element-id="elm_HuJZ70YsFrNImBOyCB3h1A"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-10%20at%201.14.54%20pm.png" width="500" height="327.53" loading="lazy" size="medium" alt="Using BloombergGPT to generate short headline suggestions in a three-shot setting." data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_OuSKq5B78wofaYj1lQFryA" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_OuSKq5B78wofaYj1lQFryA"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><div style="color:inherit;"><p>Financial technology companies are increasingly turning to artificial intelligence to help analyze data, automate processes, and improve decision making. Bloomberg recently unveiled a new AI system called BloombergGPT that is specially designed to understand financial and business language.</p><p>BloombergGPT is what's known as a large language model (LLM).&nbsp;</p><p><br></p><p>LLMs are AI systems that are trained on massive amounts of text data so they can generate human-like text and engage in tasks like question answering and summarization. They have become very popular in recent years thanks to advances in computing power and AI techniques. Some well-known examples are systems like GPT-4 from OpenAI and BLOOM from Anthropic.</p><p><br></p><p>What makes BloombergGPT different is that it was explicitly trained on financial data sources curated by Bloomberg analysts over decades. This includes 360 billion tokens (words) from sources like financial news, earnings reports, regulatory filings, press releases, and more. The goal was to create an AI system optimized for understanding nuanced financial language.</p><p><br></p><p>To supplement the financial data, BloombergGPT was also trained on 345 billion tokens of more general text from publicly available sources like Wikipedia, books, academic papers, and web content. The combination of financial and general data was intended to make the system adept at both financial tasks and general natural language processing abilities.</p><p><br></p><p>In terms of its technical details, BloombergGPT contains 50 billion parameters. Parameters refer to the adjustable settings inside the model that are tuned during training. More parameters allow the system to learn more complex patterns and relationships. For comparison, GPT-3 has 175 billion parameters.</p><p><br></p><p>BloombergGPT uses an architecture based on the transformer, which is the most common framework used today for large language models. Transformers process text by looking at the entire context rather than processing words one-by-one. This allows them to develop a more holistic understanding of language.</p><p><br></p><p>During training, BloombergGPT was optimized using techniques like activation checkpointing and mixed precision to lower memory usage and increase speed. The result was a system capable of 102 teraflops, or 102 trillion floating point operations per second. This level of compute power was needed to effectively train the 50 billion parameter model.</p><p><br></p><p>To evaluate the capabilities of BloombergGPT, the researchers tested it on a range of financial natural language tasks as well as standard AI benchmarks. On financial tasks, BloombergGPT achieved state-of-the-art results, outperforming other models by significant margins. It performed very well on financial question answering, named entity recognition, and sentiment analysis.</p><p><br></p><p>On general benchmarks, BloombergGPT proved competitive with some models over 100 times its size. While it wasn't always the top performer, it consistently outscored similarly sized models on benchmarks measuring abilities like reasoning, reading comprehension and common sense knowledge.</p><p><br></p><p>The researchers attribute BloombergGPT's effectiveness to three main factors. First and foremost was the high-quality domain-specific data. Second, they believe <a href="https://www.nownextlater.ai/Insights/post/turning-text-into-data-how-bloomberggpt-tokenizes-financial-language" title="their choice of tokenizer" rel="">their choice of tokenizer</a> - the system responsible for breaking text into pieces the model can process - was beneficial. Finally, the model architecture and training techniques allowed them to efficiently train a model competitive with far larger systems.</p><p></p><p><br></p><p>The researchers highlight that creating an AI system of this scale still poses challenges. Instabilities can arise during training that require careful monitoring and intervention. They logged their training process in detail to aid future work.</p><p><br></p><p>Additionally, the authors considered ethical issues like potential biases in financial data and possible misuse of the system. Bloomberg has extensive procedures in place to reduce risks and ensure responsible AI development. However, the company chose not to publicly release the model to minimize chances of data leakage or misuse.</p><p><br></p><p>In conclusion, BloombergGPT represents a milestone for domain-specific natural language AI. Its training process and strong performance on financial tasks demonstrate the value of curated in-domain data. While specialized, the system retains competitive general abilities as well. As financial institutions continue adopting AI, expect systems like BloombergGPT to play an increasing role in driving automation and insights.</p><p><br></p><p>Sources:</p><a href="https://arxiv.org/pdf/2303.17564.pdf" title="BloombergGPT: A Large Language Model for Finance" rel="">BloombergGPT: A Large Language Model for Finance</a></div></div><p></p></div>
</div><div data-element-id="elm_JFp2iWupUd22UPKvlveDzw" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_JFp2iWupUd22UPKvlveDzw"] .zpimage-container figure img { width: 800px ; height: 344.00px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_JFp2iWupUd22UPKvlveDzw"] .zpimage-container figure img { width:500px ; height:215.00px ; } } @media (max-width: 767px) { [data-element-id="elm_JFp2iWupUd22UPKvlveDzw"] .zpimage-container figure img { width:500px ; height:215.00px ; } } [data-element-id="elm_JFp2iWupUd22UPKvlveDzw"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-large zpimage-tablet-fallback-large zpimage-mobile-fallback-large "><figure role="none" class="zpimage-data-ref"><a class="zpimage-anchor" href="/aibooks" target="" rel=""><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Untitled%20design%20-4-.png" width="500" height="215.00" loading="lazy" size="large"/></picture></a></figure></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 13:43:14 +1000</pubDate></item><item><title><![CDATA[Turning Text into Data: How BloombergGPT Tokenizes Financial Language]]></title><link>https://www.nownextlater.ai/Insights/post/turning-text-into-data-how-bloomberggpt-tokenizes-financial-language</link><description><![CDATA[When designing BloombergGPT, a new large language model optimized for financial data, the researchers at Bloomberg made an important decision around tokenization.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_PpY_r34GRc6y7cF_plYJgw" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_BAutgZsvRw2AA4u_wjzjUA" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_kU6fS6xMRwKzNS0fF1BgrA" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_ybwgRWJlSROZfRXPob3qLg" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_ybwgRWJlSROZfRXPob3qLg"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>One of the fundamental challenges in natural language processing is transforming free-form human language into data that machines can understand. The first step in this process is called tokenization – breaking down sentences and documents into small chunks or “tokens”.</p><p><br></p><p>When designing <a href="https://www.nownextlater.ai/Insights/post/A-New-LLM-for-Finance-BloombergGPT" title="BloombergGPT, a new large language model optimized for financial data" rel="">BloombergGPT, a new large language model optimized for financial data</a>, the researchers at Bloomberg made an important decision around tokenization. They opted to use a method called unigram tokenization rather than the more standard approach of byte pair encoding (BPE).</p><p></p><p>To understand this choice, it helps to first understand what tokenization involves and the strengths of different techniques.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">The Role of the Tokenizer</span></p><p></p><p><br></p><p>The job of a tokenizer is to split input text into tokens that contain one or more characters. Tokens become the basic unit that the model manipulates when processing language.</p><p><br></p><p>With English text, an obvious tokenization would be to split on spaces to get words and punctuation. But for machine learning, it’s common to break words down even further into subword units.</p><p><br></p><p>This offers two main advantages. First, it limits the size of the vocabulary that the model needs to represent. For example, instead of separate tokens for “training” and “trainer”, they can share a common root like “train”.</p><p><br></p><p>Second, the model can process words it hasn’t seen before by recognizing their subword components. This helps the model generalize.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Byte Pair Encoding</span></p><p></p><p><br></p><p>The most popular subword tokenization algorithm used in NLP is byte pair encoding (BPE). Originally developed for data compression, it builds up a vocabulary by scanning text and greedily merging frequent pairs of tokens.</p><p><br></p><p>BPE starts by assigning each character as a token. It then iteratively merges the most common pair of tokens until reaching a target vocabulary size. Popular pairs like “in+the” get merged early on.</p><p><br></p><p>This simple probabilistic approach works decently well in practice. BPE is fast to train and produces a vocabulary with reusable subword chunks.</p><p><br></p><p>However, BPE discards the probabilistic model after training. Each merge decision is hard-coded based on frequency alone. At test time, BPE tokenizes text greedily using this fixed vocabulary.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Unigram Tokenization</span></p><p></p><p></p><p><br></p><p>BloombergGPT implements a more advanced approach called unigram tokenization. A unigram tokenizer models the probability of tokens directly using techniques from statistical language modeling.</p><p><br></p><p>The probabilities allow unigram tokenizers to capture uncertainty and make “soft” decisions when tokenizing new text. Rather than greedily splitting input using a fixed vocabulary, it chooses the most probable segmentation based on context.</p><p><br></p><p>Unigram tokenizers are first trained on large datasets to learn these token probabilities. The training process gradually removes unlikely token candidates until reaching the target vocabulary size.</p><p><br></p><p>The main advantage over BPE is that unigram keeps the probabilistic model which allows for smarter, context-aware tokenization at inference time.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Implementation in BloombergGPT</span></p><p></p><p><br></p><p>For BloombergGPT, the unigram tokenizer was trained specifically on The Pile, a diverse dataset containing both general and domain-specific data.</p><p>After experimenting with different sizes, Bloomberg settled on a 217 token vocabulary (about 131,000 tokens). This is larger than typical for NLP models.</p><p>The researchers argue the customized tokenizer better captures financial terminology. And the large vocabulary encodes more meaning within each token.</p><p><br></p><p>In initial evaluations, BloombergGPT’s unigram tokenizer reduced the size of encoded text compared to BPE and other tokenizers. This suggests it forms an efficient token vocabulary for financial language.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;font-size:16px;">Tradeoffs and Considerations</span></p><p></p><p><br></p><p>Unigram tokenization has tradeoffs to consider. The probabilistic modeling approach requires more data and compute to train compared to BPE. The tokenization decisions are also less explainable than simple frequency counts.</p><p><br></p><p>Additionally, vocabularies tailored for specific domains may not generalize as well to other data. So there are open questions around how to balance domain-specific and general tokenization.</p><p><br></p><p>The choice ultimately depends on use cases. For a financial model like BloombergGPT, the benefits of smarter domain-aware tokenization appear to outweigh the costs. But simple, fast methods like BPE remain appealing in many scenarios.</p><p><br></p><p>As language models continue advancing, we’re likely to see more specialized tokenization strategies like unigram. The tokenizer plays a key role in how systems represent and process language. Improvements here translate to downstream gains, enabling models like BloombergGPT to push the state-of-the-art in domains like finance.</p></div><p></p></div>
</div><div data-element-id="elm_2jK8b-RORhTos0e4wduTqg" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_2jK8b-RORhTos0e4wduTqg"] .zpimage-container figure img { width: 800px ; height: 344.00px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_2jK8b-RORhTos0e4wduTqg"] .zpimage-container figure img { width:500px ; height:215.00px ; } } @media (max-width: 767px) { [data-element-id="elm_2jK8b-RORhTos0e4wduTqg"] .zpimage-container figure img { width:500px ; height:215.00px ; } } [data-element-id="elm_2jK8b-RORhTos0e4wduTqg"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-large zpimage-tablet-fallback-large zpimage-mobile-fallback-large "><figure role="none" class="zpimage-data-ref"><a class="zpimage-anchor" href="/aibooks" target="" rel=""><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Untitled%20design%20-4-.png" width="500" height="215.00" loading="lazy" size="large"/></picture></a></figure></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 13:43:14 +1000</pubDate></item></channel></rss>