<?xml version="1.0" encoding="UTF-8" ?><!-- generator=Zoho Sites --><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><atom:link href="https://www.nownextlater.ai/Insights/tag/translation/feed" rel="self" type="application/rss+xml"/><title>Now Next Later AI - Blog #Translation</title><description>Now Next Later AI - Blog #Translation</description><link>https://www.nownextlater.ai/Insights/tag/translation</link><lastBuildDate>Wed, 26 Nov 2025 21:23:59 +1100</lastBuildDate><generator>http://zoho.com/sites/</generator><item><title><![CDATA[No Language Left Behind: The Future of Human-Centered Machine Translation]]></title><link>https://www.nownextlater.ai/Insights/post/no-language-left-behind-the-future-of-human-centered-machine-translation</link><description><![CDATA[While machine translation has improved dramatically for major languages, quality translation still lags for thousands of smaller languages.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_ALAXDqC_QR-T1RkNxbnnkQ" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_7Wn1QaltR3S2LWxVZlhqHQ" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_ws-tX1SeQL-ffS7jcmtWyA" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_oavVHq8mRomFgifCnxLQRw" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_oavVHq8mRomFgifCnxLQRw"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><p style="font-size:16px;font-weight:400;text-indent:0px;"><strong style="font-weight:600;"></strong>In our increasingly interconnected world, language remains a persistent barrier. While machine translation has improved dramatically for major languages like English, Spanish, and Mandarin, quality translation still lags for thousands of smaller languages.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">A new effort called No Language Left Behind (NLLB) aims to close this gap by developing high-quality translation systems covering over 200 languages. The project employs a human-centered approach to translation, collaborating with speakers of diverse languages to ensure cultural relevance. This responds directly to the needs illuminated in interviews with translators and language communities, who found existing systems unreliable.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">Creating useful systems requires quality training data. But most languages lack large parallel corpora needed for machine translation. NLLB’s ingenious solution mines the web to extract potential translations from multilingual websites. Using specialized cross-lingual sentence embedding models, they align sentences conveying similar meanings in different languages with high accuracy. This novel mining process enabled building datasets for low-resource languages at unprecedented scale.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">For evaluation, NLLB created the Flores benchmark spanning 200 languages. Careful protocols ensure accurate human translation and alignment. Flores helps quantify progress, with higher scores indicating better translation quality. NLLB models achieve state-of-the-art results on Flores, with many languages nearing usability. The project also conducted human studies confirming considerable gains over past systems.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">Another key innovation was tailoring the training process to prevent overfitting on high-resource languages. Curriculum learning introduces low-resource languages slowly to focus model capacity. A mixture-of-experts architecture further reduced interference between languages. Together these advances allowed a single model to translate between all 200+ directions while maintaining specialization.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">The NLLB project has committed to ethics and inclusion. Their partnership with Wikipedia is enabling content translation for many smaller wikipedias. With potential harms like cultural erasure in mind, the project also interviewed speakers from vulnerable language communities to guide decision making. Framing translation as a human right, the goal is to enhance global communication.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">NLLB's ambitions embody AI's expanding potential to connect us. Machine translation was long considered nearly hopeless for low-resource languages. But the project upends these assumptions by proving quality translation possible for all tongues when equity becomes a design principle.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">NLLB believes openness and collaboration will be central to fulfilling this vision. Models and datasets are fully public to catalyze innovation. The project also plans grants, support for adapting their tools, and partnerships with researchers invested in specific languages. Only through such collective effort can we ensure no language gets left out.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">NLLB faces challenges in improving under-served language pairs, capturing informal dialects, and evaluation beyond text. But these are solvable given dedicated investment and commitment. Progress shows what's possible when innovation grounds itself in lived experience.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">Past AI has often encoded the priorities of English-speaking elites. NLLB points toward an alternative future, designed for international users. When technology accommodates our differences rather than erasing them, new opportunities for connection emerge.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">No Language Left Behind offers a roadmap for inclusive AI. Its translation advancements - enabled by novel datasets, training methods, and modeling architectures - carry lessons for ensuring equity more widely. Expanding access requires investigating deficiencies and confronting barriers directly.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">The project's mission resonates far beyond translation, to all intelligent systems. For as long as exclusion persists, the work remains unfinished. These efforts bring us closer to AI that knows no language boundaries, only the poetry of voices united across divides.</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;">source:</p><p style="font-size:16px;font-weight:400;text-indent:0px;"><a href="https://arxiv.org/pdf/2207.04672.pdf" title="arxiv" rel="">arxiv</a><br></p><p style="font-size:16px;font-weight:400;text-indent:0px;"></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:04:29 +1000</pubDate></item><item><title><![CDATA[Taming the Tower of Babel - Making AI Translate Many Languages]]></title><link>https://www.nownextlater.ai/Insights/post/taming-the-tower-of-babel-making-ai-translate-many-languages</link><description><![CDATA[Machine translation has improved immensely thanks to AI, but handling multiple languages remains tricky. Researchers from Tel Aviv University and Meta studied this challenge. Through systematic experiments, they uncovered what really causes the most interference.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_reUMZ3q-Rbe5xFjCr7z9QA" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_7zoS_ESzSPWiPODR08jRyQ" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_KdaN5wPRR3GsDetE1cxFxg" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_hPA5GJeuQV2pmtjXUjbcKQ" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_hPA5GJeuQV2pmtjXUjbcKQ"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><div style="color:inherit;"><p><span style="color:inherit;">Machine translation has improved immensely thanks to AI, but handling multiple languages remains tricky. When you train one model to translate between English, Spanish, French and more, the languages can “interfere” with each other.</span></p><p><span style="color:inherit;"><br></span></p><p><span style="color:inherit;">Researchers from Tel Aviv University and Meta studied this challenge. Through systematic experiments, they uncovered what really causes the most interference.</span></p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;">What Causes the Problems?</span></p><p><br></p><p>The researchers found two main things cause issues:</p><ol><li>Small models struggle with diverse data. When you add more languages, they get confused.</li><li>Low-data languages don't get enough examples. Spanish has tons of text to learn from. But Swahili does not.</li></ol><p><br></p><p>Other factors like language similarity mattered much less. Having more languages wasn't so bad either.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;">How to Make Translation Better</span></p><p><br></p><p>The team found two simple solutions:</p><ol><li>Use bigger models. Large AI models handle diverse languages better. The extra capacity reduces the mixed-up translations.</li><li>Balance data proportions. Ensure low-data languages get sampled enough during training. Tuning this hyperparameter helped low-data languages improve.</li></ol><p>Other papers invented algorithms to reduce the language mixing-up. But this study showed basic solutions of scaling up and balancing data work well.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;">The Lesson for Business</span></p><p><br></p><p>This teaches an important lesson about AI. Advances often come from more computing power and data, not just clever new ideas. Getting the basics right matters most.</p><p><br></p><p>For business leaders, it shows the value of dedicating resources to train large AI models. Bigger models accommodate diverse data better. It also reduces the need for exotic algorithms that may not work that well.</p><p><br></p><p>Investing in computing enables handling diverse data well. Keeping solutions simple is usually the best path to success in AI. <span style="color:inherit;">Scaling up does have downsides - it costs more and has environmental impacts. </span></p></div><p><br></p><p>Sources:</p><p><a href="https://arxiv.org/abs/2212.07530" title="arxiv" rel="">arxiv</a><br></p><p></p></div><p></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 08:00:44 +1000</pubDate></item><item><title><![CDATA[Lost in Translation: When Context Matters for AI]]></title><link>https://www.nownextlater.ai/Insights/post/lost-in-translation-when-context-matters-for-ai</link><description><![CDATA[Researchers at Carnegie Mellon University and University of Lisbon systematically studied when context is needed for high-quality translation across 14 languages.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm__V34qgNpQTioLhZ-fnvtZQ" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_b2hcOGJoQg6A2N6Yf0Mp_g" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_HoSWrQVTRDyH2EhLNpZbHQ" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_zAgC_YT8n3k--dCY0u8Aww" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_zAgC_YT8n3k--dCY0u8Aww"] .zpimage-container figure img { width: 1090px ; height: 472.79px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_zAgC_YT8n3k--dCY0u8Aww"] .zpimage-container figure img { width:723px ; height:313.60px ; } } @media (max-width: 767px) { [data-element-id="elm_zAgC_YT8n3k--dCY0u8Aww"] .zpimage-container figure img { width:415px ; height:180.01px ; } } [data-element-id="elm_zAgC_YT8n3k--dCY0u8Aww"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-fit zpimage-tablet-fallback-fit zpimage-mobile-fallback-fit hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-08%20at%206.43.46%20pm.png" width="415" height="180.01" loading="lazy" size="fit" alt="Examples of high P-CXMI tokens and corresponding linguistic phenomena." data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_dFEu9et1RIOb9GhP9IH05g" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_dFEu9et1RIOb9GhP9IH05g"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Recent advances in generative artificial intelligence have led to great improvements in machine translation, with AI systems able to translate text between languages with impressive quality. However, new research reveals these systems still struggle with ambiguity and discourse phenomena that require broader context beyond individual sentences.</p><p><br></p><p>Researchers at Carnegie Mellon University and University of Lisbon systematically studied when context is needed for high-quality translation across 14 languages. By analyzing AI model predictions, they identified common ambiguity patterns including:</p><ul style="margin-left:40px;"><li>Lexical cohesion - Using consistent words to refer to the same entity across a document.</li><li>Formality - Properly translating formal vs informal pronouns based on social relationships.</li><li>Pronouns - Identifying the correct gendered pronoun when translating.</li><li>Verb forms - Choosing the right verb tense and conjugation to match the tone and flow.</li><li>Ellipsis - Filling in omitted verbs and subjects that require inferred meaning.</li></ul><p><br></p><p>The study found even state-of-the-art commercial systems struggle with these discourse phenomena compared to translating simpler, unambiguous sentences. The AI models only showed marginal improvements when given surrounding context, suggesting current techniques are inadequate.</p><p>For business leaders employing machine translation, these findings reveal a gap between today's AI capabilities and the nuance of human language. While fine for simple uses, ambiguity and discourse still pose a major challenge. For mission-critical applications, companies should be aware of potential errors with discourse and not fully trust black-box AI systems.</p><p><br></p><p>Advancing natural language AI to properly incorporate discourse and context remains an open research problem. As models ingest more diverse training data and learn complex reasoning, the goal is to reach human-level mastery of pragmatics. Moving forward, evaluation benchmarks that test contextual understanding will be crucial to drive progress.</p><p><br></p><p>The road to fully fluent AI is still long, but identifying specific weaknesses around ambiguity lays the groundwork for breakthroughs. With focused efforts to handle discourse, translation models that know what truly matters in the broader context may soon get the gist across languages.</p><p><br></p><p>Source:</p><p><a href="https://arxiv.org/pdf/2109.07446.pdf" title="arxiv" rel="">arxiv</a><br></p><p></p></div><p></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Thu, 10 Aug 2023 07:58:11 +1000</pubDate></item></channel></rss>