Close Menu
TFFH – The Financial Freedom Hub
    What's Hot

    Mindfulness Meditation | Longevity | Effortless Performance | Flow

    14/05/2025

    5 Highest-Yielding Dividend Aristocrats® That Could Sink or Soar (According to Analysts)

    14/05/2025

    1 Magnificent Defense Stock Down 23% to Buy and Hold Forever – TFFH – The Financial Freedom Hub

    14/05/2025
    Facebook X (Twitter) Instagram
    Facebook X (Twitter) Instagram
    TFFH – The Financial Freedom HubTFFH – The Financial Freedom Hub
    • Home
    • Money Basics
    • Budgeting 101
    • Saving Strategies
    • Debt Management
    • Emergency Funds
    • Credit & Loans
    • Youtube
    TFFH – The Financial Freedom Hub
    Home»Tech»Computing»This AI Paper Investigates Test-Time Scaling of English-Centric RLMs for Enhanced Multilingual Reasoning and Domain Generalization
    Computing

    This AI Paper Investigates Test-Time Scaling of English-Centric RLMs for Enhanced Multilingual Reasoning and Domain Generalization

    MathsXP.com By MathsXP.com14/05/2025No Comments4 Mins Read
    Facebook Twitter Pinterest LinkedIn Tumblr Email
    Share
    Facebook Twitter LinkedIn Pinterest Email

    Reasoning language models, or RLMs, are increasingly used to simulate step-by-step problem-solving by generating long, structured reasoning chains. These models break down complex questions into simpler parts and build logical steps to reach answers. This chain-of-thought (CoT) approach has proven effective in improving output quality, especially in mathematical and logical tasks. Despite multilingual capabilities in many modern large models, the focus of research and training has remained largely centered on English, leaving a gap in understanding how well these reasoning skills translate to other languages.

    One major challenge is that most RLMs are fine-tuned on English data, which limits their ability to reason effectively in other languages. This becomes especially problematic for low-resource languages that have limited training examples. The models may default to English thinking patterns, producing lower-quality outputs when prompted in another language. Furthermore, differences in language structure can cause reasoning errors, particularly when a model trained in one language is expected to infer logic in another without adequate linguistic alignment.

    Current techniques employ zero-shot or few-shot prompting strategies to manage these limitations, often using English as a pivot language. Some efforts involve presenting prompts in the same language as the query to preserve linguistic consistency. However, small models have minimal benefits due to limited capacity, and even large models show inconsistent performance when reasoning in low-resource languages. Despite multilingual pretraining, the gap between the training and reasoning language continues to hinder accurate multilingual reasoning.

    The Brown University and MBZUAI research team focused on evaluating how increasing test-time computation, particularly through extended reasoning chains, can affect the multilingual reasoning abilities of English-centric RLMs. They investigated using s1 models based on the Qwen2.5-Instruct architecture and fine-tuned on 1,000 English STEM reasoning samples. These models were tested across various languages using benchmarks like MGSM and Global-MMLU to answer four core questions: the effectiveness of crosslingual test-time scaling, language-mixing behaviors, performance under language-forcing, and cross-domain generalization.

    In-depth experiments showed that models with more parameters significantly benefited from increased test-time thinking tokens. The 14B s1 model, when scaled to 8,000 thinking tokens, achieved an average accuracy of 81% across non-English languages in MGSM. It outperformed models like Qwen2.5-14B-Instruct by +23.1% in French and +41.6% in Swahili. Even though the model was trained only in English, its performance surpassed that of larger models such as DeepSeek’s R1-Distill-Qwen-32B in several high-resource languages. The study also found that reasoning in high-resource languages like Chinese and English is more efficient, requiring fewer tokens and delivering better results than in low-resource languages like Swahili or Telugu.

    A key observation was the “quote-and-think” behavior, where the model quoted non-English phrases from prompts and reasoned in English. This consistent pattern across languages like Japanese and Russian suggested that the model used its multilingual understanding to interpret non-English input without direct translation. Language-forcing experiments further confirmed that forcing reasoning in high-resource languages yielded better results, while strict reasoning in low-resource languages led to significant accuracy drops and computational inefficiencies.

    Despite strong results in STEM-related tasks, performance gains did not transfer to domains like cultural commonsense or humanities. In benchmarks like FORK, increasing thinking tokens sometimes reduced performance, indicating overthinking. The study concludes that while test-time scaling enhances multilingual reasoning in high-resource languages, it does not generalize effectively to out-of-domain tasks or low-resource languages, indicating the need for further research on balanced multilingual training and domain adaptation.


    Check out the Paper. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter and don’t forget to join our 90k+ ML SubReddit.

    Here’s a brief overview of what we’re building at Marktechpost:

    Nikhil is an intern consultant at Marktechpost. He is pursuing an integrated dual degree in Materials at the Indian Institute of Technology, Kharagpur. Nikhil is an AI/ML enthusiast who is always researching applications in fields like biomaterials and biomedical science. With a strong background in Material Science, he is exploring new advancements and creating opportunities to contribute.


    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    MathsXP.com
    • Website

    Related Posts

    Mindfulness Meditation | Longevity | Effortless Performance | Flow

    14/05/2025

    1 Magnificent Defense Stock Down 23% to Buy and Hold Forever – TFFH – The Financial Freedom Hub

    14/05/2025

    Wall Street’s sudden rebound catches investors ‘offside’

    14/05/2025
    Add A Comment
    Leave A Reply Cancel Reply

    Latest post

    Mindfulness Meditation | Longevity | Effortless Performance | Flow

    14/05/2025

    5 Highest-Yielding Dividend Aristocrats® That Could Sink or Soar (According to Analysts)

    14/05/2025

    1 Magnificent Defense Stock Down 23% to Buy and Hold Forever – TFFH – The Financial Freedom Hub

    14/05/2025

    Investing for Your Kid’s Future

    14/05/2025

    Wall Street’s sudden rebound catches investors ‘offside’

    14/05/2025

    Whoop Has Changed Their Upgrade Policy

    14/05/2025

    How to Get Back on Track with YNAB

    14/05/2025

    Will AI Search Engines Cripple Google’s Dominance?

    14/05/2025

    Online Tennis Instruction Home – Online Tennis Instruction – Learn How To Play Your Best Tennis, Free Tennis Tips

    14/05/2025

    Pisces Man Secrets by Relationship Astrologer Anna Kovach » Books

    14/05/2025
    About The Financial Freedom Hub

    The Financial Freedom Hub is your go-to resource for mastering personal finance. We provide easy-to-understand guides, practical tips, and expert advice to help you take control of your money, budget effectively, save for the future, and manage debt. Whether you're just starting out or looking to refine your financial strategy, we offer the tools and knowledge you need to build a secure financial future. Start your journey to financial freedom with us today!

    Company
    • Home
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms and conditions
    Latest post

    Mindfulness Meditation | Longevity | Effortless Performance | Flow

    14/05/2025

    5 Highest-Yielding Dividend Aristocrats® That Could Sink or Soar (According to Analysts)

    14/05/2025

    1 Magnificent Defense Stock Down 23% to Buy and Hold Forever – TFFH – The Financial Freedom Hub

    14/05/2025

    Investing for Your Kid’s Future

    14/05/2025
    TFFH – The Financial Freedom Hub
    Facebook X (Twitter) Instagram YouTube
    • Home
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms and conditions
    © 2025 The Financial Freedom Hub. All rights reserved.

    Type above and press Enter to search. Press Esc to cancel.