Large language models (LLMs) often default to English for reasoning, even when responding in other languages, which can diminish their effectiveness and consistency in multilingual contexts. Researchers Shan Chen, Jirui Qi, and colleagues explore methods to encourage LLMs to maintain reasoning in the language of the user's query, revealing that small-scale supervised fine-tuning (SFT) can promote language consistency but sometimes at the expense of accuracy. They found that combining SFT with math-focused reinforcement learning (GRPO) can enhance accuracy on complex tasks without reverting to English reasoning, although challenges remain in low-resource languages like Japanese. The study suggests that model merging and targeted fine-tuning can help balance accuracy with language consistency, offering practical strategies for improving multilingual reasoning in AI models.