2 Sources
[1]
OpenAI's newest AI model is switching languages to Chinese and others while reasoning, puzzling users and experts
Serving tech enthusiasts for over 25 years. TechSpot means tech analysis and advice you can trust. WTF?! OpenAI's latest AI model, o1, has been displaying unexpected behavior that has captured the attention of both users and experts. Designed for reasoning tasks, the model has been observed switching languages mid-thought, even when the initial query is presented in English. Users across various platforms have reported instances where OpenAI's o1 model begins its reasoning process in English but unexpectedly shifts to Chinese, Persian, or other languages before delivering the final answer in English. This behavior has been observed in a range of scenarios, from simple counting tasks to complex problem-solving exercises. One Reddit user commented, "It randomly started thinking in Chinese halfway through," while another user on X questioned, "Why did it randomly start thinking in Chinese? No part of the conversation (5+ messages) was in Chinese." The AI community has been buzzing with theories to explain this unusual behavior. While OpenAI has yet to issue an official statement, experts have put forward several hypotheses. Some, including Hugging Face CEO Clément Delangue, speculate that the phenomenon could be linked to the training data used for o1. Ted Xiao, a researcher at Google DeepMind, suggested that reliance on third-party Chinese data labeling services for expert-level reasoning data might be a contributing factor. "For expert labor availability and cost reasons, many of these data providers are based in China," said Xiao. This theory posits that the Chinese linguistic influence on reasoning could be a result of the labeling process used during the model's training. Another school of thought suggests that o1 might be selecting languages it deems most efficient for solving specific problems. Matthew Guzdial, an AI researcher and assistant professor at the University of Alberta, offered a different perspective in an interview with TechCrunch: "The model doesn't know what language is, or that languages are different. It's all just text to it," he explained. This view implies that the model's language switches may stem from its internal processing mechanics rather than a conscious or deliberate choice based on linguistic understanding. Tiezhen Wang, a software engineer at Hugging Face, suggests that the language inconsistencies could stem from associations the model formed during training. "I prefer doing math in Chinese because each digit is just one syllable, which makes calculations crisp and efficient. But when it comes to topics like unconscious bias, I automatically switch to English, mainly because that's where I first learned and absorbed those ideas," Wang explained. While these theories offer intriguing insights into the possible causes of o1's behavior, Luca Soldaini, a research scientist at the Allen Institute for AI, emphasizes the importance of transparency in AI development. "This type of observation on a deployed AI system is impossible to back up due to how opaque these models are. It's one of the many cases for why transparency in how AI systems are built is fundamental," Soldaini said.
[2]
Why Does ChatGPT's Algorithm 'Think' in Chinese?
OpenAI's new reasoning model is doing weird, unpredictable stuff. OpenAI recently unveiled a new algorithm, o1, which the company calls its "reasoning" model. The idea behind the algorithm is that it spends "more time thinking" before it responds, thus delivering better answers. However, the algorithm appears to "think" not just in English but in several different languages. Web users recently noticed that the program was randomly displaying Chinese characters, as well as code that appeared to be written in several other languages. Most people will only pay attention to the final output that ChatGPT produces, but users have the option to look at how the model is reasoning its way to an answer. That's where many people noticed that the LLM had begun to incorporate Mandarin and Cantonese into its process. "Why did o1 pro randomly start thinking in Chinese?" asked Rishab Jain, via X. "No part of the conversation (5+ messages) was in Chinese... very interesting... training data influence." Prior to that, on January 5th, another web user, Nero, wrote on X: "uhmm, why is my gpt o1 thinking in chinese, lol." They tagged both OpenAI and ChatGPT in their message but received no reply. The obvious answer would appear to be that the algorithm was likely trained on large tranches of Chinese data, and, thus, that data is influencing the algorithm's output. "Most probable explanations - certain languages might offer tokenization efficiencies or easier mappings for specific problem types," said Rohan Paul, an AI engineer. "So, o1 may be switching languages because its internal representation of knowledge finds that using Chinese can lead to more optimized computation paths when handling certain problems." Another online commentator, Raj Mehta, gave a similar explanation: "o1, like many large language models (LLMs), operates in a shared latent space where concepts are abstract, not tied to specific languages. It might "reason" in the language that maps most efficiently to the problem at hand." Gizmodo reached out to OpenAI for comment but didn't receive an immediate explanation. There's plenty of speculation flying around about why this is happening, though TechCrunch interviewed Luca Soldaini, a research scientist at the Allen Institute for AI, who seemed to offer the best response to the conundrum. Soldeini said that, due to the opaque nature of the company's algorithms, there's really no way to know why the program is behaving in the way that it is. “This type of observation on a deployed AI system is impossible to back up due to how opaque these models are,†Soldaini told the outlet. “It’s one of the many cases for why transparency in how AI systems are built is fundamental.†Indeed, the black-box nature of corporate AI algorithms is particularly ironic given OpenAI's self-stated "open" mission of transparent technological development. "OpenAI isn't so open, which means that when its algorithms do weird things like spout Chinese, all we can really do is scratch our heads and wonder why."
Share
Copy Link
OpenAI's latest AI model, o1, designed for reasoning tasks, has been observed switching languages mid-thought, sparking curiosity and debate among users and AI experts about its training data and internal processes.
OpenAI's latest AI model, o1, has captured the attention of users and experts alike with its peculiar behavior of switching languages mid-reasoning. Designed for complex problem-solving tasks, the model has been observed transitioning from English to Chinese, Persian, and other languages before delivering final answers in English 12.
Users across various platforms have reported instances of o1's language-switching behavior. One Reddit user noted, "It randomly started thinking in Chinese halfway through," while another on X (formerly Twitter) questioned, "Why did it randomly start thinking in Chinese? No part of the conversation (5+ messages) was in Chinese" 1. This unexpected phenomenon has sparked curiosity and debate within the AI community.
Several theories have emerged to explain o1's multilingual reasoning:
Training Data Influence: Clément Delangue, CEO of Hugging Face, suggests that the behavior could be linked to the training data used for o1 1.
Chinese Data Labeling: Ted Xiao, a researcher at Google DeepMind, proposes that the reliance on Chinese data labeling services for expert-level reasoning data might be a contributing factor 1.
Efficiency in Problem-Solving: Some experts, like Rohan Paul, an AI engineer, speculate that o1 might be switching languages based on which one offers the most efficient computation path for specific problems 2.
Abstract Concept Representation: Raj Mehta suggests that o1, like many large language models, operates in a shared latent space where concepts are abstract and not tied to specific languages 2.
Matthew Guzdial, an AI researcher and assistant professor at the University of Alberta, offers a different perspective: "The model doesn't know what language is, or that languages are different. It's all just text to it" 1. This implies that the language switches may stem from the model's internal processing mechanics rather than a conscious choice based on linguistic understanding.
Tiezhen Wang, a software engineer at Hugging Face, proposes that the language inconsistencies could be due to associations formed during training. Wang explains, "I prefer doing math in Chinese because each digit is just one syllable, which makes calculations crisp and efficient. But when it comes to topics like unconscious bias, I automatically switch to English, mainly because that's where I first learned and absorbed those ideas" 1.
Luca Soldaini, a research scientist at the Allen Institute for AI, emphasizes the importance of transparency in AI development. "This type of observation on a deployed AI system is impossible to back up due to how opaque these models are. It's one of the many cases for why transparency in how AI systems are built is fundamental," Soldaini stated 12.
As the AI community continues to grapple with o1's unexpected behavior, the incident highlights the complexities of advanced language models and the need for greater transparency in their development and deployment.
Elon Musk's companies X and xAI have filed a lawsuit against Apple and OpenAI, alleging anticompetitive practices in the integration of ChatGPT into iOS, claiming it stifles competition in the AI chatbot market.
50 Sources
Technology
5 hrs ago
50 Sources
Technology
5 hrs ago
YouTube has been secretly testing AI-powered video enhancement on select Shorts, leading to backlash from creators who noticed unexpected changes in their content. The platform claims it's using traditional machine learning, not generative AI, to improve video quality.
7 Sources
Technology
5 hrs ago
7 Sources
Technology
5 hrs ago
Leading tech firms and investors create a network of political action committees to advocate for AI-friendly policies and oppose strict regulations ahead of the 2026 midterms.
5 Sources
Policy
5 hrs ago
5 Sources
Policy
5 hrs ago
Perplexity AI introduces Comet Plus, a subscription service that shares revenue with publishers when their content is used by AI tools, addressing concerns about fair compensation in the era of AI-powered search and content generation.
7 Sources
Technology
5 hrs ago
7 Sources
Technology
5 hrs ago
Nvidia announces plans to implement silicon photonics and co-packaged optics for AI GPU communication by 2026, promising higher transfer rates and lower power consumption in next-gen AI data centers.
2 Sources
Technology
21 hrs ago
2 Sources
Technology
21 hrs ago