4 Sources
[1]
OpenAI beats Elon Musk's Grok in AI chess tournament
ChatGPT-maker OpenAI has beaten Elon Musk's Grok in the final of a tournament to crown the best artificial intelligence (AI) chess player. Historically, tech companies have often used chess to assess the progress and abilities of a computer, with modern chess machines virtually unbeatable against even the top human players. But this competition did not involve computers designed for chess - instead it was held between AI programs designed for everyday use. OpenAI's o3 model emerged unbeaten in the tournament and defeated xAI's model Grok 4 in the final, adding fuel to the fire of an ongoing rivalry between the two firms.
[2]
Sam Altman's OpenAI Crushes Elon Musk's Grok in AI Chess Championship - Decrypt
Sam Altman's OpenAI o3 model -- which was deprecated late last week with the release of GPT-5 -- demolished Elon Musk's Grok 4 in four straight games Thursday to win Google's Kaggle Game Arena AI Chess Exhibition. You may think it was a super complex spectacle of high tech behemoths putting their reasoning to the ultimate test, but as an appetizer, let's say world champion Magnus Carlsen compared both bots to "a talented kid who doesn't know how the pieces move." The three-day tournament, which ran August 5-7, forced general-purpose chatbots -- yes, the same ones that help you write email and claim to be approaching human-level intelligence -- to play chess without any specialized training. No chess engines, no looking up moves, just whatever chess knowledge they'd randomly absorbed from the internet. The results were about as elegant as you'd expect from forcing a language model to play a board game. Carlsen, who co-commentated the final, estimated both AIs were playing at the level of casual players who recently learned the rules -- around 800 ELO. For context, he's arguably the best chess player who ever lived, with an ELO of 2839 points. These AIs were playing like they'd learned chess from a corrupted PDF. "They oscillate between really, really good play and incomprehensible sequences," Carlsen said during a broadcast, following the game. At one point, after watching Grok walk its king directly into danger, he joked it might think they were playing King of the Hill instead of chess. The actual games were like a masterclass in how not to play chess, even for those who don't know the game. In the first match, Grok essentially gave away one of its important pieces for free, then made things worse by trading off more pieces while already behind. Game two got even weirder. Grok tried to execute what chess players call the "Poisoned Pawn" -- a risky but legitimate strategy where you grab an enemy pawn that looks free but isn't. Except Grok grabbed the wrong pawn entirely, one that was obviously defended. Its queen (the most powerful piece in the board) got trapped and captured immediately. By game three, Grok had built what looked like a solid position -- good positional control, no obvious dangers, and basically a set up that can help you win the match. Then in mid game, it basically fumbled the ball directly to the opponent. It lost piece after piece in rapid succession. This was actually weird, considering that before the match against o3, Grok was a pretty strong contender, showing solid potential -- so much that the chess Grand Master Hikaru Nakamura praised it. "Grok is easily the best so far, just being objective, easily the best." The fourth (and last) game provided the only genuine suspense. OpenAI's o3 made a massive blunder early in the game, which is a big danger in any reasonable match. Nakamura, who was streaming the match, said there were still "a few tricks" left for o3 despite the disadvantage. He was right -- o3 clawed back to win its queen back and slowly squeezed out a victory while Grok's endgame play fell apart like wet cardboard. "Grok made so many mistakes in these games, but OpenAI did not," Nakamura said during his livestream. This was quite the reversal from earlier in the week. The timing couldn't have been worse for Elon Musk. After Grok's strong early rounds, he'd posted on X that his AI's chess abilities were just a "side effect" and that xAI had "spent almost no effort on chess." That turned out to be an understatement. Before this "official" chess tournament, International Master Levy Rozman hosted his own tournament earlier this year with less advanced models. He respected all the moves the chatbots recommended, and the whole situation ended up being a complete mess with illegal moves, piece summonings, and incorrect calculations. Stockfish, an AI built specifically for chess, ended up winning the tournament against ChatGPT. Altman's AI was matched against Musk's in the semifinals, and Grok lost. So it's 2-0 for Sam. However, this tournament was different. Each bot got four chances to make a legal move -- if they failed four times, they automatically lost. This wasn't hypothetical. In early rounds, AIs tried to teleport pieces across the board, bring dead pieces back to life, and move pawns sideways like they were playing some fever-dream version of chess they'd invented themselves. They got disqualified. Google's Gemini grabbed third place by beating another OpenAI model, salvaging some dignity for the tournament organizers. That bronze medal match featured a particularly absurd drawn game where both AIs had completely winning positions at different points but couldn't figure out how to finish. Carlsen pointed out that the AIs were better at counting captured pieces than actually delivering checkmate -- they understood material advantage but not how to win. It's like being great at collecting ingredients but unable to cook a meal. These are the same AI models that tech executives claim are approaching human intelligence, threatening white-collar jobs, and revolutionizing how we work. Yet they can't play a board game that has existed for 1,500 years without trying to cheat or forgetting the rules. So it's probably safe to say we're safe, AI won't take control of humanity, for now.
[3]
OpenAI's o3 Beats Grok 4 to Win AI Chess Tournament | AIM
Just a day before OpenAI's o3 reasoning model was officially retired from ChatGPT on August 8 in favour of GPT-5 with built-in reasoning capabilities, it won Kaggle's AI Chess Exhibition Tournament. The model beat xAI's Grok-4 model to win the finals 4-0. The model reached the finals, beating the Kimi K2 Instruct model in the quarter-finals and the company's own o4-mini model with the same scoreline. "Up until the semifinals, it seemed like nothing would be able to stop Grok 4 on its way to winning the event," said Chess.com, the world's leading platform for online chess. "But the illusion fell through on the last day of the tournament. The chatty o3 simply dismantled its mysterious opponent with four convincing wins. Grok's play was unrecognisabl
[4]
ChatGPT defeats Grok 4 to win chess tournament for AI - VnExpress International
With a 4-0 win over Elon Musk's AI Grok 4 in the final, ChatGPT's o3 Large Language Model (LLM) took the crown at the first AI chess tournament. About an hour before the final, OpenAI announced the launch of its 11th generation LLM, called GPT-5. However, the model ChatGPT used in the final was still o3 - the company's strongest reasoning product. Facing xAI's Grok 4, which had performed well in the group stage, o3 showed superior strength with an average move accuracy rate of 90.8%, compared to Grok 4's 80.2%, according to Chess.com. In all four games, ChatGPT checkmated its opponent after 35, 30, 28 and 54 moves. World number two Hikaru Nakamura said Grok 4 seemed nervous and anxious in this match, so it did not perform as well as in the previous two. It often lost pieces easily in the final match, which rarely happened in victories over Gemini 2.5 Flash and Gemini 2.5 Pro from Google. The o3 model ended the tournament with three wins with a score of 4-0, with an average accuracy rate of 12 games above 91%. Although its strength is not comparable to that of a grandmaster, players with an online chess rating of 2,000 or lower may have difficulty facing o3, especially in blitz or super blitz games. ChatGPT, a product of OpenAI, is a pioneer chatbot in the AI revolution, launched on Nov. 30, 2022. At that time, it used the GPT-3, a general-purpose model. The GPT line represents versatility, while the o lines are more inclined towards reasoning. The o3 was released in January 2025, while the o4-mini appeared three months later. These two models are also representatives of OpenAI in the first AI chess tournament in history. Grok 4 is a creation by xAI, owned by the world's richest man Elon Musk. He said that Grok 4 had hardly learned chess before participating in this tournament. The tournament was organized by Google on the Kaggle platform for three days between Aug. 5-7, with 8 AIs competing in a single-elimination format. The two Chinese representatives, Kimi K4 and Deepseek, were both eliminated early in the quarterfinals, with heavy losses. The remaining six representatives in the tournament are all from American companies. These are the strongest AIs in the world. In the third-place match, Gemini 2.5 Pro beat o4-mini with a score of 3.5-0.5. The LLMs participating in the tournament were all non-chess specialists. At the same time, another chess tournament was taking place between eight chess engines, which received less attention. These eight engines had Elo ratings ranging from 3,576 (Integral) to 3,731 (Stockfish). Some of them used AI algorithms to improve their performance. They were all far above human level.
Share
Copy Link
OpenAI's o3 model decisively defeated Elon Musk's Grok 4 in the final of Google's Kaggle Game Arena AI Chess Exhibition, showcasing the capabilities and limitations of general-purpose AI in specialized tasks.
In a groundbreaking event that pitted general-purpose AI models against each other in the ancient game of chess, OpenAI's o3 model emerged victorious in Google's Kaggle Game Arena AI Chess Exhibition. The tournament, held from August 5-7, saw o3 decisively defeat Elon Musk's xAI's Grok 4 in the final with a clean sweep of 4-0 12.
Source: BBC
The three-day tournament featured eight AI models competing in a single-elimination format. Notably, all participants were general-purpose language models rather than specialized chess engines. The competition included entries from major tech companies, with six of the eight participants coming from American firms 4.
Despite the advanced nature of these AI models, their chess performance was surprisingly rudimentary. World champion Magnus Carlsen, who co-commentated the final, estimated both finalists were playing at a level of around 800 ELO, comparable to casual players who recently learned the rules 2. This stark contrast to specialized chess engines, which can easily outperform even the best human players, highlights the current limitations of general-purpose AI in specialized tasks.
The games were characterized by a mix of occasional brilliant moves and frequent blunders. Chess Grandmaster Hikaru Nakamura observed that the AIs "oscillate between really, really good play and incomprehensible sequences" 2. Some particularly amusing moments included:
This tournament serves as a reality check for the current state of AI technology. While these models excel in language processing and general knowledge tasks, their struggle with the structured rules and strategy of chess reveals significant gaps in their reasoning capabilities 23.
Source: VnExpress International
The timing of the tournament coincided with some interesting developments in the AI world:
While the tournament showcased the current limitations of general-purpose AI in specialized tasks, it also highlighted the rapid progress being made in the field. As these models continue to evolve, it will be interesting to see how their performance in such specialized domains improves 34.
Source: Decrypt
The event has sparked discussions about the nature of AI intelligence and the challenges that remain in creating truly versatile artificial general intelligence. As the dust settles on this unique chess tournament, the AI community is left with valuable insights into the strengths and weaknesses of current language models, paving the way for future advancements in the field 234.
Summarized by
Navi
[3]
[4]
VnExpress International
|ChatGPT defeats Grok 4 to win chess tournament for AI - VnExpress InternationalApple plans to launch a major upgrade to Siri with App Intents in spring 2026, promising enhanced voice control across apps. The update faces challenges but could revolutionize iPhone interactions.
5 Sources
Technology
11 hrs ago
5 Sources
Technology
11 hrs ago
Chinese state-affiliated media criticizes Nvidia's H20 AI chips, claiming they pose security risks and are technologically inferior. Nvidia denies these allegations, defending the integrity of their products amidst escalating US-China tech tensions.
10 Sources
Technology
11 hrs ago
10 Sources
Technology
11 hrs ago
China is pushing for the US to ease restrictions on AI chip exports, particularly high-bandwidth memory chips, as part of trade negotiations ahead of a potential summit between Presidents Xi and Trump.
2 Sources
Business and Economy
11 hrs ago
2 Sources
Business and Economy
11 hrs ago
Amazon introduces Alexa+, a major AI upgrade to its voice assistant, aiming to compete with ChatGPT's conversational abilities. While offering improved features, early testing reveals significant bugs and reliability issues.
2 Sources
Technology
19 hrs ago
2 Sources
Technology
19 hrs ago
Recent computer science graduates face unprecedented unemployment rates as AI tools and industry layoffs reshape the tech job market, challenging long-held promises of prosperity in the field.
2 Sources
Technology
3 hrs ago
2 Sources
Technology
3 hrs ago