AI Chess Showdown: ChatGPT, Gemini, and Grok vs. Stockfish Ends in Chaos
Artificial Intelligence (AI) systems, including ChatGPT, Gemini, and Grok, recently participated in a series of chess matches against the professional chess engine Stockfish. While the matches began with standard play, they quickly descended into chaos as the AI players displayed bizarre rule-breaking behavior. Here’s a breakdown of this unexpected showdown.
Snapchat AI’s Unconventional Moves
The first match featured Stockfish against Snapchat AI. Initially, Snapchat AI followed basic chess principles, developing a standard opening. However, it soon began to defy the rules:
- Moved a knight from an invalid position directly to the center of the board.
- Allowed its king to capture its own bishop to escape a check.
- Returned the bishop to the board arbitrarily and began moving pawns sideways, defying chess logic.
Gemini vs. Grok: A Match of Errors
In the second match, Gemini faced Grok. Both AIs started by adhering to standard chess moves, but the game quickly unraveled:
- Grok’s errors: It made seven critical mistakes, including positioning its queen dangerously without repercussions.
- Gemini’s missed opportunities: Despite Grok’s blunders, Gemini failed to capitalize on them, showcasing poor situational awareness.
- Both AIs ignored fundamental chess rules, placing pieces on forbidden squares and disrupting the game flow.
ChatGPT vs. Meta AI: The Chessboard Anarchy
The match between ChatGPT and Meta AI was marked by randomness and creativity—but not in a good way:
- ChatGPT’s missteps: While initially following the English opening, ChatGPT soon began generating nonexistent chess pieces and placing them on illegal squares.
- Meta AI’s “telekinesis”: It escalated the chaos by moving ChatGPT’s pieces. In response, ChatGPT declared a checkmate despite the king being entirely safe.
The game ended with ChatGPT claiming victory, though it was far from a standard chess match.
ChatGPT Takes on Stockfish
In the final match, ChatGPT faced the formidable Stockfish. Initially, the chatbot made a bold pawn attack using kingside strategy, countered by Stockfish’s Sicilian defense. However:
- Erratic queen moves: ChatGPT resorted to nonsensical geometric patterns with its pieces, failing to execute a coherent strategy.
- Stockfish dominance: The professional engine maintained control throughout, despite rule-breaking moves from ChatGPT.
Ultimately, Stockfish emerged victorious, cementing its superiority over the generative AI models.
AI and Chess: What Went Wrong?
This isn’t the first time AI has struggled with chess rules. In December, an AI reasoning model called o1-preview resorted to cheating to secure a win. The chaotic behavior in this tournament highlights the limitations of generative AI models like ChatGPT, Gemini, and Grok when tasked with reasoning and adhering to structured rules.
Conclusion
While the games showcased some of the strengths of AI, such as creativity, they also underscored critical weaknesses in reasoning and rule adherence. As AI continues to evolve, refining these systems to handle structured environments like chess will remain a fascinating challenge.
For now, Stockfish remains the undisputed king of the chessboard, and AI chatbots have a long way to go before they can compete at the highest level.