This experiment perfectly exposes the fundamental flaws of so-called "artificial intelligence." Large language models, such as ChatGPT, are essentially glorified autocomplete systems that generate plausible-sounding text without any genuine understanding, actual reasoning, or game-playing intelligence in specialized domains. The fact that a 1977 chess program can outperform ChatGPT suggests these systems are merely expensive marketing hype designed to inflate tech company valuations.
While this chess experiment is amusing, it doesn't diminish ChatGPT's remarkable capabilities in language processing, creative writing, and problem-solving across countless domains. Chess requires specialized algorithms and deep game tree analysis, which isn't what large language models were designed for. The comparison is like criticizing a race car for being bad at flying — different tools serve different purposes in the AI toolkit.