San Francisco, CA – In a move lauded by some and questioned by others, OpenAI has announced the public release of the chain of thought (CoT) reasoning process for its o3-mini model. This allows both free and paid users to observe the model’s internal thought process as it arrives at an answer. The move is seen by many as a step towards greater transparency in AI, but some experts are raising concerns about the authenticity and utility of the released data.
The announcement, initially reported by Chinese tech media outlet [量子位 – Quantum Bit], sparked immediate reaction online, with some users humorously thanking DeepSeek, a competitor in the AI space, for potentially prompting OpenAI’s decision.
The chain of thought is a technique that allows large language models to break down complex problems into smaller, more manageable steps, theoretically mimicking human reasoning. By revealing these steps, OpenAI aims to provide users with a better understanding of how the model arrives at its conclusions.
In one example provided by OpenAI, when asked Why isn’t today Friday?, the o3-mini model reportedly recognized the question as humorous and attempted to provide a witty response. It then supposedly used the Zeller’s congruence formula to confirm that the day was not Friday, even checking for leap year exceptions. Finally, the model playfully blamed the calendar for designating the day as Thursday, and encouraged the user to hang in there, tomorrow is closer to the weekend!
However, the release has not been without its critics. Some users have expressed skepticism about whether the publicly available CoT data is the raw, unfiltered output of the model. Concerns have been raised about the relatively slow display speed and discrepancies in character count compared to what would be expected from a truly original output.
The AI claimed to use Zeller’s formula but didn’t show the calculation process, which raised a lot of people’s vigilance, one user commented. Many users suspect that this is still a post-summary of the AI’s thinking process, rather than the original data.
Adding fuel to the fire, comments made by OpenAI’s Chief Product Officer, Kevin Weil, during a recent Responding to Everything event, hinted at the possibility of a curated CoT release. Weil stated that while releasing the full thought chain could be exploited by competitors for model distillation, OpenAI recognized the desire for such information, particularly among sophisticated users. He suggested that OpenAI would strive to find a balance.
Developer Mckay Wrigley echoed these concerns, stating, I worry that the summarized chain of thought is actually worse than no chain of thought. He argued that a true CoT acts as a prompt debugger, helping users guide and correct the model. A summarized version, he warned, could add confusion and introduce errors, making the model harder to debug.
In response to these concerns, an OpenAI spokesperson confirmed to [量子位 – Quantum Bit] that the publicly released CoT data undergoes post-processing. This process involves removing unsafe content, simplifying complex ideas, and improving the experience for non-English speaking users.
Despite the controversy, users are already experimenting with the newly available CoT data. One OpenAI researcher showcased their method of using o3-mini to play Tic-Tac-Toe, providing the model with a game state and asking for the optimal next move.
While the long-term impact of OpenAI’s decision remains to be seen, the release of the o3-mini’s chain of thought has undoubtedly sparked a crucial debate about the balance between transparency, security, and the practical utility of AI explanations. Whether this is a genuine window into the mind of an AI or a carefully curated performance remains a key question for the AI community to address.
References:
- 量子位 (Quantum Bit). (2024). OpenAI突然公开o3思维链!网友:让我们谢谢DeepSeek. Retrieved from [Insert Original Article URL Here – if available, otherwise remove this line]
Views: 0