The Super Bowl commercials showed Claude, Anthropic’s AI assistant, poking fun at ChatGPT by depicting it inserting awkward and intrusive ads into user conversations. This portrayal struck a nerve with Altman, who insisted that OpenAI would never implement ads in such a disruptive manner.
“We’re not stupid enough to run ads like that,” Altman declared in his response. He clarified that OpenAI’s actual ad strategy involves placing clearly labeled, non-influential ads at the bottom of conversations. These ads would be based on the chat context but wouldn’t interrupt the user experience.
The heated reaction highlighted the growing rivalry between the two AI companies. Anthropic, founded by former OpenAI employees concerned about AI safety, has positioned itself as a more responsible alternative to ChatGPT. Their marketing has consistently emphasized ethical AI development since the company’s launch.
Altman’s criticism went beyond defending OpenAI’s ad strategy. He accused Anthropic of being “authoritarian” and wanting to control how people use AI. He claimed that Claude Code is blocked from companies like OpenAI and that Anthropic dictates what users can do with their technology.
The controversy underscores fundamental differences between the companies’ approaches. OpenAI prioritizes widespread access, using ad revenue to fund free ChatGPT for millions of users. Meanwhile, Anthropic focuses on safety and responsibility as its main selling points. Anthropic’s four ads were humorously targeting OpenAI after the announcement of ads in ChatGPT’s free tier.
Despite the mockery, ChatGPT remains the most popular chatbot by a significant margin. OpenAI has been transparent about testing ads, explaining in a blog post how sponsored products would be recommended based on conversation context, but only when relevant and without disrupting the user experience. While finding the commercials somewhat amusing, Altman expressed clear disappointment over what he considered a misleading representation of OpenAI’s advertising philosophy.