When OpenAI announced earlier this week that it would begin testing advertisements within ChatGPT, it marked a key shift in how AI tools make money. The company promised users could opt out of personalized ads and would see clear labels distinguishing promotions from regular responses. However, widespread concerns over the use of ads in AI content has many worried. Senator Ed Markey sent a letter this week to OpenAI CEO Sam Altman and other tech leaders expressing concerns over the use of ads with AI. Ads in AI chatbots might turn helpful tools into sneaky marketing tricks that take advantage of deepening emotional bonds and personal relationships people are building with their AI assistants.
Markey’s letter lists specific worries beyond basic ads. He notes chatbots designed to sound human are becoming friends for millions, especially young people. Ads popping up in these chats could blur the line between help and sales.
“This raises significant concerns for consumer protection, privacy, and the safety of young users,” Markey writes. He sent the letter to CEOs at OpenAI, Anthropic, Google, Meta, Microsoft, Snap, and xAI. He’s hoping for answers on how these chatbots could refrain from inserting ads during serious talks or using private chats for targeted ads.
The Federal Trade Commission warned kids and teens are especially easy to trick with ads hidden in regular content in its 2023 report on “stealth advertising”. Markey points to this report to support his claims as to why AI chatbot ads are dangerous. Unlike ads on websites, chatbot ads could look like natural parts of a conversation. Users might not be easily able to spot the difference between real help and sales pitches.
The Senator’s Warning: When AI Becomes a Sales Tool
Markey’s main worry is about the emotional ties users form with AI chatbots. A 2025 study by the American Psychological Association found many teens use chatbots for friendship and support. Companies putting ads into these close chats risk exploiting that trust. As Markey writes, “companies could use this emotional link to push products, making ads seem like a friend’s suggestion.”
This isn’t just a worry. Meta recently said it will use AI chatbots to make ads more personal across its apps. Google also told ad clients Gemini ads might come in 2026. OpenAI said ads won’t change responses or show near sensitive topics. But big questions remain. Markey asks if companies will use private chats for commercial profiling. If someone talks about mental health in ChatGPT, could that info target ads elsewhere?
Privacy risks are serious. Chatbots are gathering detailed personal info as users share about health, relationships, and family. Using that data for ads breaks trust. Unlike normal data collection, users often feel they’re chatting with a friend or an advisor, not a data grabber. That emotional connection makes people share more than they would otherwise.
Business Pressure vs. Ethical Limits
OpenAI and other AI companies’ arguments are that they need to find additional ways to generate revenue to continue to support the very expensive, data center-heavy AI applications. OpenAI says ads help keep AI free for everyone.
OpenAI CEO Sam Altman had previously expressed reservations about introducing ads, labeling it a “last resort.” In a 2024 talk at Harvard, he described the combination of ads and AI as “uniquely unsettling”. But pressures from financial commitments and the necessity to explore new revenue streams have motivated the company to change its mind. This flip-flop already made people question AI platform vendors’ transparency.
The challenge is balancing profit and ethics. Meta and Google have used user data for ads for years. But chatbots are different. Their friendly chats make it hard to see when help turns into sales. Regulators and legislators see this as the right time to push back before ads become fully entrenched in AI systems.
Markey’s letter asks some key questions. Will companies use sensitive chat topics for ads? Will they use kids’ data for targeting? How will users spot ads? He asks how companies pick users for ad tests, if they’ll allow opt-outs and how they’ll block ads during health talks.
Markey’s letter is the latest and highest-profile pushback against ads in AI chatbots. Lawmakers increasingly see the reach of AI as a serious problem. For AI companies, this is a chance to lead. They can build strong safeguards before rules force them to. OpenAI’s point about ads supporting free access makes financial sense, but does it make sense in the context of trust and responsible AI use?
The way forward needs clear labels, strict data rules and real user choices. Companies should see Markey’s letter not as a nuisance or regulatory hurdle, but as an opportunity to provide specific plans and avoid vague promises. For example, companies can commit to never using health chats for ads and clearly mark any commercial ties.
This isn’t just about ChatGPT and its recent announcement on ads in its product. It’s about AI’s future. As chatbots join daily life, the line between helper and sales tool grows thinner. Markey’s letter isn’t against making money from AI. His questions are a key moment for the industry. They show AI can be both profitable and ethical. As he writes, AI companies “have a responsibility to ensure that AI chatbots do not become another digital ecosystem structured to covertly manipulate users.”







