AI Companies Want to Read Your Chatbot’s Thoughts—And That Might Include Yours AI Companies Want to Read Your Chatbot’s Thoughts—And That Might Include Yours

AI Companies Want to Read Your Chatbot’s Thoughts—And That Might Include Yours

In brief

  • More than 40 top AI researchers propose monitoring chatbots’ internal “chain of thought” to catch harmful intent before it becomes action.
  • Privacy experts warn that monitoring these AI thought processes could expose sensitive user data and create new risks of surveillance or misuse.
  • Researchers and critics alike agree that strict safeguards and transparency are needed to prevent this safety tool from becoming a privacy threat.

Forty of the world’s top AI researchers just published a paper arguing that companies need to start reading their AI systems’ thoughts. Not their outputs—their actual step-by-step reasoning process, the internal monologue that happens before ChatGPT or Claude gives you an answer.

The proposal, called Chain of Thought monitoring, aims to prevent misbehavior, even before the model comes up with an answer and can help companies to set up scores “in training and deployment decisions,” the researchers argue

But there’s a catch that should make anyone who’s ever typed a private question into ChatGPT nervous: If companies can monitor AI’s thoughts in deployment—when the AI is interacting with users—then they can monitor them for anything else too.

When safety becomes surveillance

“The concern is justified,” Nic Addams, CEO at the commercial hacking startup 0rcus, told Decrypt. “A raw CoT often includes verbatim user secrets because the model ‘thinks’ in the same tokens it ingests.”

Everything you type into an AI passes through its Chain of Thought. Health concerns, financial troubles, confessions—all of it could be logged and analyzed if CoT monitoring is not properly controlled.

“History sides with the skeptics,” Addams warned. “Telecom metadata after 9/11 and ISP traffic logs after the 1996 Telecom Act were both introduced ‘for security’ and later repurposed for commercial analytics and subpoenas. The same gravity will pull on CoT archives unless retention is cryptographically enforced and access is legally constrained.”

Career Nomad CEO Patrice Williams-Lindo is also cautious about the risks of this approach.

“We’ve seen this playbook before. Remember how social media started with ‘connect your friends’ and turned into a surveillance economy? Same potential here,” she told Decrypt.

She predicts a “consent theater” future in which “companies pretend to honor privacy, but bury CoT surveillance in 40-page terms.”

“Without global guardrails, CoT logs will be used for everything from ad targeting to ’employee risk profiling’ in enterprise tools. Watch for this especially in HR tech and productivity AI.”

The technical reality makes this especially concerning. LLMs are only capable of sophisticated, multi-step reasoning when they use CoT. As AI gets more powerful, monitoring becomes both more necessary and more invasive.

Tej Kalianda, a design leader at Google, is not against the proposition, but emphasizes the importance of transparency so users can feel comfortable knowing what the AI does.

“Users don’t need full model internals, but they need to know from the AI chatbot, ‘Here’s why you’re seeing this,’ or ‘Here’s what I can’t say anymore,'” she told Decrypt. “Good design can make the black box feel more like a window.”

She added: “In traditional search engines, such as Google Search, users can see the source of each result. They can click through, verify the site’s credibility, and make their own decision. That transparency gives users a sense of agency and confidence. With AI chatbots, that context often disappears.”

Is there a safe way forward?

In the name of safety, companies may let users opt out of giving their data for training, but those conditions may not necessarily apply to the model’s Chain of Thought—that is an AI output, not controlled by the user—and AI models usually reproduce the information users give to them in order to do proper reasoning.

So, is there a solution to increase safety without compromising privacy?

Addams proposed safeguards: “Mitigations: in-memory traces with zero-day retention, deterministic hashing of PII before storage, user-side redaction, and differential-privacy noise on any aggregate analytics.”

But Williams-Lindo remains skeptical. “We need AI that is accountable, not performative—and that means transparency by design, not surveillance by default.”

For users, right now, this is not a problem—but it can be if not implemented properly. The same technology that could prevent AI disasters might also turn every chatbot conversation into a logged, analyzed, and potentially monetized data point.

As Addams warned, watch for “a breach exposing raw CoTs, a public benchmark showing >90% evasion despite monitoring, or new EU or California statutes that classify CoT as protected personal data.”

The researchers call for safeguards like data minimization, transparency about logging, and prompt deletion of non-flagged data. But implementing these would require trusting the same companies that control the monitoring.

But as these systems become more capable, who will watch their watchers when they can both read our thoughts?

Generally Intelligent Newsletter

A weekly AI journey narrated by Gen, a generative AI model.



Source link

Jose Antonio Lanz

https://decrypt.co/330647/ai-companies-want-read-your-chatbots-thoughts-might-include-yours

2025-07-17 17:27:03

bitcoin
Bitcoin (BTC) $ 111,173.00 1.29%
ethereum
Ethereum (ETH) $ 3,812.74 0.47%
tether
Tether (USDT) $ 1.00 0.02%
bnb
BNB (BNB) $ 1,126.27 0.93%
xrp
XRP (XRP) $ 2.36 2.27%
solana
Solana (SOL) $ 178.82 4.00%
usd-coin
USDC (USDC) $ 0.999899 0.01%
staked-ether
Lido Staked Ether (STETH) $ 3,809.69 0.34%
tron
TRON (TRX) $ 0.31485 0.54%
dogecoin
Dogecoin (DOGE) $ 0.186675 4.02%
cardano
Cardano (ADA) $ 0.639193 1.43%
wrapped-steth
Wrapped stETH (WSTETH) $ 4,634.02 0.34%
wrapped-bitcoin
Wrapped Bitcoin (WBTC) $ 111,179.00 0.92%
wrapped-beacon-eth
Wrapped Beacon ETH (WBETH) $ 4,091.33 1.33%
figure-heloc
Figure Heloc (FIGR_HELOC) $ 1.00 3.39%
ethena-usde
Ethena USDe (USDE) $ 1.00 0.03%
chainlink
Chainlink (LINK) $ 17.24 1.54%
stellar
Stellar (XLM) $ 0.321038 1.12%
bitcoin-cash
Bitcoin Cash (BCH) $ 515.67 3.23%
wrapped-eeth
Wrapped eETH (WEETH) $ 4,110.30 0.30%
hyperliquid
Hyperliquid (HYPE) $ 37.44 2.66%
sui
Sui (SUI) $ 2.51 8.12%
avalanche-2
Avalanche (AVAX) $ 21.24 3.11%
weth
WETH (WETH) $ 3,811.06 0.39%
leo-token
LEO Token (LEO) $ 9.66 0.36%
binance-bridged-usdt-bnb-smart-chain
Binance Bridged USDT (BNB Smart Chain) (BSC-USD) $ 1.00 0.00%
usds
USDS (USDS) $ 0.999538 0.10%
coinbase-wrapped-btc
Coinbase Wrapped BTC (CBBTC) $ 111,187.00 1.22%
usdt0
USDT0 (USDT0) $ 1.00 0.03%
litecoin
Litecoin (LTC) $ 94.20 6.22%
hedera-hashgraph
Hedera (HBAR) $ 0.169508 0.42%
whitebit
WhiteBIT Coin (WBT) $ 41.75 0.87%
shiba-inu
Shiba Inu (SHIB) $ 0.00001 1.93%
ethena-staked-usde
Ethena Staked USDe (SUSDE) $ 1.20 0.01%
monero
Monero (XMR) $ 299.44 3.52%
crypto-com-chain
Cronos (CRO) $ 0.156186 1.00%
the-open-network
Toncoin (TON) $ 2.10 0.28%
mantle
Mantle (MNT) $ 1.62 8.63%
dai
Dai (DAI) $ 0.99991 0.12%
polkadot
Polkadot (DOT) $ 3.01 3.59%
zcash
Zcash (ZEC) $ 267.69 22.93%
memecore
MemeCore (M) $ 2.20 0.41%
okb
OKB (OKB) $ 172.02 4.88%
uniswap
Uniswap (UNI) $ 6.01 2.91%
aave
Aave (AAVE) $ 231.03 0.73%
bitget-token
Bitget Token (BGB) $ 4.67 5.45%
world-liberty-financial
World Liberty Financial (WLFI) $ 0.119462 9.22%
bittensor
Bittensor (TAO) $ 299.16 2.03%
near
NEAR Protocol (NEAR) $ 2.28 5.57%
blackrock-usd-institutional-digital-liquidity-fund
BlackRock USD Institutional Digital Liquidity Fund (BUIDL) $ 1.00 0.00%