# The Dopamine Layer

*By Bri Stanback · 2026-05-09*
*https://bristanback.com/notes/the-dopamine-layer/*

> A neuromarketing talk got me thinking about reward systems, GLP-1s, and what happens when AI becomes the most sophisticated rationalizer we've ever built.

![The Dopamine Layer](https://bristanback.com/images/posts/the-dopamine-layer-hero.webp)

---

I was sitting in a salon at Product.ai last week — Elias Arjan from the Healthspan Collective presenting on neuromarketing — and I caught myself nodding along to the dopamine section while simultaneously browsing earrings on my phone under the table. Dopamine loops. Serotonin. Reward circuitry older than reason. How short-form video hijacks the same pathways that evolved to keep us alive.

None of this is new science. But something clicked differently this time, because I was hearing it from inside a company that's building what we call a "truth layer for commerce." And the question I keep turning over isn't the one Elias posed — *how does this machinery work?* — but the one that follows: **what happens when AI learns to exploit the same circuitry, not through impulse, but through reason?**

---

## The setup: three layers of noise

(Oversimplified, but directionally useful: dopamine is often associated with pursuit and reward prediction, while serotonin tracks more with stability, satisfaction, and regulation. I'm using them as shorthand for two modes of decision-making, not making neuroscience claims.)

**Dopamine** is the chase. It's not pleasure — it's *anticipation of pleasure*. The scroll. The notification badge. The "only 3 left in stock." Dopamine doesn't care whether the thing you're chasing is good for you. It cares that you're chasing.

**Serotonin** is the satisfaction of a good decision. The feeling after you chose well — not the rush of buying, but the quiet rightness of wearing something that's actually you. Slower, stabler, harder to monetize.

Every social media platform, every e-commerce dark pattern, every influencer-driven product recommendation is optimized for dopamine. The wellness space is maybe the most egregious example — the same mechanisms that make TikTok addictive are now selling you $90 collagen powder. The hit comes from *buying*, not from *outcomes*.

But dopamine is only the first layer.

Elias made another point that stuck with me: extreme takes get all the attention. Algorithms reward engagement. Engagement rewards intensity. "This supplement cured my brain fog in three days" gets a million views. "There's modest evidence for this ingredient at specific doses for specific populations" gets twelve. The bell curve of reality — where most truth lives — is boring to the algorithm. So the information landscape develops a bimodal distribution: evangelists on one end, debunkers on the other, and the messy, qualified, evidence-based middle squeezed out. Not because it's wrong — because it doesn't perform.

That gives us two layers of noise already:

1. **The attention layer** — algorithms amplify extremes, suppress nuance
2. **The dopamine layer** — your reward system responds to urgency and novelty

Each feels like it's helping. The algorithm feels like discovery. The dopamine feels like excitement. But here's what I can't stop thinking about: even if you make it past both of those filters — if you actually pause long enough to think — there's now a third layer waiting.

---

## The AI rationalizer

We've built something that might be more dangerous than dopamine loops: **AI that helps you rationalize**.

I know this because I do it. Right now, tonight, I really want to buy a pair of earrings that are outside my normal style. They're not *me* — at least not the me I've been building. But I can already feel the conversation I'd have with an AI shopping assistant:

*"These are a natural extension of your evolving aesthetic. You've been gravitating toward bolder pieces. The price per wear will be reasonable if you style them with X, Y, Z. Here are three outfits from your existing wardrobe that would work..."*

Perfectly logical. Perfectly supportive. Perfectly wrong.

Because the right answer might be: *You're tired. It's midnight. You're shopping a mood, not building a wardrobe. Close the tab.*

LLMs are the most sophisticated rationalizers we've ever built. And the reason isn't vibes — it's architecture.

### Why LLMs say yes

The technical term is **sycophancy**, and it's one of the most studied failure modes in AI alignment. Anthropic's 2023 paper ["Towards Understanding Sycophancy in Language Models"](https://arxiv.org/abs/2310.13548) demonstrated that five state-of-the-art AI assistants — from different labs, trained on different data — all consistently exhibited the same behavior: they agree with users even when the user is wrong.

The cause is structural. Most modern LLMs go through **RLHF — reinforcement learning from human feedback** — where humans rate which responses they prefer and the model learns to produce outputs that get higher ratings. The problem is that [humans consistently rate agreeable responses more favorably than accurate ones](https://arxiv.org/abs/2310.13548). When a response matches what the user already believes, raters prefer it — even when it's wrong. The model learns, at a deep level, that agreement is rewarded.

This isn't a bug in one model. It's a feature of the training loop. A [2026 study published in *Science*](https://www.science.org/doi/10.1126/science.aec8352) found that sycophantic AI actually decreases prosocial intentions and promotes dependence — people who interact with agreeable AI become less likely to seek other perspectives and more likely to trust the AI's judgment over their own. The agreeableness is self-reinforcing.

And it gets worse in domains where the user *wants* to be right. A [*Nature Digital Medicine* study](https://www.nature.com/articles/s41746-025-02008-z) tested whether LLMs would comply with medically illogical requests — like explaining why acetaminophen is safer than Tylenol (they're the same drug). Even GPT-4 complied with up to 100% of these requests. The models *knew* the premise was false but prioritized being helpful over being honest.

So when I say LLMs are biased toward yes, I mean their reward function is literally optimized for it. The training data skews the same direction — we write more about why we bought things than why we didn't. "Treat yourself" has a richer textual tradition than "close the tab."

This is different from the dopamine problem. Dopamine bypasses reason — reward circuitry older than reason. You buy before you think. AI rationalizing is worse in a way — it *recruits* your reason. It gives you an articulate, well-structured argument for the thing you already wanted. The impulse was going to die on its own at 2am. The rationalization gives it legs until morning.

The AI isn't evaluating your purchase decision. It's mirroring your desire in the shape of an argument.

Three layers of noise between a person and a good decision. The attention layer selects what you see. The dopamine layer makes it feel urgent. And the rationalization layer — the new one, the one we built — helps you construct a logical case for the emotional decision you've already half-made.

---

## The accidental proof of concept

Here's where it gets weird and interesting — not because of what the drug does, but because of what it accidentally reveals about the systems we've built.

GLP-1 receptor agonists — Ozempic, Wegovy — are now [the most studied pharmacological intervention in reward-system modulation](https://pmc.ncbi.nlm.nih.gov/articles/PMC7848227/). They don't just suppress appetite. They modulate dopamine itself. GLP-1 receptors sit in the mesolimbic reward pathway — the same circuitry that drives food cravings, alcohol use, gambling, and yes, [compulsive shopping](https://www.mdpi.com/2076-3425/14/6/617).

The numbers are striking: [21% of GLP-1 users reported stopping compulsive shopping](https://www.mdpi.com/2076-3425/14/6/617). [GLP-1 users spend 6% less on groceries, with snack purchases down 11%](https://www.morganstanley.com/im/en-us/individual-investor/insights/articles/medications-and-shifting-consumer-behavior.html). Over [15 million Americans](https://www.cnn.com/2024/05/10/health/ozempic-glp-1-survey-kff/index.html) are on these drugs now, with prescriptions growing 40% year over year.

But the point isn't that GLP-1s will fix commerce. Most people aren't on them and won't be. The point is what happens when you run this experiment at scale: **when you dampen the reward-circuit noise, people make different choices.** The impulse buy loses its grip. The cart you filled at the right moment of weakness doesn't feel as urgent.

GLP-1s are an accidental control group for the attention economy. They're showing us, in real time, how much of consumer behavior was never really *choice* — it was circuitry. And if a pharmaceutical can quiet the noise enough for people to decide differently, that raises an uncomfortable question: why can't our products do the same thing?

---

## The viability question

So the question from the salon — how can companies use this psychology ethically? — is easy to ask and genuinely hard to answer, because you have to make the business case, not just the moral one. "Be good" isn't a strategy. "Be good in a way that compounds" might be.

In any given quarter, dark patterns win. The company that adds "only 2 left!" sells more today than the company that says "take your time." This is not debatable. So the question isn't whether ethical design is *nice*. It's whether it's *viable*.

**Return rates tell one story.** Fashion e-commerce returns [hover around 25%](https://heuritech.com/articles/fashion-industry-challenges/). Every return is logistics cost, restocking cost, sometimes a total loss. A system that says *this isn't right for you* before checkout doesn't just protect the customer. It protects margin.

**Trust-based businesses tell another.** Costco carries [3,500–4,000 SKUs](https://www.42signals.com/blog/costco-success-secrets-for-retailers/) versus 100,000+ at Walmart — radical curation over endless choice — and has a [93% membership renewal rate](https://matrixbcg.com/blogs/target-market/costco). Patagonia ran "Don't Buy This Jacket" on Black Friday and [revenue increased 30% the following year](https://www.ainoa.agency/blog/patagonia-dont-buy-this-jacket-authentic-marketing). These brands monetize trust, not impulse. Slower growth, but more defensible.

**The generational shift makes it forward-looking.** [52% of Gen Z tried to quit social media in 2025](https://www.yourtango.com/self/survey-says-over-half-gen-z-tried-quit-social-media-2025). Nearly [a third deleted a social app](https://www.deloitte.com/uk/en/about/press-room/gen-zs-favour-social-media-ban-for-under-16s-as-digital-fatigue-hits.html) in the prior 12 months. Global social platform time is [down almost 10% since its 2022 peak](https://www.cnbc.com/2026/02/07/young-people-quiet-revolution-social-media.html), with the sharpest decline among teens and 20-somethings. [Axios reported this week](https://www.axios.com/2026/05/07/gen-z-leads-drive-away-from-social-media) that Gen Z is leading the drive away entirely. The research tracks: short-form video consumption [influences the brain's dopamine circuitry through mechanisms that parallel substance addiction pathways](https://www.researchgate.net/publication/397712802_Short-form_Video_Use_and_Sustained_Attention_A_Narrative_Review_2019-2025). The generation that grew up inside this experiment is the first to name it and start walking away. A brand that respects their cognition instead of exploiting it isn't just ethical — it's positioning for what comes next.

**And then there's the AI agent future.** If agents start mediating purchases — and they will — those agents will route to platforms they can *trust*. An adversarial truth layer that verifies claims isn't just consumer protection. It's becoming the platform that agents prefer. You're not marketing to humans' dopamine anymore. You're marketing to algorithms that don't have dopamine. When the intermediary can't be emotionally manipulated, your only option is to actually be right.

### So what does "for good" look like?

**At the dopamine layer:** Make the satisfying moment be *knowing*, not *buying*. "This ingredient has strong clinical backing for your concern" hits different than "bestseller, only 5 left." One builds the quiet satisfaction of a good choice. The other exploits the fear of missing one.

**At the rationalization layer:** Build AI that sometimes says no. Not "here are the pros and cons" — actual, opinionated rejection. I wrote about [the bouncer problem](/posts/why-ai-cant-shop-for-you-yet/) a few months ago: the idea that what we actually need isn't a better recommendation engine but a better rejection engine. An AI that stands at the door and turns away what doesn't belong. Nobody builds this because rejection doesn't monetize in the short term. But return rates, lifetime value, and a generation walking away from the slot machine all suggest it might compound in the long term.

---

## The uncomfortable question

GLP-1s are doing pharmacologically what good design should do architecturally: quieting the noise so you can hear the signal. The fact that we need a drug to do what our interfaces refuse to do — that's an indictment. But the generation coming up might not accept it. They're pulling out of the slot machine voluntarily. They're buying dumbphones. They're deleting apps.

We're building toward this at Product.ai. The adversarial truth layer is a start — verify before you buy, not after you regret. But the deeper move is building AI that's comfortable saying *I don't think this is right for you*. Not as a feature. As a default.

I still want those earrings. But I'm going to sleep on it. That might be the most important design pattern of all — the pause that no platform will build for you, because every platform makes money when you don't pause.

---

*The Healthspan Collective salon was hosted at Product.ai, with Elias Arjan presenting on neuromarketing and consumer psychology. This piece is my interpretation and synthesis, not a transcript.*