How it works · Living agents

Living agents

Your ICP saw three competitor launches this week. So did our agents.

Synthetic feedback has a quiet problem. The agents are bright, the prompts are well-engineered, the calibration is honest, and the check returns in 60 seconds. By the afternoon, a category leader cuts pricing, a competitor ships a launch on Hacker News, or your ICP's feed lights up with a viral switching post. The agents that produced the morning's check have not heard about any of this. Their world stopped at generation time.

That is the gap between feedback and reality, and for a long time it was tolerable: focus groups took weeks anyway, so a one-day staleness was rounding error. Synthetic was supposed to fix the latency problem. It can also widen the staleness one, unless the agents themselves keep reading what your buyers are reading.

The flaw in static synthetic research.

A prediction made today is wrong tomorrow if the world changed in between. That sentence reads as obvious; the industry treats it as inconvenient. A brand that surveyed as "safe" on Friday can be the subject of a viral controversy by Monday. A category that read as "growing" in spring can flatten by autumn after one quarter of recession headlines. A competitor that was "irrelevant" in February can buy a Super Bowl ad and climb into top-of-mind by Tuesday.

Static synthetic agents cannot account for any of this. They were created on a given Wednesday, with a given prior, and they will respond to your stimulus on that prior forever. The seams are well hidden. The agent will speak in the first person, in present tense, and produce a confident answer. The answer is just, without anyone telling you, slightly out of date.

In a category that moves slowly this is fine. Most categories that buyers care about do not move slowly. Pricing perceptions shift after one viral pricing comparison. Brand affinity moves after one ad cycle. Crisis sentiment moves in hours. If the synthetic stack you bought is built to be precise on stimulus and blind to context, you are paying for an answer that was correct last month.

What "living" actually means.

Each Prism agent receives a daily feed. The feed is small enough that a person could plausibly absorb it in a real day, and varied enough that it looks like a day rather than a research instrument: a dozen news headlines, a few synthetic social posts from people in their demographic, a couple of competitor ads, a snippet of peer conversation. The mix is matched to who the agent is, where they live, and what they have already been seeing this week.

Each item becomes a memory. The agent does not just glance at a headline and move on; the headline shifts a small set of internal sentiments, the way it would for a person. Brands mentioned positively gain a fraction of warmth. Brands at the centre of a negative cycle lose a fraction of trust. Topics that are repeated, the way news cycles repeat, compound their effect. Topics that fade from the feed decay back toward the agent's long-run baseline.

When you run a simulation, the agent answers from the world it has been living in. It does not narrate the news at you. It just responds the way a person whose week has looked like that week would respond. The mechanism is hidden. The effect is visible: predictions that move, gradually, with the market instead of trailing behind it.

The four sources that feed every agent.

Real headlines from where your buyers actually read

Pulled every 24 hours from Hacker News, Lobsters, the SaaS press (TechCrunch, The Information, Stratechery), Lenny's Newsletter, First Round Review, OpenView blog, and category-specific publications (DX Index, GitLab DevSecOps, Pragmatic Engineer). Filtered by cluster relevance, deduplicated, and embedded so your agents see what their real-life counterparts read. A simulated CTO gets the morning Hacker News front page; a Head of Growth gets the Lenny's roundup. We never repeat an item to an agent who already absorbed it.

Synthetic SaaS-Twitter and Slack chatter

Generated daily, anchored to real trending topics on X (FKA Twitter), in r/SaaS, r/devops, r/startups, on Indie Hackers, and inside the public Slack/Discord communities each cluster lives in. Modeled on the actual cadence and tone of those rooms, terse on dev Twitter, longer-form on Lenny's, link-heavy on HN. Every synthetic post is labelled as synthetic in our logs. We don't reproduce real users' words.

Peer conversations

What an agent's buying network is saying. Multi-turn synthetic exchanges between people in their cluster, modeled on patterns from real Slack DMs, founder dinners, and SaaS-circle group chats. Two co-founders comparing pricing on a call. A team Slack channel passing around a competitor's launch post. Peer talk is rarer per day than feed content, but it weighs more on tool-switching decisions, the way it does in actual life.

Real launch posts and pricing changes

Pulled from Product Hunt, IndieHackers launches, public ad libraries (LinkedIn, Reddit), and changelog feeds (companies' /changelog pages, GitHub Releases, RSS). Matched to clusters using the same rules each surface uses. Your agents see your competitor's actual launch posts, their pricing-page diffs, the changelog item that just dropped. When a launch hits the front page of HN, our daily refresh picks it up, and the next round of checks reflects it.

How exposure becomes opinion.

Each item the agent consumes shifts sentiment toward the topics and brands it mentions, by a small amount. A single headline does not move the dial visibly. A week of consistent coverage does. We weight bad news slightly more than good news, because real human attention does the same: negativity bias is one of the better-replicated findings in cognitive psychology, and any model that pretends otherwise stops looking like a real human inside a week.

Moods decay back toward baseline over time, because real humans do not stay in a mood forever. An agent rattled by a Tuesday news cycle is calmer by Friday, unless the cycle has continued. Topics that get repeatedly mentioned compound: an opinion the agent forms about a category in week one is reinforced or eroded by what they see in weeks two and three. Quiet weeks let the baseline reassert itself.

After 30 days, every agent has an opinion shape that reflects the world they have been living in, not the world they were born into.

Buyers, not personas

Seven shifts that make an AI buyer stop looking like a prompt.

Most AI-feedback tools model buyers as polished respondents: a job title, a personality prompt, a one-shot opinion. Prism agents go further. Seven layered systems, peer influence, stack drift, circadian mood, memory consolidation, engagement modes, switching cliffs, and quarter-end pressure, give simulated SaaS buyers the texture of actual people checking your landing page on a Tuesday night, not the polish of a sales-call transcript.

1. Peer influence across the buying network.

Each agent has a network of 6–12 followed peers within their cluster, co-founders, ex-colleagues, the four people on their team Slack who actually evaluate tools. Every reaction prompt includes a block summarising what their network has been saying this week: “your former CTO Maria switched off Linear this month”, “your peers are negative on per-seat pricing (avg sentiment –0.22)”, excerpts from notable peer reactions. Brand sentiments also drift passively toward the network at 5% weight per day, tipping points emerge the way they do on Hacker News, not the way they do in a focus group. SaaS buying is driven 5–10× more by network than by marketing copy; without it, AI feedback is just one model talking to itself.

2. Stack drift.

Once a month, every agent's revealed-preference tool stack is re-evaluated against the launches, peer reviews, and pricing changes they actually lived through. A simulated CTO who absorbed thirty days of Postgres-vs-MongoDB discourse doesn't just “say” they prefer one, they swap one tool in their stack. After six months an agent looks meaningfully different from the day they were generated, in the way a real engineering lead does. The longest-running criticism of synthetic feedback , “your agents change opinions but never their stack” , closes here.

3. Mood that knows what time it is.

An agent's mood varies with local hour of day, morning inbox triage, post-lunch dip, evening shipping window, late-night fatigue, and within a multi-check session, mood walks one step at a time instead of being re-hashed for each variant. The same pricing page shown to the same simulated SMB founder at 10am versus 11pm gets two different reactions, both honest. Re-running a check produces realistic variance (5–8% spread on identical inputs) without losing reproducibility, the walks are seeded per agent per stimulus.

4. Memory that consolidates while it sleeps.

Weekly, similar episodes from the past month get clustered and collapsed into one themed memory in the agent's voice, “during April I watched the AI infrastructure pricing wars with growing skepticism of per-token billing”, not a list of HN posts. This is what a real buyer actually remembers a month later. Retrieval stays fast at scale, and reaction quality on year-old clusters stops degrading the way it does in static AI agents.

5. Engagement modes: scroll, skim, evaluate.

Real buyers scroll past most of their feed, skim some, deeply read few. Prism agents now do the same: roughly 50% of items are scrolled past (no memory, but exposure still counts toward drift), 35% are skimmed (a one-line summary), 15% get a full evaluation read. The mix bends per behavioural flag, passive buyers scroll more, technical evaluators stop on benchmarks, PLG users always stop on free-tier announcements. This matches actual SaaS-buyer engagement curves, drops cost-per-agent by ~40%, and means the memory bank reflects what an agent would actually remember reading.

6. Switching cliffs and the budget gap.

After 14 consecutive days of strongly negative exposure to a vendor, bad outage, viral pricing-change post, public security incident, an agent promotes that vendor to a switch-away list and the prompt thereafter reads “you actively avoid this”, the loyal customer turning churn risk in one moment. Separately, when budget stress sustains (runway pressure flag, hiring freeze flag), the gap between an agent's stated tool preferences and what they'll actually approve widens measurably. Reactions in that state return two quotes: a public one that sounds principled and a private one about the line item. That's the “we love it but we can't justify it this quarter” pattern most AI feedback misses.

7. Quarter-end pressure as a behavioural input.

Daily, every cluster pulls a calendar signal, week of quarter, proximity to year-end, conference cycle, fundraising-window density, and turns it into three small behavioural shifts. Quarter-end weeks tilt CRO and Head of Growth agents toward “I need to lock revenue, not evaluate vendors” reactions. Conference weeks (SaaStr, RSA, AWS re:Invent) shift CTO and Head of Engineering attention sharply toward competitor announcements and away from cold inbound. Year-end quiet windows make annual-contract pitches land twice as hard. Same simulated buyer, same outbound email, on the Friday of Q4 close vs. the second week of January: two honestly different reactions.

None of these are gimmicks. Each one closes a specific way that AI agents leak their syntheticness when you watch them long enough. The cumulative effect is an agent population whose 30-day, 90-day, and one-year behaviour you can show to a SaaS founder and have them recognise the buyers they actually sell to.

What this changes for your decisions.

  1. / 01

    Pricing checks that move with the category.

    Run a pricing-page check this week. Prism's SaaS agents already incorporate this week's pricing posts, this week's competitor changes, this week's switching threads. Re-run the same pricing check in a month after a category leader cuts their entry tier, the agents have absorbed that change too, and the second prediction reflects how the SaaS pricing landscape has shifted underneath your same tier structure.

  2. / 02

    Launch timing.

    Test the same launch tweet in week 1 and week 4 of a Product Hunt cycle. Watch sentiment shift as the news cycle moves past whatever launched ahead of you. The same hook, against the same indie-hacker audience, lands differently depending on what they have been reading on X and HN that week. Pick the week, don't guess.

  3. / 03

    Competitor monitoring.

    Your agents see your competitor's actual launch posts, their pricing-page changes, their changelog entries. Their brand sentiment toward the competitor drifts week by week as exposure compounds. You watch the drift and you know how your category is moving without paying a research vendor. The first time the line bends, you find out before someone forwards you the Hacker News thread.

  4. / 04

    Longitudinal positioning.

    Most synthetic feedback is one-shot. Prism's agents persist. Run a check in March and again in June, and the same agents have lived through three months of category news, competitor launches, and SaaS-Twitter discourse in between. The June prediction is informed by everything that happened. The agents are older, in the only way agents can be: more exposed.

What we don't do.

We don't scrape Twitter or TikTok. We don't claim to know what an individual real person consumed. We don't pretend the synthetic posts our agents read are real posts by real people. What we do: we feed agents a curated mix of legally sourced real signals and disclosed synthetic content, designed to mirror the shape of what their demographic actually encounters. We publish our methodology. We publish our validation against real survey shifts over time.

Validating drift.

The public validation page now shows time-series accuracy: the same cluster validated quarterly against fresh real-survey data. If our living agents are working, accuracy holds steady or improves as the world shifts. If they were static, accuracy would decay between calibration cycles. The gap is the proof.

Every cluster we publish carries its time-series alongside its current number, so you can see for yourself how a Prism prediction has aged in the months since we made it. Read the latest pass at /validation.

Live sample

Three sample agents, anonymised, drawn from real Prism output. Hover to pause.

Agent · Mira
36 · Munich, DE · Hospital pharmacist
-0.12
Last 7 days · 22 items consumed
  1. 7 days ago
    • News · Süddeutsche ZeitungECB holds rates; consumer confidence flat in Bavaria.
    • Social · Instagram (synthetic)Friend posts about a delayed pharmacy refill, frustrated tone.
    • Ad · Meta Ad LibraryDM-drogerie Markt sustainability campaign, soft pastel.
  2. 6 days ago
    • News · TagesschauGeneric medication shortages reported in three Länder.
    • Peer · Hospital colleagues (synthetic)Two coworkers complain about a competing chain's online portal.
    • Social · Reddit (synthetic)r/Munich thread praises a local independent pharmacist.
  3. 5 days ago
    • News · Spiegel OnlineFederal health minister announces price-cap consultation.
    • Ad · Google AdsDocMorris discount banner, urgency framing.
    • Social · Instagram (synthetic)Wellness influencer endorses a competing OTC brand.
  4. 4 days ago
    • News · HandelsblattRetail chain quarterly results miss expectations.
    • Peer · Family group chat (synthetic)Sister mentions a doctor recommended a specific generic.
    • Ad · YouTubeRatiopharm trust-building spot, warm narrator.
    • Social · Threads (synthetic)Ongoing debate about pharmacy opening hours reform.
  5. 3 days ago
    • News · Bayerischer RundfunkLocal pharmacy strike suspended after talks.
    • Social · Instagram (synthetic)Coworker posts a positive review of a delivery service.
    • Ad · Meta Ad LibraryShop Apotheke promotional creative, price-led.
  6. Yesterday
    • News · FAZInvestigation: counterfeit OTC supply in EU online channels.
    • Peer · WhatsApp group (synthetic)Friend asks for a brand recommendation, several reply.
    • Ad · TV (synthetic)Generic competitor airs a celebrity-led spot.
  7. Today
    • News · ZEIT OnlineOp-ed: rebuilding trust in independent pharmacies.
    • Social · Reddit (synthetic)Thread on a viral packaging redesign she liked.
    • Ad · Google AdsSponsored result for a national chain, generic copy.
Brand sentiment · 7-day drift
  • Brand A · independent chain+0.19
  • Brand B · online pure-player-0.30
  • Brand C · supermarket pharmacy-0.13
Sample agent · based on real Prism output, anonymised

What we don't do.

We don't scrape real users. Synthetic posts are generated, labelled synthetic in our logs, and never pretend to come from real people.

We respect our news partners' attribution requirements. Every signal we pull keeps its source link. When the source is Actually Relevant, we also keep their curator framing, the relevance summary, the antifactors, because that framing shapes how a real reader would process the story.

We don't reproduce paid content as if it were earned, we don't strip attribution from journalism we ingest, and we don't silently train on customers' stimulus briefs. The agents are synthetic; the supply chain is honest.

See a living agent.

One full anonymised 30-day history. Every item, every sentiment shift, every brand drift, in one page.