Updated · 8 min read
Generative AI for lifecycle content: where it earns its place and where it embarrasses you
Picture a marketer at 4pm on a Friday with a subject line block. They open the AI panel inside their email tool, hit generate, get five options, pick one, ship the campaign, go home. That is generative AI doing honest work — a copy partner that compresses the blank-page step from fifteen minutes to thirty seconds. Now picture the same tool writing every subject line for every send to every user, with nobody reading the output. Same technology, different mode, completely different risk. One of those programs ships clean lifecycle copy on a Friday afternoon. Another one ships hallucinated prices, off-brand jokes, and tone-deaf condolences to grieving customers. This guide draws the line between them — and tells you where on the spectrum your program should actually live.

By Justin Williames
Founder, Orbit · 10+ years in lifecycle marketing
Three speeds you can run generative AI at — risk goes up at each one
Risk in generative content sits in the gap between a model's output and a brand voice you've spent years building. Every mode that closes the human review gap also widens the brand drift one.
Generative AI — models that produce new text from a prompt, like ChatGPT or Claude — can be plugged into your email program at three different speeds. Same underlying technology, very different risk profiles. Most ESPs (email service providers — Braze, Iterable, Klaviyo, the platforms that actually send your emails) sell features at every speed without making the trade-off explicit. So here it is, plainly.
Speed 1 — Acceleration. Model produces multiple variants for a human to choose from: subject lines, preview text, body copy variations. A marketer picks one (or refines into a final). Risk profile: low, because a human is still the final filter. Productivity gain: real — 3–5x on copy ideation. This is what BrazeAI (Braze's built-in copy assistant, formerly called Sage AI) is best at.
Speed 2 — Approved variant rotation. Model produces a candidate set; a human approves a subset; from there, your ESP rotates approved variants automatically — usually via a multi-armed bandit (an algorithm that gradually sends more traffic to whichever variant is performing best, like a slot machine that learns which arm pays out). Unapproved copy never ships, but no human reviews each individual send. Risk profile: medium. Best for high-volume programs where per-send review is impossible. Iterable and Klaviyo support patterns like this.
Speed 3 — Per-user real-time generation. Model writes copy at the moment of send, personalised to the individual user, with no human in the loop. Most futuristic mode, also the one with the most failure modes — hallucinations (a model confidently inventing facts that aren't true), brand voice drift, off-topic tangents, inappropriate content for the user's situation. Risk profile: high. Worth doing only with strong guardrails and high-frequency human audit. Few programs need this; even fewer should run it without infrastructure most teams don't have.
Most programs should live in Speed 1 with selective use of Speed 2 for high-volume templates. Genuinely Speed-3 use cases exist — large e-commerce catalogues with tens of thousands of SKUs, dynamic news digests, programmatic content recommendations — but for nearly everything else, Speed 3 is the wrong answer.
Where it actually pulls its weight
Five places generative AI does honest work in a lifecycle program — meaning it saves real time without introducing risk a human can't catch.
Subject line ideation. Highest-ROI use case in the program, full stop. Model generates variants in seconds, marketer picks two or three to test, test produces lift. Compresses the "think of three different angles" step from 15 minutes to 30 seconds. BrazeAI, Klaviyo's subject line generator, and most chat LLMs (large language models — the technology behind ChatGPT, Claude, Gemini) accessed via your warehouse all do this well.
Preheader and preview text. Preheader is the short line of text that appears after the subject in an inbox preview — a second hook before someone clicks. Most teams under-think it. Generative AI is good at producing preheaders that complement (rather than repeat) the subject line. Our preheader guide walks through the standard, with examples.
Body copy variants for A/B testing. When a program needs three or four genuinely different body angles to test, generative AI can produce them faster than a copywriter writing from scratch. Copywriter still edits and approves. All the compression is in the blank-page step.
Translation and localisation drafting. First-pass translation of approved English copy into other languages, for a localisation team to refine. Faster than full retranslation, lower-quality than native local copy. Right answer for cost-constrained localisation; wrong answer for brand voice in your primary market.
Programmatic product / content descriptions across a large catalogue. Ecommerce catalogue of 50,000 SKUs needs description copy for each. Hand-writing is impossible; generative AI is genuinely useful here. Pattern that works: per-SKU prompts grounded in product attributes, structured templates, human spot-checks for quality. Risk to watch is hallucinated specifications — model invents features the product doesn't have. Always ground generation in verified product data.
Four ways it goes wrong — all preventable
Every brand-damaging AI email story you've seen lands in one of these four buckets. Nothing surprising; each one is avoidable if you know what to watch for.
1. Brand voice drift. Models have a default voice — chatty, slightly enthusiastic, fond of American idioms, partial to certain rhetorical patterns. Without strong prompting and review, output drifts toward this default voice over weeks. First email reads like the brand. By the hundredth, it sounds like every other brand using the same model. Mitigation: a tightly defined brand voice spec (a document describing how your brand sounds) included in every prompt, plus a sample-based audit weekly. Our brand voice guide walks through the discipline.
2. Hallucinated facts. A hallucination is when a model confidently states something untrue — invents a feature, a price, a sale end date, a discount terms line. Especially common when prompts are loose ("write a promotional email about our new product") and grounding data — verified facts you give the model to work from — is incomplete. Mitigation: ground every generation in verified facts from your product database, never let the model invent specifics, run a fact-check pass on any output that contains numbers, claims, or commitments.
3. Tone-deaf in context. Models don't know your customer just had a refund denied, that your brand is in the middle of a PR crisis, or that this recipient unsubscribed once and resubscribed during a different mood. Generated copy in the abstract may be fine; in the actual user's context it can range from awkward to insulting. Mitigation: keep generative content out of high-context moments — post-complaint, sensitive contexts, escalations.
4. Compliance and regulatory exposure. Financial services, healthcare, alcohol, gambling — categories where every claim is regulated and an unsupervised model can manufacture liability. Asymmetric risk: small upside on faster copy, large downside if a regulated claim slips through. Mitigation: human review on every send in regulated categories, period. If volume makes that impossible, your program shouldn't be using generative content.
BrazeAI: where it earns its keep, where it embarrasses you
BrazeAI — a generative layer baked into Braze's composer, formerly Sage AI — currently handles subject lines, body copy variations, and image generation. Useful in Speed 1 (acceleration with human review). Less useful when treated as autonomous content generation.
Pattern that works: use BrazeAI to generate 3–5 variants for whichever element you're testing (typically subject line). Pick two to A/B test. Read the result. Use the winner as input for the next test. Model accelerates the variation step in a discipline (subject line testing) your team already runs.
Pattern that fails: ship BrazeAI's first suggestion unedited. Output is fluent but generic. Model has no memory of your previous campaigns, your voice norms, or what worked last week. Without human filtering, your program ends up sounding like a competent but anonymous brand.
Same pattern applies to Iterable Copy Assist, Klaviyo's subject line generator, and any LLM accessed directly via API. Feature names differ; discipline is identical.
The boring layer that stops the program embarrassing you
Governance is the part most teams skip. Style guide, review process, audit trail (a record of which prompt produced which copy on which date, so you can debug after the fact) — that's the trio. Without them, programs accumulate small problems until a big one ships. Four pieces, working together, do the job.
A documented brand voice spec your team uses in every prompt. Not the marketing department's aspirational voice doc — an operator-usable version with examples of what to write and what to never write.
A review SLA per speed (SLA = service-level agreement, the rule for how fast something gets reviewed). At Speed 1 (acceleration), every output reviewed before send. At Speed 2 (approved variant rotation), each variant reviewed before approval and rotation runs automatic from there. At Speed 3 (real-time generation), sample audit at high frequency, with rollback authority — meaning someone can pull the plug fast if drift appears.
A do-not-generate list. Sensitive contexts, regulated claims, customer service escalations, post-incident messaging. A list of programs and templates where generative content is explicitly off the table. Our AI Personalisation skill covers the canonical do-not-generate list.
An audit trail. Which version of which prompt produced which copy that shipped to which audience on which date. Without this, debugging "why did we send that?" becomes forensic archaeology. Most ESP-native generative features ship with weak audit trails, so building a thin layer of your own is worth the effort.
None of this is exciting work. All of it separates generative content as a productivity tool from generative content as a slow-motion brand incident.
Read to the end
Scroll to the bottom of the guide — we'll tick it on your reading path automatically.
Frequently asked questions
- Is it safe to ship BrazeAI subject lines without human review?
- No. BrazeAI (formerly Sage AI) is an acceleration tool, not autonomous copy generation. Fluent-but-generic output drifts off-brand quickly without human filtering. Pattern that works: generate 3–5 variants, pick the strongest two for A/B testing, ship the winners. Pattern that fails: ship the first suggestion. Thirty seconds saved per send is not worth brand drift accumulating across hundreds of sends.
- How do I prompt generative AI for consistent brand voice?
- Include explicit voice context in every prompt: tone (one sentence), three things your brand never says (specific phrases or constructions), an audience description, and an example of past copy that exemplifies the voice. Treat this as a reusable system prompt, not something written from scratch each time. Most ESP-native generative features now support saved prompt templates — use them.
- What's the right mix of generative AI and human-written copy?
- For most lifecycle programs: human-written for evergreen flagship templates (welcome, key transactional, brand moments); generative-accelerated (Speed 1) for tests, variants, and high-volume but lower-stakes templates; rarely generative-autonomous (Speed 3) and only with strong guardrails. The mix evolves over time based on which templates you've validated against holdouts (a random group you don't message, used as a control to measure incremental impact).
- Should I use AI to generate copy in regulated industries?
- Yes for ideation and acceleration with human review on every send (Speed 1). No for autonomous variant rotation (Speed 2) or per-user generation (Speed 3) in regulated categories — financial services, healthcare, alcohol, gambling, and similar. Asymmetric risk profile (small productivity gain, large compliance exposure) doesn't justify the autonomy. Programs that need to scale in regulated categories are better served by a structured template library than by generative content.
- How do I detect brand voice drift over time?
- Sample-based audit at a regular cadence — 10 sent emails per week, reviewed against your brand voice spec by someone with authority to flag drift. Compare voice metrics (sentence length, exclamation density, idiom use) month over month. Most programs catch drift only when it's loud — by which point months of generated copy have already shipped. The audit cadence is unsexy and load-bearing.
This guide is backed by an Orbit skill
Related guides
Browse allPredictive models in lifecycle: churn, propensity, and recommendations without the magic
Predictive models in lifecycle are mostly three things: churn risk, conversion propensity, and product recommendations. Each one earns or loses its place based on whether its score actually changes a decision. Here's the operator view of what's worth deploying, what to expect from ESP-native suites, and when to build your own.
The SMS playbook from the operator's seat
SMS is the highest-engagement and highest-risk channel in the lifecycle stack. Here's the compliance architecture, the copy discipline, and the frequency rules that keep SMS from destroying the goodwill it's uniquely positioned to create.
AI personalisation at scale: the architecture that actually works
Every ESP now sells an AI personalisation layer. Most teams turn it on and quietly notice the lift is smaller than the sales deck promised. The model isn't the problem — the plumbing underneath is. Here's the data, content and activation stack that decides whether AI personalisation moves revenue or just moves dashboards.
Building a personal chief-of-staff AI on Claude Routines
A real chief of staff used to mean a salary line on an exec's budget. Anthropic's Routines feature — Claude running on a schedule with access to your work tools — pulls the job inside reach of one operator. This is the architecture: morning brief, hourly interactive layer, midday drift check, evening debrief with end-of-day reconciliation, Sunday weekly review. Plus the draft-react protocol that lets the assistant act without auto-sending, calendar work blocks that double as the task tracker, and memory files the system writes into. Brain in a GitHub repo, runtime in claude.ai, no servers.
Liquid for lifecycle marketers — the complete Braze reference
Every personalised field in every Braze message runs through Liquid. Get it right and personalisation quietly improves every send. Get it wrong and 50,000 people see 'Hi {{${first_name}}}'. This reference covers the syntax and the production habits that stop that happening.
Brand voice in lifecycle: how to sound like you, not the generic SaaS CRM voice
Lifecycle emails — the automated sequences that go out on signup, after a purchase, when a cart is abandoned — drift toward a polished, generic SaaS voice because it's the default in every template library. Here's how to actually write in your brand's voice across an entire program.
Found this useful? Share it with your team.
Use this in Claude
Run this methodology inside your Claude sessions.
Orbit turns every guide on this site into an executable Claude skill — 63 lifecycle methodologies, 91 MCP tools, native Braze integration. Free for everyone.