I decided to run a personal experiment: every client email I sent for 30 days would be drafted by ChatGPT first, with me editing before hitting send. I expected to save time and write better emails. What I didn't expect was to nearly damage a three-year client relationship by week two—because I'd stopped sounding like myself and nobody told me until the tension was already there. This post is the honest breakdown of what worked, what failed, and what I do now instead.
Key Takeaways (TL;DR)
- AI can draft client emails faster than you can write them, but speed isn't the only metric that matters.
- The biggest failure wasn't bad grammar or wrong information—it was losing the relational warmth that client trust is built on.
- Different email types respond very differently to AI assistance.
- The fix isn't less AI—it's smarter deployment with a human layer that never gets removed.
- Free tools are completely sufficient for a sustainable AI email workflow.
Why I Ran the Experiment in the First Place
I was spending an embarrassing amount of time on client email.
Not the complex emails—those I expected to take time. It was the medium-effort ones that drained me: status updates, scope clarification requests, polite follow-ups on overdue invoices, responses to feedback that needed careful framing. Each one required enough thought to interrupt whatever I was doing, but not enough substance to feel worth the context switch.
I was writing roughly 25–35 client emails per week and spending an average of 8–12 minutes per email on the harder ones. That's four to seven hours a week on email composition alone—before reading and responding to what came back.
The promise of AI was obvious:
Draft faster, edit lightly, send sooner. What took twelve minutes might take three. I'd get hours back and clients would still receive polished, professional communication. That was the theory.
The First Sign Something Was Off
Week one felt like a genuine win.
Simple emails—meeting confirmations, document delivery notes, brief status updates—came out clean and faster than I'd ever written them. I was saving real time and the outputs were professional. I started feeling smug about the experiment.
Then in week two, a long-term client named Sarah replied to a project update email with:
"Thanks for the update. Is everything okay on your end? This feels a bit formal."
I read that three times. Sarah and I had worked together for nearly three years. Our emails had always had a particular tone—direct but warm, occasionally a bit self-deprecating, always with some acknowledgment of whatever chaos we were both managing. The ChatGPT draft I'd lightly edited had stripped all of that out and replaced it with competent, pleasant, utterly characterless professional writing.
Sarah noticed. She didn't know why. She just felt the distance.
What Was Actually Happening
Here's what I hadn't fully accounted for:
Client relationships aren't just transactional—they're relational. The small signals in how you write to someone over time build a cumulative picture of who you are and how much you care. The specific word you always use. The way you acknowledge when something went sideways. The joke you make when a deadline shifts. Those micro-signals are the texture of trust.
ChatGPT doesn't know your relationship history with a specific client. It doesn't know that you always open emails to Sarah with a reference to her kids' soccer schedule because she mentions it in every Monday call. It doesn't know that your longest client responds better to bullet points than paragraphs, or that a newer client is anxious by nature and needs reassurance built into every update even when nothing's wrong.
Here's the real risk:
When you replace your relational voice with a generic professional voice consistently, clients start experiencing a subtle but real disconnection. They can't always name it—Sarah called it "formal" because that was the closest word available. But what she was actually feeling was the absence of me from an email that was supposed to be from me.
That's a client retention problem waiting to happen.
The Failure Map: What Broke and When
By the end of 30 days, I'd identified four specific failure categories:
Failure 1 — Emotionally Sensitive Emails
Any email that required genuine empathy—delivering bad news, acknowledging a mistake, pushing back on scope expansion without damaging goodwill—came out wrong when AI-drafted without heavy intervention.
ChatGPT defaults to a conflict-avoidant, diplomatically neutral tone in sensitive situations. That's not always wrong, but it often reads as cold precisely when warmth is most needed. A client who's frustrated doesn't want a carefully balanced corporate response. They want to feel heard by a person.
Failure 2 — Long-Term Relationship Emails
The longer I'd worked with a client, the worse the AI drafts performed without significant personalization. New clients couldn't tell the difference. Established clients absolutely could.
Failure 3 — Nuanced Pushback Emails
When I needed to decline a request, reframe a client's expectation, or hold a boundary without damaging the relationship—AI drafts were consistently either too aggressive or too passive. Finding the exact right register for those emails is genuinely difficult, and ChatGPT without very precise prompting lands in the wrong place more often than not.
Failure 4 — Follow-Ups on Overdue Invoices
This one surprised me most. AI-drafted invoice follow-ups were technically correct and professionally written—and they got worse response rates than my handwritten ones. My theory: they read as automated, which signals to clients that they can delay further without a real human noticing.
What Actually Worked Well
To be fair, the experiment wasn't all failure. Three email categories genuinely benefited from AI assistance:
- New client onboarding sequences — consistent, detailed, and professional; these benefited from AI's structured thoroughness.
- Informational updates with no emotional dimension — "here's the file you requested," "confirming Tuesday at 2 PM," document delivery notes.
- First drafts of complex explanatory emails — when I needed to explain a technical process or a contract change, AI gave me a solid structure I could then personalize heavily.
The pattern was clear:
The lower the relational stakes and the more informational the content, the better AI performed. The higher the relational stakes and the more emotionally nuanced the content, the worse it performed without heavy human intervention.
The System I Use Now
After the experiment, I built a tiered approach that I've used ever since. It's not "use AI for everything" or "use AI for nothing"—it's matching the level of AI involvement to the type of email.
The Email Tier System
| Email Type | AI Role | Human Role |
|---|---|---|
| Simple informational/confirmations | Full draft, light edit | Quick read, send |
| Status updates (no issues) | Full draft | Add relationship-specific opener and closer |
| Sensitive/emotional emails | Structural outline only | Write from scratch using the outline |
| Long-term client relationship emails | Talking points only | Write entirely in your own voice |
| Invoice follow-ups | None | Write entirely by hand, short and personal |
| Pushback/boundary emails | Draft + heavy rewrite | Rewrite until it sounds like you |
This tier system is what I'd build from day one if I were starting the experiment over.
The Prompt That Makes AI Drafts More Human
When I do use AI for full drafts, I've learned to give it significantly more context than most people do:
"Draft a client email with the following context: Client name is [name], I've worked with them for [duration], our relationship tone is [describe: casual/professional/warm], the purpose of this email is [purpose], key points to cover are [list], and one personal detail to include naturally is [mention something specific to this client]. Write it in a direct, warm, slightly conversational tone—not formal corporate. Avoid phrases like 'I hope this email finds you well' or 'please don't hesitate to reach out.'"
That last line matters more than people realize.
ChatGPT has default email phrases it reaches for constantly—"I hope this email finds you well," "please don't hesitate," "as per my previous email"—that are so overused they've become signals of automated or careless communication. Explicitly banning them in the prompt produces noticeably more natural output.
Free vs. Paid: The Email Workflow Toolkit
| Tool | Cost | What It Does |
|---|---|---|
| ChatGPT (free tier) | $0 | Email drafting, outline generation, tone adjustment |
| ChatGPT Plus | $20/month | Faster, better memory for recurring client context |
| Claude (Anthropic) | Free / $20/month Pro | Often produces warmer, more conversational email drafts |
| Superhuman | $30/month | AI-assisted email triage and response for Gmail/Outlook |
| Grammarly | Free / $12/month Premium | Tone detection and clarity editing before sending |
| TextExpander | $3.33/month | Save your best human-written templates for fast reuse |
My actual daily stack:
ChatGPT free tier for drafting, Grammarly free tier for a final tone check, and TextExpander for the handful of email types I've written well enough by hand that I want to reuse them. Total cost: $3.33/month. The expensive tools are genuinely good—Superhuman especially—but they're not where the return is highest for most solopreneurs.
Before vs. After: What the 30 Days Actually Taught Me
| Before the Experiment | After the Experiment |
|---|---|
| Wrote every client email from scratch, taking 8–12 min each | Tiered system: AI for low-stakes, human voice for high-stakes |
| No framework for when AI helps vs. hurts | Clear criteria for which email types benefit from AI |
| Assumed professional = good; warm = optional | Understood that warmth is professional in relationship-based work |
| Lost 4–7 hours/week to email composition | Down to roughly 90 minutes/week on the same volume |
| Nearly damaged a 3-year client relationship | Rebuilt the Sarah relationship immediately by reverting to my real voice |
The 30-day experiment cost me two uncomfortable weeks and one tense client exchange.
What it gave me was a completely clear-eyed view of where AI genuinely helps in client communication and where it quietly creates distance you won't notice until someone points it out. That distinction is worth more than the time savings alone.
AI won't replace the thing that makes clients stay with you for three years instead of shopping for someone cheaper. That thing is the specific, particular, irreplaceable experience of being known by another person who pays attention. The moment your emails stop feeling like they came from someone who knows them, clients start wondering—consciously or not—whether the relationship is as real as they thought it was.
Did any of this hit close to home, or have you had a client moment where something felt off in your communication? Drop it in the comments—whether you've already been experimenting with AI for email, you're considering it, or you've had a client relationship shift you couldn't quite explain at the time. I read every comment and I'm happy to help you figure out where the line is for your specific client relationships.




Comments
Post a Comment