I Asked 4 Different AI Tools the Same Question — Here's What Actually Happened

I Asked 4 Different AI Tools the Same Question — Here's What Actually Happened

Last Tuesday I needed help writing a cold email to a potential client. Not a template — an actual email for an actual person I'd researched. Instead of picking my usual tool, I decided to run an experiment. Same prompt, four different AI assistants, timed and compared. What I found surprised me, and not in the way you'd expect from those "AI showdown" videos.

The Setup: One Prompt, Four Contenders

I used ChatGPT (GPT-4), Claude 3.5 Sonnet, Google Gemini, and Perplexity. The prompt was specific: write a cold email to a marketing director at a mid-sized SaaS company, mention their recent product launch, keep it under 150 words, and focus on how my content services could help their launch content perform better.

I gave each tool the exact same context — the company name, what they do, the product they just launched, and what I wanted the email to accomplish. Then I hit send on all four within about two minutes of each other.

Here's the thing most comparison posts miss: raw output quality barely matters anymore. All four tools produced usable emails. Every single one. The differences came down to things that only show up when you're actually trying to use the output for something real.

ChatGPT gave me the most polished initial draft — smooth transitions, professional tone, exactly 143 words. Claude's version was slightly more conversational and included a specific reference to the product launch that felt more natural. Gemini produced something functional but generic. Perplexity surprised me by pulling in a recent news mention I hadn't included in my prompt.

The Real Difference Nobody Talks About

So if the outputs were all decent, what actually separated them? The revision process. And this is the kick — the insight that changed how I use these tools entirely.

When I asked each tool to revise the email to sound "less salesy," the responses diverged dramatically. ChatGPT made surface-level word swaps: "opportunity" became "idea," "use" became "use." The structure and approach stayed identical.

Claude did something different. It asked me a clarifying question: "When you say less salesy, do you mean you want to lead with curiosity instead of a value proposition?" That single question led to a completely different email structure that actually solved my problem.

Gemini's revision felt like it ran the same email through a "casual tone" filter. Perplexity essentially rewrote from scratch, which wasn't what I wanted.

My take: the best AI tool for any task isn't the one that produces the best first draft. It's the one that collaborates best on revisions. Most of my work happens in rounds two through five, not round one. I'd never noticed this until I ran them side by side.

What Each Tool Actually Excels At

After this experiment — and honestly, after dozens more like it over the past few months — I've landed on some opinions that might sound controversial.

ChatGPT remains my go-to for structured content with specific formatting requirements. It follows instructions precisely. When I need a bulleted list that's exactly five items, or a paragraph that hits a word count, it delivers consistently. But it rarely surprises me.

Claude has become my choice for anything requiring nuance or back-and-forth refinement. That clarifying question it asked? That's not a one-time thing. It consistently engages with the why behind requests, which means fewer wasted iterations.

Gemini works best when I need something integrated with Google's ecosystem — pulling from Docs, checking my calendar context, that kind of thing. The output quality alone doesn't justify switching, but the workflow integration sometimes does.

Perplexity I now use exclusively for research-heavy prompts where I want current information baked in. That news mention it found? Turns out it was relevant — the marketing director had been quoted in it. Including that reference made my email genuinely stand out.

The Workflow That Actually Saves Time

Here's what I do now: I start research tasks in Perplexity, move to Claude for drafting anything that needs revision, and finish in ChatGPT if I need precise formatting or a specific output structure. Sounds complicated, but it's become automatic.

The real time savings come from knowing which tool to open first. Before this experiment, I'd start everything in ChatGPT, get frustrated with revision loops, switch to Claude, lose my context, and waste twenty minutes. Now I pick the right tool upfront based on what the task actually requires.

One more thing that surprised me: the free tiers are often fine. I pay for ChatGPT Plus and Claude Pro, but honestly? For quick tasks like this email, the free versions would have worked. The paid versions shine on longer, more complex projects — not everyday stuff.

Anyway, that cold email? I ended up using Claude's revised version with Perplexity's news mention added in. Got a response within four hours. Sometimes the boring answer — use multiple tools for what they're each good at — is the right one.

Heads up: Some links in this post may be affiliate links. I only recommend tools I've personally tested. Opinions are entirely my own.

댓글

이 블로그의 인기 게시물

How to Use Claude AI to Organize Your Messy Inbox (Without Losing Your Mind)

The Perplexity AI Feature That Makes Google Feel Outdated

What Happens When You Give AI Tools a Real Deadline to Meet