[Day 3] Comparing AI Tools for Handwriting Transcription
[Day 3] Comparing AI Tools for Handwriting Transcription
I put Claude, ChatGPT, and Gemini to the test with the same handwritten notes. One misread a key number, one nailed the formatting with icons, and one gave me three different options I didn’t ask for. Here’s what happened—and what it taught me about choosing the right AI tool for the job.
One of my favourite uses of AI is transcribing handwritten notes. I think better when I write by hand. There’s something about pen on paper that helps ideas flow more freely than typing. The accuracy of AI transcription has become so good now that it’s one of the first things I recommend to anyone looking for a practical, immediate use case for these tools.
This is especially powerful in team workshops. After a brainstorming session covered in sticky notes and whiteboard scribbles, someone used to be stuck with the tedious job of typing everything up. Now? Snap a photo, upload to an AI tool, and you’ve got structured notes in seconds. It’s a clear example of how AI can be a genuine productivity booster.
Today I tested this workflow across three different AI assistants and the results revealed something important about how each tool interprets human input differently.
Today’s Experiment
Goal: Transcribe my handwritten 2026 goals from paper to digital, formatted for Notion (my second brain).
The Task: I photographed a page of handwritten goals covering four life areas: Mind, Body, Finances, Relationships, and asked Claude, ChatGPT, and Gemini to:
Transcribe the handwritten notes
Reformat the output for Notion with clear structure
Why This Matters: This is exactly the kind of analog-to-digital workflow that knowledge workers encounter constantly. Meeting notes, brainstorm sessions, whiteboard captures—all need to go from messy handwriting to structured digital formats. The question is: which AI handles this best?
Process
Step 1: Capture the Source Material
I started with a single page of handwritten notes in my notebook listing my 2026 goals across four categories:
Mind: Daily morning practice for intention-setting
Body: Mobility exercises, gym 2x per week, food tracking with Cal AI
Finances: Automated savings and investments monthly
Relationships: Reach out to two people daily
A sample of my handwriting that borders on being illegible even to me at times.
Step 2: Test Each AI Tool
I used the same image and same initial prompt across all three tools:
Initial Prompt: “Transcribe”
Follow-up Prompt: “Revise the transcript for Notion”
This kept the experiment consistent while testing how each AI interprets minimal instructions.
Step 3: Compare the Results
I evaluated each output on three criteria:
Accuracy: Did it correctly read my handwriting?
Formatting: How well did it structure for Notion?
Usefulness: Could I copy-paste directly into my workspace?
Outputs
🟠 Claude (Opus 4.5)
Transcription Accuracy: Made one notable error—read “2x per week” for gym as “4x per week.” The rest was accurate.
Notion Formatting: Clean structure with headers and bullet points. Intelligently separated Finances into its own category (my handwriting grouped it under Body). No icons.
Sample Output:
First prompt using Claude
Follow-up prompt using Claude
Verdict: Good structure, helpful category separation, but the accuracy error on a specific number is a reminder to always verify AI outputs.
🟢 ChatGPT
Transcription Accuracy: Correctly read “2x per week” for gym. All other details accurate.
Notion Formatting: Excellent. Added emoji icons for each category (🧠 Mind, 💪 Body, 💰 Finances, 🤝 Relationships), used consistent structure, and included bold formatting for key terms like “Cal AI” and “2× per week.”
Sample Output:
First prompt using ChatGPT
Follow-up prompt using ChatGPT
Verdict: The “Goldilocks” option—accurate transcription, visually appealing with icons, and immediately copy-paste ready for Notion.
🔵 Gemini
Transcription Accuracy: Correct on all details, including the 2x per week. Interestingly, it interpreted my “$” symbol for finance as “Finance” category—showing deeper contextual understanding.
Notion Formatting: Offered three different options:
Clean Dashboard View - Simple goals page with checkboxes
Daily Habit Tracker Setup - Full database schema with property types
Verdict: Most comprehensive and accurate, but felt like overkill for my immediate need. The multiple options required more decision-making energy. Best for users who want to build a full Notion system from scratch.
What I Learned Today
1. Accuracy Varies on Specifics
Claude misread “2x” as “4x”—a small error with real consequences if I’d built my fitness routine around it. Always verify numbers, dates, and specific details in AI transcriptions.
2. Same Prompt, Different Interpretations
“Revise for Notion” meant different things to each AI:
Gemini: Multiple architectural options and database schemas
3. More Options ≠ Better Output
Gemini’s three options were impressive but added cognitive load. Sometimes you just want the AI to make a reasonable choice and give you something usable. ChatGPT’s single, well-formatted output was immediately actionable.
4. Context Interpretation Shows Intelligence
Gemini recognizing “$” as “Finance” demonstrated contextual reasoning beyond literal transcription. This kind of interpretation adds value but only if it aligns with your intent.
5. The Human Still Curates
Each tool produced something useful, but I still needed to:
Verify accuracy (caught Claude’s 4x error)
Choose which format fits my system
Decide whether I wanted icons or not
Adapt the output to my actual Notion workspace
This is the human-AI collaboration in action: AI handles the heavy lifting of transcription and formatting; humans provide quality control and context-specific decisions.
Try It Yourself
Time Required: 5-10 minutes per tool
What You’ll Need:
A photo of handwritten notes
Access to Claude, ChatGPT, and/or Gemini
A destination app (Notion, Obsidian, Apple Notes, etc.)
The Experiment:
Take a clear photo of your handwritten notes
Upload to your AI tool of choice
Start with a simple prompt: “Transcribe”
Follow up with: “Format this for [your app]”
Compare the output to your original—note any errors or interpretations
Copy into your destination and see what needs adjustment
Questions to Consider:
Which errors would matter most in your context?
Do you prefer icons/visual formatting or clean minimalism?
How much do you want the AI to interpret vs. literally transcribe?
Every year since ChatGPT launched, I've found myself asking a completely different question about generative AI—not the same question with better answers, but an entirely new question that only became possible because the previous one had been resolved. In 2022, I wondered what AI actually knew (and learned the hard way about hallucinated salary data in job descriptions). By 2025, I was producing interactive training decks and Excel templates without writing a single line of code. Now, at the start of 2026, I'm asking how to build automated workflows that free me up to focus on what humans do best. These five questions trace the extraordinary pace of change in just five years—and reveal where the real opportunities lie for knowledge workers ready to ask the next question
I've tried using Claude, ChatGPT, and Gemini to create presentations and the results were always underwhelming. The content was decent, but the formatting? A manual nightmare. So when I tested Gamma.app with a genuinely complex use case (employee offboarding compliance across six Asia-Pacific countries), I wanted to see if a specialized tool could actually deliver what general-purpose AI assistants can't: polished, presentation-ready output in minutes, not hours. Here's exactly what happened, step by step, including the limitations they don't tell you about upfront.
What happens when you give three AI tools the same complex virtual training scheduling problem and one of them walks straight into a cultural landmine? I ran an experiment with Claude, ChatGPT, and Gemini: identical prompts asking each to recommend the best approach for scheduling a 100-person virtual training across six APAC countries during Lunar New Year season. One tool recommended a culturally tone-deaf date to save $2,500, and one framed the time zone constraint as a "Golden Window" that actually helped me think through the problem differently. This is what AI collaboration looks like in practice and why the human still needs to be in the loop.