Deep Dive Case Studies

Learning ThroughDocumented Experiments

In-depth explorations of prompt engineering techniques, AI limitations, and real-world applications. Each case study includes interactive examples, bilingual videos, and downloadable resources.

5 Case Studies
Bilingual Content (EN/FR)
Interactive Examples

What Are Case Studies?

Unlike traditional blog posts, these case studies are detailed experiments where I test prompt engineering techniques, document every step, analyze results critically, and share everything—including what didn't work.

Each case study may include the following and more:

  • Bilingual video walkthroughs (EN & FR)
  • Downloadable prompts and templates
  • Before/after interactive comparisons
  • Expandable deep-dive sections
  • Live demos and deployed applications
  • Detailed analysis and key findings

All Case Studies (5)

XML-Structured AI Agents: Optimizing for How Models Actually Work
Featured
ENFR
technicalprocessintermediate

XML-Structured AI Agents: Optimizing for How Models Actually Work

I discovered XML structuring could optimize my 7 Context Engineers for token efficiency, minimize attention degradation, and enforce priority hierarchies. Here's the redesign.

Anthropic recommends XML tags for prompt clarity. Before implementing the 7 Context Engineers for 777-1, I restructured their definitions from prose to hierarchical XML. Here's why this matters for token efficiency, attention management, and context engineering.

18 min read3
Visual QA on Autopilot: Building a Self-Correcting AI Pipeline
Featured
ENFR
technicalprocessintermediate

Visual QA on Autopilot: Building a Self-Correcting AI Pipeline

I built a pipeline where Claude screenshots apps, finds visual bugs, and fixes them itself.

Too lazy to screenshot every UI bug? I built a workflow where Claude sees and fixes its own visual mistakes. Score improved from 63 to 77 in the first phase - then things got interesting.

15 min read6
Meet the Team: 7 Custom Subagents Built from 129 Code Reviews
Featured
ENFR
technicalprocessintermediate

Meet the Team: 7 Custom Subagents Built from 129 Code Reviews

Real names. Real personalities. Real job descriptions. Download the markdown files and use them yourself.

I analyzed 129 code reviews and extracted the 7 most common issues. Then I turned each one into a subagent with a name, a personality, and a detailed job description. Here they are.

15 min read7
From Trained Instance to Portable Prompt: Extracting Nancy's Blog Refiner
Featured
ENFR
generalintermediate

From Trained Instance to Portable Prompt: Extracting Nancy's Blog Refiner

How a financial blogger captured 3 months of AI training and made it work across Claude, ChatGPT, Gemini, and Grok

I had the perfect blog refiner. Then I lost it. Here's how I extracted my trained AI's personality using prompt extraction and made it work everywhere.

12 min read4
Meta-Prompting in Action: What Happens When You Let Claude Code Redesign Your App
Featured
ENFR
technicaldesignadvanced

Meta-Prompting in Action: What Happens When You Let Claude Code Redesign Your App

A brutally honest experiment in trusting AI without human oversight

I asked Claude to write detailed prompts for Claude Code, then executed them without any testing or modifications. Here's what worked brilliantly—and what broke completely.

25 min read2