Confidently Improve your AI Feature
Remove the guesswork. Get immediate clarity on what's holding your AI back, and exactly how to fix it.
Get in touch
Reliable AI Improvement
Here’s how I typically get my clients unstuck, and systematically help them unlock real progress
Diagnostic Sprint
Quickly uncover what's really blocking your progress.
Iterative Improvement Cycles
Rapidly solve your biggest blockers in clear, measurable cycles.
Scale & Secure
Confidently scale your evaluation methods, and secure long-term impact.
1 - Diagnostic Sprint
Before making any commitment, we'll start with a focused Diagnostic Sprint (90–120 min).
What happens during this Sprint
→ Live review of real examples from your system (successful and failed interactions).
→ Immediate identification and naming of your core failure modes.
→ Clear understanding of your current evaluation gaps (where and why your metrics fail).
What You'll Get Immediately
Instant clarity on exactly what's worth fixing first. No guessing, no wasted effort.
→ A simple, actionable failure mode catalog you can use right away.
Concrete next steps for immediate improvement.
2 - Iterative Improvement Cycle
We'll move into a structured, high-clarity improvement cycle lasting typically between 4 to 8 weeks (about 2–4 cycles).
Each week you'll make clear, measurable progress toward solving your top-priority blockers, giving you complete visibility into how your AI is improving.
A Typical Week Together
Weekly Sync (1h/week)
We explicitly review progress, identify immediate next actions, and solve blockers live.
Mid-cycle Check-ins (asynchronous)
Tactical support precisely when you need it, ensuring continuous momentum between syncs.
Direct Working Sessions (as needed)
Ad-hoc sessions for specific tasks like manual tagging, error analysis, or creating your first lightweight evaluations.
What You'll Get Immediately
Fast, visible, measurable progress on the issues blocking your LLM improvement.
Reliable evaluation methods customized to your system.
Full confidence in your ability to measure AI progress week after week.
3 - Scale & Secure
Once your improvement processes are reliably delivering results, we smoothly transition into scaling mode.
Our goal here isn't just sustaining your progress, but making your evaluation methods highly efficient, automated, and deeply integrated into your workflow, setting you up for long-term success.
How We Scale
→ Automate Robust Evaluations
Reliable and automated evals fully validated against your manual insights.
MLOps Setup (as needed)
Robust infrastructure to keep your AI improvements rapid and sustainable.
Internal Training & Upskilling (optional)
Equip your internal team with proven methodologies, ensuring continued clarity and improvement independently.
What You'll Get Immediately
→ Sustainable, scalable processes for continuously measuring and improving AI.
→ Dramatic reduction in manual effort without sacrificing clarity.
→ Long-term confidence that your AI quality will only continue to improve.
About Me
I’ve been building and shipping AI for over eight years.
Along the way I fell into every possible trap, eventually spending months optimising things that didn’t move the needle.
Over time, I learned how to cut through the noise. How to isolate what matters, learn from every iteration, and build systems that actually improve.
Now, when AI teams hit a wall, when nothing seems to improve, I help them break through it. Quickly, and with clarity.
Made with Gamma