How AI Workshops Transformed Our Engineering Team's Productivity by 70%
How AI Workshops Transformed Our Engineering Team's Productivity by 70%
Last year, I noticed a pattern at Radius Agent. Our engineers were spending hours every week on tasks that felt... mindless.
Writing boilerplate code. Drafting documentation. Converting data between formats. Writing test cases. Debugging with console.log scattered everywhere.
These tasks weren't hard. They were just tedious. And they were stealing time from the work that actually required human creativity and judgment.
That's when I started the AI workshops.
The Problem: Death by a Thousand Cuts
Let me paint you a picture of a typical day for our engineers before the workshops:
- Morning: Write boilerplate for a new API endpoint (30 min)
- Late Morning: Hunt through logs trying to understand a bug (45 min)
- Afternoon: Write documentation for a feature (1 hour)
- Late Afternoon: Convert JSON responses to TypeScript types manually (30 min)
- End of Day: Write unit tests for the code written earlier (1 hour)
That's nearly 4 hours of work that, while necessary, didn't require deep thinking. It was mechanical. Repetitive.
What if AI could handle 80% of this?
Starting Small: The First Workshop
I didn't want to dump a bunch of AI tools on the team and call it a day. That never works.
Instead, I started with a single workshop focused on one use case: using LLMs for code generation.
The Format
- Duration: 2 hours
- Size: 6-8 engineers
- Structure:
- 30 min: Demo of the tool and approach
- 60 min: Hands-on exercises with real problems from our codebase
- 30 min: Discussion and Q&A
The First Exercise
I gave everyone the same task: generate a CRUD API endpoint for a new entity in our system.
The traditional way: copy an existing endpoint, modify it, update tests, add documentation. About 45 minutes of work.
The AI-assisted way: describe what you need in natural language, let the LLM generate the boilerplate, review and refine. About 10 minutes.
The room got very quiet. Then the questions started flowing.
Building a Prompt Library
The next workshop focused on something that became our secret weapon: a shared prompt library.
We created a repository of prompts that anyone could use:
## Generate TypeScript Types from JSON
Prompt:
"Given this JSON response, generate TypeScript interfaces.
Use strict types, avoid 'any'. Add JSDoc comments for non-obvious fields.
[paste JSON here]"
Example:
Input: {"id": 1, "name": "John", "createdAt": "2024-01-01"}
Output: interface User { id: number; name: string; createdAt: string; }
We had prompts for:
- Converting between data formats
- Writing unit tests
- Generating documentation
- Debugging assistance
- Code review checklists
- API design suggestions
The key insight: prompts are shareable institutional knowledge. When one engineer figures out a good prompt, the whole team benefits.
The Transformation
Over six months of weekly workshops, here's what changed:
Metric: Time Spent on Boilerplate
Before: ~40% of engineering time After: ~12% of engineering time
Engineers weren't writing less code. They were writing less boring code.
Metric: Documentation Coverage
Before: Maybe 30% of functions had good documentation After: Over 80% had AI-assisted documentation
When writing docs takes 30 seconds instead of 10 minutes, people actually do it.
Metric: Test Coverage
Before: We aimed for 60% and often missed it After: Consistently hitting 80%+
AI-generated test cases aren't perfect, but they're a great starting point. Engineers review and refine instead of writing from scratch.
The 70% Productivity Improvement
When I say 70% improvement, I don't mean everyone suddenly worked 70% faster. I mean:
Engineers spent 70% more time on complex, creative, high-value work.
The kind of work that requires human judgment. Architectural decisions. Debugging complex distributed system issues. Thinking about user experience. Mentoring juniors.
The stuff that actually moves the needle.
What We Learned Along the Way
Lesson 1: AI is a Tool, Not a Replacement
The engineers who resisted AI were often worried about being replaced. We addressed this head-on:
"AI won't replace engineers. Engineers who use AI will replace engineers who don't."
The goal was never to reduce headcount. It was to multiply each engineer's impact.
Lesson 2: Quality Review is Non-Negotiable
AI-generated code needs review. Period.
We established a simple rule: AI can generate, humans must validate. Every AI-assisted contribution went through the same code review process as manual code.
This actually improved code quality because reviewers were fresher - they hadn't just spent an hour writing the code themselves.
Lesson 3: Start with Pain Points
We didn't try to AI-ify everything. We asked engineers: "What's the most tedious part of your day?" Then we focused workshops on those specific pain points.
For some teams, it was documentation. For others, it was test writing. For our data team, it was format conversion.
Targeted solutions beat generic "use AI for everything" advice.
Lesson 4: Make It Social
The workshops weren't just training sessions. They became a space for engineers to share discoveries:
- "I found this prompt that generates perfect error messages"
- "Did you know you can feed the AI your linting rules?"
- "Check out how I use AI for code review prep"
This organic knowledge sharing was as valuable as the formal instruction.
The Prompts That Made the Biggest Impact
Here are a few prompts from our library that became team favorites:
The Debug Assistant
I'm debugging an issue where [describe symptoms].
Here's the relevant code: [paste code]
Here's the error: [paste error]
Here's what I've tried: [list attempts]
Help me understand what might be going wrong and suggest next steps.
The Test Generator
Generate unit tests for this function using Jest.
Cover: happy path, edge cases, error handling.
Use describe/it blocks. Mock external dependencies.
[paste function]
The Documentation Writer
Write JSDoc documentation for this function.
Include: description, @param for each parameter, @returns, @throws if applicable.
Be concise but complete.
[paste function]
Looking Forward
We're now exploring more advanced uses:
- AI-assisted code review - Pre-screening PRs for common issues
- Automated refactoring suggestions - Identifying code that could be cleaner
- Intelligent monitoring - Using LLMs to analyze error patterns
The workshops continue, but now engineers bring their own discoveries to share. The culture has shifted from "AI is scary" to "AI is a superpower."
For Teams Considering This
If you're thinking about starting AI workshops at your company:
- Start small - One use case, one tool, one workshop
- Focus on pain points - Ask your team what's tedious
- Build shared resources - Prompt libraries, templates, examples
- Make it hands-on - Theory is boring; doing is learning
- Celebrate wins - Share success stories to build momentum
The goal isn't to turn every engineer into an AI expert. It's to give them tools that make the boring parts of their job disappear, so they can focus on the parts that made them want to be engineers in the first place.
Interested in running AI workshops for your team? I'm happy to share more details about our format and learnings. Reach out on LinkedIn.