The courses are complementary rather than competitive, serving different market segments with aligned principles but different depth/breadth.
- Model vs Product Evaluation: Both make identical distinction using different terminology
- Benchmark Illusion: Aishwarya's concept = your "illusory benefit of generic metrics"
- Positioning: Both target practitioners building AI products
- Input-Expected-Actual framework maps to your reference-based/reference-free metrics
- Three evaluation methods (Human, Code, LLM Judge) = identical to your Chapter 5 hierarchy
- Calibration principle: Both emphasize LLM judges must be validated against human judgment
- Flywheel concept: Same continuous improvement cycle, identical terminology
- Guardrails definition: Both define as synchronous, real-time safety checks
- Lifecycle frameworks: Video 9's 7-step checklist complements your Analyze-Measure-Improve cycle
Your course addresses 4 of 5 misconceptions:
- ✅ Don't rely solely on benchmarks
- ✅ Evaluation is continuous/cross-functional
- ✅ Both simple metrics AND LLM judges have roles
- ✅ Context-specific evaluation required
⚠️ Gap: Offline evals vs A/B testing relationship not explicitly covered in your materials
- No contradictions: Phoenix implements your tool-agnostic principles
- Complementary: Shows "how to implement" what you teach as "why it matters"
- Risk: Students might conflate Phoenix features with fundamental methodology
- Aishwarya: Start with ~50 examples
- Your course: Start with ~100 traces
- Resolution: Both are rough heuristics; your course uses different sizes for different contexts (20-100 depending on task)
- Aishwarya: Accessible overview with practical frameworks (10-min videos)
- Your course: Comprehensive methodology with statistical rigor (150-page reader, 4 weeks)
- Aishwarya: Phoenix-centric hands-on tutorials
- Your course: Tool-agnostic principles, explicitly states "if new LLM comes out tomorrow, methodology still applies"
- Aishwarya: Doesn't cover open coding/axial coding depth
- Your course: Detailed grounded theory methodology (Chapter 3)
- Three Gulfs Model - Unique conceptual framework
- Open/Axial Coding - Rigorous error analysis methodology
- Statistical Validation - TPR/TNR, bootstrapping, confidence intervals
- Collaborative Practices - Inter-annotator agreement (IAA), alignment sessions
- Architecture-Specific Coverage - RAG (Chapter 7), multi-turn (Chapter 6), agentic systems (Chapter 8)
- Custom Tooling Emphasis - Chapter 10 on review interfaces
- Cost Optimization - Dedicated Chapter 11
- Lower price point - Free videos vs your $5,000 (major accessibility)
- Shorter commitment - 13 videos vs 4-week intensive
- Practical prioritization - 2x2 matrices for metric selection
- Phoenix integration - Hands-on platform tutorials
- OpenAI association - Kiriti Badam's credibility
- Problem-first branding - Clear frustrated buyer positioning
- Entry-level/introductory - "AI Evals for Everyone"
- Frustrated buyer segment - "stuck with noisy, expensive, useless evals"
- Lead generation funnel - Free → $2,500-$3,000 paid courses
- Phoenix ecosystem - Tool-specific implementation
- Professional/comprehensive - Deep technical training
- Serious practitioners - Moving beyond POCs to production
- Premium positioning - $5,000 signals depth and rigor
- Tool-agnostic - Methodology over platform features
Relationship: Sequential market segments rather than direct competition. Her free course could actually funnel students who want deeper methodology to yours.
- Add FAQ on dataset sizing - Address 50 vs 100 discrepancy
- Consider Phoenix comparison - Students will encounter it; explain how it implements your principles
- Add A/B testing section - Only gap where Aishwarya covers something you don't
- Reference complementary resources - Could acknowledge her free intro as supplementary
- Emphasize depth differentiators - Your statistical rigor, error analysis methodology, custom tooling
- Position as "advanced implementation" - For teams beyond basic tutorials
- Highlight 150-page reader - Major material differentiator
- Feature enterprise credentials - 2,000+ students, 35+ implementations
Her course validates demand for evals education and shares your core principles. No contradictory advice that would confuse students. The $2,500 price difference and Phoenix-specific focus creates clear market segmentation.
Zero contradictions, high complementarity, different market segments. Students who take both courses would receive reinforcing messages with Aishwarya providing accessible entry and you providing comprehensive mastery. Your course's depth, statistical rigor, and tool-agnostic methodology remain strong differentiators worth the premium pricing.
- Carnegie Mellon University - Language Technologies Institute (Class of 2019)
- Research affiliations with University of Michigan, MIT, and Oxford
- 1,002+ citations on Google Scholar
- 35+ research papers published at top-tier conferences (NeurIPS, AAAI, CVPR, EMNLP)
Primary research areas:
- Social media aggression detection - Trolling, cyberbullying, flaming, hate speech
- Fake news detection & fact verification - Created Factify5WQA and Factify 2 shared tasks
- Meme interpretation - Multimodal ML approaches
- Artificial Social Intelligence - How AI and social media merge to create misinformation risks
- Current: Founder & CEO, LevelUp Labs (AI advisory & implementation startup)
- Previous: Tech Lead & Forward-Deployed AI Engineer at AWS Generative AI Innovation Center
- Led and implemented 30+ AI solutions for AWS clients
- Clients: Hitachi Digital, Deloitte, multiple Fortune 500 companies
- Claim: "50+ enterprise implementations" across OpenAI, Google, Amazon, Databricks
- Taught professional courses at MIT and Oxford
- TEDx Jacksonville speaker (March 2025) - "Social media and the age of AI misinformation"
- 95,000+ LinkedIn followers (recognized as "one of the most prominent voices in enterprise AI")
Link: https://problem-first-ai.kit.com/84cdfacd2b
What it includes:
- 10-day email course with daily lessons
- 4 live sessions included
- Topics: AI agents, workflows vs agents, RAG, MCP (Model Context Protocol), multi-agent systems
- Purpose: Primary lead generation funnel for paid Maven courses
Link: https://github.com/aishwaryanr/awesome-generative-ai-guide
Performance:
- 24,200+ stars (massive engagement)
- 5,200+ forks
- Updated regularly
What it includes:
- Monthly GenAI paper summaries
- 60+ GenAI interview questions
- 30+ curated free GenAI courses
- Code notebooks for generative AI applications
What it includes:
- 90+ hand-curated FREE resources
- 5 structured learning paths (fundamentals to production)
- 60+ project ideas categorized by difficulty
- "Don't Build AI Products Like Traditional Software"
- "Why AI Agents Aren't Enough for Real-World Applications"
- "Build your AI Moat as Software Developer in 2026"
Published: January 2026
Purpose:
- Thought leadership positioning
- Establishes "evals aren't enough" narrative
- Creates demand for their "improvement flywheel" solution
- "Machine Learning: Teach by Doing" series - 37 videos
- 100,000+ views on the playlist
- Free AI Evals course with 13 videos
- Posts #genai content daily on LinkedIn
- Simplifies complex generative AI research into digestible insights
- Focus on providing free, valuable educational resources
- Resource sharing - GitHub repository updates, free courses
- Course promotions - Maven course launches
- Research insights - Latest AI papers
- Thought leadership - Industry trends, best practices
- Daily LinkedIn posts (95K followers) → Free value
- GitHub repository (24.2K stars) → Developer community
- O'Reilly article → Thought leadership
- TEDx talk → Credibility
- YouTube videos → 100K+ views
- Free 10-day email course → Email capture, 4 live sessions
- Year-End AI Resources Handbook → Gated resource
- Free resource bundles → Demonstrates expertise
- Lenny's Newsletter feature → Third-party validation
- $3,000 flagship course (Building Agentic AI) → 1,500+ students, 4.9/5 rating
- $2,500 evals course (Beyond Evals) → New launch
- LevelUp Labs consulting → Enterprise services
-
AI Evals For Engineers & PMs - Hamel Husain & Shreya Shankar
- Price: $5,000 | Next: March 16, 2026 | Rating: 4.8/5
-
AI Evals for PMs Certification - Marily Nika et al.
- Price: $999 | Next: March 2, 2026 | Rating: 4.6/5
-
AI Evals and Analytics Playbook - Stella Liu & Amy Chen
- Price: $2,250 | Next: February 21, 2026 | Rating: 5.0/5
-
AI Evals for Product Development - Shane Butler
- Price: $1,500 | Next: April 6, 2026
-
Beyond Evals: Designing Improvement Flywheels - Aishwarya Reganti & Kiriti Badam
- Price: $2,500 | Next: March 14, 2026
-
Building AI Applications for Data Scientists and Software Engineers - Hugo Bowne-Anderson & Stefan Krawczyk
- Price: $2,100 | Next: March 10, 2026 | Rating: 4.7/5
- Daily content machine - 95K LinkedIn followers, daily posts
- Free email course - Strong lead gen (10 days + 4 live sessions)
- GitHub authority - 24K stars demonstrates massive reach
- Lower pricing - $2,500 vs your $5,000
- OpenAI association - Kiriti's role carries weight
- Multiple lead magnets - More entry points
- O'Reilly thought leadership - Industry voice
- Lenny's Newsletter feature - Same platform as your course
- Enterprise client logos - F500 social proof
- Less comprehensive - 3 weeks vs your 4 weeks
- No course reviews yet - New course vs your 4.8/5
- Lighter on methodology - Less error analysis depth
- No 150-page reader equivalent