πŸ¦‰
AI Engineering Handbook
About
  • Hey
  • Tech Stack
  • ML to AI Shift
  • Agentic Systems
  • Anti Patterns
  • AI Engineer
  • Roles Reimagined
  • Dream Team
  • System
    • Architecture
    • Start Simple
    • Enhance Context
    • Implement Guardrails
    • Add Routing and Gateway
    • Optimise with Caching
    • Enable Agent Patterns
Powered by GitBook
On this page

Anti Patterns

PreviousAgentic SystemsNextAI Engineer

Last updated 3 days ago

Common failures that waste time and money. Extracted from by Chip Huyen

⛔️ Using GenAI Unnecessarily

Teams adopt generative AI for status, not solutions. Companies want bragging rights: "we use AI."

Example: Energy startup built GenAI for optimisation. A simple algorithm checking cheap hours works better.

⛔️ Premature Complexity

Teams rush to complex solutions first. "I must climb that mountain" mentality. Some common mistakes

  • Vector databases before basics

  • Agent frameworks too early

  • Fine-tuning before prompting

⛔️ Bad Product or Bad AI? or Both?

πŸ’‘ Technical excellence β‰  product success.

Meeting summaries example:

  • ❌ Summary length (3 vs 5 sentences)

  • βœ… Action items from meetings

LinkedIn job recommendations example:

  • ❌ Match accuracy scores

  • βœ… Skill gap identification

Tax chatbot example:

  • ❌ Users don't know what to ask

  • βœ… Suggest relevant FAQs

⛔️ Not Understanding Users

User experience means understanding human behaviour.

πŸ’‘ Users want help, not judgment.

⛔️ Overcomplicating RAG Systems

Teams rush to vector databases.

  • Start here: Keyword search, clean data, smart chunking

  • Data prep matters: Add context, include titles and keywords

⛔️ Using Agent Frameworks too early

Agent frameworks create hidden complexity.

  • ❌ Framework bugs become your bugs

  • ❌ Starter prompts contain typos

  • ❌ Performance fluctuates with framework changes

  • ❌ Hard to diagnose root causes

  1. Build core logic with direct API calls

  2. Measure performance with 100+ test cases

  3. Identify specific bottlenecks frameworks solve

  4. Only then evaluate if framework complexity is worth it

⛔️ Lack of Human Evaluation

AI judges help but have limits. Add human evaluation.

Limitations:

  • Models change

  • Hard to track progress

  • Teams lose touch with users

  • Tools score differently

⛔️ Not Localising Failures

Multi-step AI needs failure isolation.

  • ❌ Resume extraction fails. Which step? PDF to text? Text to data?

  • βœ… Debug each step separately.

⛔️ Overindexing on Early Success

Weekend demos don't predict launch timelines. Demo = 20% of total work.

Production needs:

  • Error handling

  • User testing

  • Performance tuning

  • Infrastructure

  • Security

  • Monitoring

  • ...and more

⛔️ Giving Up Too Early

Teams quit after first failures. Find root causes. Fix them. Common causes:

  • Bad product design

  • No user research

  • Poor evaluation

πŸ€“Issues

beat frameworks.

Teams on 0.0.* versions

One team built POC with LangChain,

cause unpredictable failures

by 25% on average

Frameworks add and harder to manage

due to complex state handling

When frameworks fail, you debug

Framework updates modify , breaking performance

rather than simplifies

AI Engineering Anti-patterns
Direct API calls for simple tasks
couldn't update without breaking changes
couldn't update without major code changes
Syntax errors in framework scripts
Direct API calls outperform LangChain
layers that make systems slower
Multi-step task failures
framework code instead of your logic
starter prompts without notice
Excessive abstraction obscures
CrewAI execution loops
CrewAI production challenges
Why developers abandon LangChain