AI Engineering Architecture and User Feedback

AI Engineering Architecture and User Feedback Study Guide

Overview

This chapter explores how to combine various AI engineering techniques into production-ready applications. It covers system architecture design and user feedback collection for continuous improvement through data flywheels.

AI Engineering Architecture

Core Principle

Start simple and gradually add complexity. The basic flow: Query → Model → Response, then progressively enhance based on needs.

Step 1: Enhance Context

Purpose: Give models necessary information to produce quality outputs

Components:

Implementation considerations:

Step 2: Implement Guardrails

Purpose: Mitigate risks and protect users/systems

Input Guardrails

Protect against:

Sensitive data detection:

Mitigation strategies:

Output Guardrails

Quality failures:

Security failures:

Handling strategies:

Trade-offs:

Step 3: Add Model Router and Gateway

Router Functions

Intent classification:

Implementation:

Gateway Benefits

Unified interface: Single point for multiple model APIs Access control: Centralized security and cost management Fallback policies: Handle rate limits and API failures Additional features: Load balancing, logging, analytics

Popular solutions: Portkey, MLflow, Wealthsimple, TrueFoundry, Kong, Cloudflare

Step 4: Reduce Latency with Caches

Exact Caching

Warning: Improper caching can cause data leaks between users

Semantic Caching

Step 5: Add Agent Patterns

Complex workflows: Loops, parallel execution, conditional branching Write actions: Email composition, order placement, bank transfers Risk considerations: Significantly increased system exposure

Monitoring and Observability

Key Performance Indicators

Metrics Categories

Format Failures

Quality Metrics

Safety Metrics

User Behavior Metrics

Performance Metrics

Logs and Traces

Comprehensive logging:

Trace requirements:

Drift Detection

Monitor changes in:

AI Pipeline Orchestration

Two-Step Process

  1. Component definition: Models, databases, tools, evaluation/monitoring systems
  2. Chaining: Function composition and data flow specification

Design Considerations

LangChain, LlamaIndex, Flowise, Langflow, Haystack

Evaluation Criteria

User Feedback Systems

Strategic Importance

Feedback Types

Explicit Feedback

Implicit Feedback

Natural Language Feedback Signals

Early Termination

Error Correction

User Edits

Complaints

Eight categories (from FITS dataset):

Sentiment Analysis

Conversational Action Feedback

Regeneration Signals

Conversation Organization

Conversation Metrics

Feedback Collection Design

Timing Strategies

Initial calibration: Optional preference setting Failure moments: Error reporting and recovery options Low confidence: Uncertainty-driven feedback requests Success moments: Optional positive feedback collection

Collection Best Practices

Design Examples

Feedback Limitations

Common Biases

Degenerate Feedback Loops

Mechanism: Predictions influence feedback, which influences next model iteration Examples:

Mitigation: Understand feedback limitations and potential biases before implementation

Key Takeaways

  1. Progressive complexity: Start simple, add components as needed
  2. System thinking: Problems often require multi-component solutions
  3. Observability first: Design for failure detection and debugging
  4. User feedback value: Critical for competitive advantage and improvement
  5. Bias awareness: Understand and design around feedback limitations
  6. Product-engineering convergence: AI engineering increasingly involves product considerations
  7. Safety considerations: Each component addition increases potential failure modes