KnowYourBasics
Get Started
🎓 Acharya Nagarjuna University – Distance Education

Diploma in Gen AI Engineering

Industry-focused diploma bridging academic excellence with real-world skills

1 Year Duration
2 Semesters
35 Credits
12 Courses
📥 Download Complete Syllabus (DOCX)
← Back to Distance Education Programs

Program Overview

Comprehensive training in gen ai engineering with industry-standard tools and practices

🎯 Program Objectives

  • Industry-focused learning experiences
  • Hands-on projects and real-world applications
  • Modern tools and technologies
  • Professional skill development
  • Career-ready graduates

📚 Learning Outcomes

  • Master core concepts and fundamentals
  • Apply advanced techniques and methodologies
  • Build production-ready applications
  • Work with industry-standard tools
  • Complete comprehensive capstone projects

🛠️ Technologies Covered

  • Modern programming languages and frameworks
  • Industry-standard tools and platforms
  • Cloud services and deployment
  • Best practices and methodologies
  • Production systems and MLOps

Course Structure

2 Semesters • 12 Courses • 35 Total Credits

S.No Course Code Course Title Credits
SEMESTER 1: Foundation & Fundamentals (16 Credits)
1 DAGE101 Introduction to Gen AI & LLMs 3
2 DAGE102 Introduction to Gen AI & LLMs Lab 2
3 DAGE103 Prompt Engineering & Model Fine-tuning 3
4 DAGE104 Prompt Engineering & Model Fine-tuning Lab 2
5 DAGE105 LLM APIs & Local Model Setup 3
6 DAGE106 Capstone Project 1 3
SEMESTER 2: Advanced & Production Systems (19 Credits)
7 DAGE201 Vector Databases & RAG Systems 3
8 DAGE202 Vector Databases & RAG Systems Lab 2
9 DAGE203 AI Agent Development with LangChain & CrewAI 3
10 DAGE204 AI Agent Development Lab 2
11 DAGE205 Production GenAI Systems & MCP 3
12 DAGE206 Capstone Project 2 6

Detailed Syllabus

Click on a course to jump to its detailed syllabus

DAGE101 DAGE102 DAGE103 DAGE104 DAGE105 DAGE106 DAGE201 DAGE202 DAGE203 DAGE204 DAGE205 DAGE206
DAGE101

Introduction to Gen AI & LLMs

📚 3 Credits 📅 Semester 1 📋 Prerequisite: Basic programming knowledge

Course Objectives

  1. To introduce Generative AI concepts and applications
  2. To understand Large Language Models (LLMs) and their architecture
  3. To learn about different LLM families and their characteristics
  4. To understand how LLMs work and their capabilities
  5. To explore LLM use cases and applications
I

INTRODUCTION TO GENERATIVE AI

What is Generative AI? – Definition and Concepts – Generative vs Discriminative Models – History and Evolution of Gen AI – Applications of Generative AI: Text Generation, Image Generation, Code Generation, Audio Generation. Large Language Models Overview – What are LLMs? – How LLMs Work: Transformer Architecture Basics – Training Process: Pre-training, Fine-tuning – Tokenization and Embeddings – Context Windows and Attention Mechanisms. LLM Capabilities and Limitations – What LLMs Can Do: Text Completion, Question Answering, Translation, Summarization – LLM Limitations: Hallucinations, Context Limits, Bias – When to Use LLMs – When Not to Use LLMs.

📝 Illustrative Problems

Identify Gen AI use cases; Compare different Gen AI applications; Understand LLM capabilities; Recognize LLM limitations; Choose appropriate use cases.

II

LLM MODEL FAMILIES

OpenAI Models – GPT-3.5, GPT-4 – Model Characteristics – Use Cases – API Access – Pricing and Limitations. Anthropic Models – Claude Models – Claude Characteristics – Use Cases – API Access – Comparison with GPT. Open Source Models – Llama Models (Meta) – Mistral Models – Model Characteristics – Local Deployment – Fine-tuning Capabilities. Other Model Families – Google Models: Gemini, PaLM – Cohere Models – Specialized Models: Code Models, Multimodal Models – Model Selection Criteria.

📝 Illustrative Problems

Compare different LLM models; Select appropriate model for use case; Understand model capabilities; Evaluate model trade-offs; Choose between proprietary and open-source models.

III

TRANSFORMER ARCHITECTURE BASICS

Transformer Architecture Overview – Attention Mechanism – Self-Attention – Multi-Head Attention – Positional Encoding – Encoder-Decoder Architecture. How LLMs Generate Text – Autoregressive Generation – Sampling Strategies: Greedy, Temperature, Top-k, Top-p – Token Prediction – Generation Parameters. Model Parameters and Scale – Parameter Counts – Model Sizes – Scaling Laws – Compute Requirements – Model Efficiency.

📝 Illustrative Problems

Understand attention mechanism; Explain text generation process; Adjust generation parameters; Compare model architectures; Understand scaling concepts.

IV

INTERACTING WITH LLMs

LLM APIs Overview – OpenAI API – Anthropic API – API Authentication – API Rate Limits – Cost Management. Basic API Usage – Making API Calls – Prompt Structure – Response Handling – Error Handling – Best Practices. Python Libraries for LLMs – OpenAI Python Library – Anthropic Python Library – LangChain Basics (Introduction) – Simple LLM Applications.

📝 Illustrative Problems

Make API calls to LLMs; Handle API responses; Build simple LLM applications; Manage API costs; Implement error handling.

V

LLM APPLICATIONS AND USE CASES

Text Generation Applications – Content Creation – Creative Writing – Code Generation – Documentation Generation – Use Case Examples. Question Answering Systems – Building Q&A Systems – Context Management – Answer Quality – Evaluation Metrics. Translation and Summarization – Machine Translation – Text Summarization – Extractive vs Abstractive Summarization – Implementation Approaches. Ethical Considerations – Bias in LLMs – Privacy Concerns – Misinformation – Responsible AI – Best Practices.

📝 Illustrative Problems

Build text generation application; Create Q&A system; Implement summarization; Address ethical concerns; Evaluate application quality.

📖 Textbooks

  1. 1. Jay Alammar, "The Illustrated Transformer", Blog Post Series
  2. 2. Tom Taulli, "Artificial Intelligence Basics", Apress, 2019
  3. 3. OpenAI Documentation and Research Papers
DAGE102

Introduction to Gen AI & LLMs Lab

📚 2 Credits 📅 Semester 1 📋 Prerequisite: DAGE101

Course Objectives

  1. To practice interacting with different LLMs
  2. To build basic GenAI applications
  3. To experiment with LLM capabilities
  4. To understand practical LLM usage

📖 Textbooks

  1. 1. OpenAI API Documentation
  2. 2. Anthropic API Documentation
DAGE103

Prompt Engineering & Model Fine-tuning

📚 3 Credits 📅 Semester 1 📋 Prerequisite: DAGE101

Course Objectives

  1. To master prompt engineering techniques
  2. To learn advanced prompting strategies
  3. To understand model fine-tuning concepts
  4. To implement few-shot learning and chain-of-thought prompting
  5. To fine-tune LLMs for specific tasks
I

PROMPT ENGINEERING FUNDAMENTALS

Introduction to Prompt Engineering – What is Prompt Engineering? – Why Prompt Engineering Matters – Elements of a Good Prompt – Prompt Structure: Instruction, Context, Examples, Output Format. Basic Prompting Techniques – Zero-Shot Prompting – One-Shot Prompting – Few-Shot Prompting – Prompt Templates – Prompt Variables – Iterative Prompt Refinement. Prompt Design Principles – Clarity and Specificity – Providing Context – Setting Constraints – Defining Output Format – Examples and Demonstrations – Common Mistakes.

📝 Illustrative Problems

Write effective zero-shot prompts; Design few-shot prompts; Create prompt templates; Refine prompts iteratively; Avoid common prompt mistakes.

II

ADVANCED PROMPTING TECHNIQUES

Chain-of-Thought Prompting – What is Chain-of-Thought? – Step-by-Step Reasoning – Implementing CoT – CoT for Complex Problems – CoT Variations. Role-Based Prompting – Assigning Roles to LLMs – System Prompts – Persona-Based Prompts – Multi-Agent Scenarios – Role Consistency. Prompt Chaining – Breaking Complex Tasks into Steps – Sequential Prompts – Conditional Prompting – Prompt Pipelines – Managing State Across Prompts.

📝 Illustrative Problems

Implement chain-of-thought prompting; Use role-based prompts; Create prompt chains; Build prompt pipelines; Manage complex prompts.

III

FEW-SHOT LEARNING AND IN-CONTEXT LEARNING

Few-Shot Learning Concepts – What is Few-Shot Learning? – In-Context Learning – Example Selection – Example Ordering – Few-Shot vs Fine-tuning. Implementing Few-Shot Learning – Selecting Good Examples – Formatting Examples – Providing Diverse Examples – Handling Edge Cases – Evaluating Few-Shot Performance. Advanced In-Context Learning – Dynamic Example Selection – Adaptive Few-Shot – Meta-Learning Concepts – Prompt Compression Techniques.

📝 Illustrative Problems

Implement few-shot learning; Select effective examples; Format examples properly; Evaluate few-shot performance; Optimize example selection.

IV

MODEL FINE-TUNING FUNDAMENTALS

Introduction to Fine-tuning – What is Fine-tuning? – When to Fine-tune vs Prompt – Fine-tuning Process Overview – Data Requirements – Fine-tuning vs Pre-training. Fine-tuning Methods – Full Fine-tuning – Parameter-Efficient Fine-tuning (PEFT) – LoRA (Low-Rank Adaptation) – QLoRA (Quantized LoRA) – Adapter Methods. Fine-tuning Workflow – Data Preparation – Data Formatting – Training Configuration – Hyperparameter Tuning – Evaluation and Validation.

📝 Illustrative Problems

Prepare data for fine-tuning; Choose fine-tuning method; Configure training parameters; Evaluate fine-tuned models; Compare fine-tuning approaches.

V

PRACTICAL FINE-TUNING AND OPTIMIZATION

Fine-tuning Open Source Models – Fine-tuning Llama Models – Fine-tuning Mistral Models – Using Hugging Face Transformers – Training Infrastructure – Cost Considerations. Fine-tuning Best Practices – Data Quality – Data Augmentation – Overfitting Prevention – Evaluation Metrics – Model Selection – Deployment Considerations. Prompt Optimization – A/B Testing Prompts – Measuring Prompt Performance – Cost Optimization – Latency Optimization – Quality vs Cost Trade-offs.

📝 Illustrative Problems

Fine-tune open-source model; Implement best practices; Optimize prompts; Measure performance; Balance quality and cost.

📖 Textbooks

  1. 1. OpenAI, "GPT Best Practices", OpenAI Documentation
  2. 2. Anthropic, "Prompt Engineering Guide", Anthropic Documentation
  3. 3. Hugging Face, "Fine-tuning Language Models", Hugging Face Course
DAGE104

Prompt Engineering & Model Fine-tuning Lab

📚 2 Credits 📅 Semester 1 📋 Prerequisite: DAGE103

Course Objectives

  1. To practice prompt engineering techniques
  2. To implement advanced prompting strategies
  3. To fine-tune models for specific tasks
  4. To optimize prompt performance

📖 Textbooks

  1. 1. OpenAI, "GPT Best Practices", OpenAI Documentation
  2. 2. Hugging Face, "Fine-tuning Language Models", Hugging Face Course
DAGE105

LLM APIs & Local Model Setup

📚 3 Credits 📅 Semester 1 📋 Prerequisite: DAGE101

Course Objectives

  1. To master LLM API integration
  2. To set up and run local LLM models
  3. To understand API vs local model trade-offs
  4. To build applications using both approaches
  5. To optimize API usage and local model performance
I

LLM API INTEGRATION

OpenAI API Deep Dive – API Authentication – API Endpoints – Chat Completions API – Completions API – Embeddings API – Function Calling – Streaming Responses. Anthropic API – Claude API Overview – API Authentication – Message API – Streaming – Tool Use – API Best Practices. API Integration Patterns – Synchronous vs Asynchronous Calls – Error Handling – Retry Logic – Rate Limiting – Cost Management – Response Caching.

📝 Illustrative Problems

Integrate OpenAI API; Handle API errors; Implement retry logic; Manage API costs; Cache API responses.

II

LOCAL MODEL SETUP WITH OLLAMA

Introduction to Ollama – What is Ollama? – Why Use Local Models? – Ollama Installation – Ollama Architecture – Supported Models. Setting Up Ollama – Installation on Different Platforms – Downloading Models – Model Management – Running Models – Ollama CLI Usage. Ollama API – REST API Endpoints – Python Integration – Streaming Responses – Model Configuration – Performance Tuning.

📝 Illustrative Problems

Install and configure Ollama; Download and run models; Use Ollama API; Integrate with Python; Optimize performance.

III

WORKING WITH LOCAL MODELS

Available Local Models – Llama Models – Mistral Models – Code Models – Multimodal Models – Model Selection Guide. Model Configuration – Context Window Settings – Temperature and Sampling – GPU vs CPU – Memory Management – Model Quantization. Performance Optimization – Hardware Requirements – GPU Acceleration – Model Quantization – Batch Processing – Optimization Techniques.

📝 Illustrative Problems

Select appropriate local model; Configure model parameters; Optimize performance; Handle memory constraints; Compare model performance.

IV

API VS LOCAL MODELS

Trade-offs Analysis – Cost Comparison – Latency Comparison – Privacy Considerations – Customization – Scalability – Use Case Selection. Hybrid Approaches – Using APIs for Some Tasks – Using Local Models for Others – Fallback Strategies – Cost Optimization – Performance Optimization. Migration Strategies – Moving from API to Local – Moving from Local to API – Hybrid Deployment – Decision Framework.

📝 Illustrative Problems

Compare API vs local models; Choose appropriate approach; Implement hybrid solution; Optimize costs; Plan migration.

V

BUILDING PRODUCTION APPLICATIONS

Application Architecture – Designing LLM Applications – API Integration Patterns – Local Model Integration – Error Handling – Monitoring. Best Practices – Security Considerations – API Key Management – Rate Limiting – Cost Monitoring – Performance Monitoring – Logging and Debugging. Deployment Strategies – Deploying API-Based Applications – Deploying Local Model Applications – Containerization – Cloud Deployment – Edge Deployment.

📝 Illustrative Problems

Design application architecture; Implement security best practices; Deploy applications; Monitor performance; Handle production issues.

📖 Textbooks

  1. 1. OpenAI API Documentation
  2. 2. Anthropic API Documentation
  3. 3. Ollama Documentation
DAGE106

Capstone Project 1

📚 3 Credits 📅 Semester 1 📋 Prerequisite: DAGE101, DAGE103, DAGE105

Course Objectives

  1. To apply Semester 1 concepts in a real-world GenAI project
  2. To demonstrate mastery of prompt engineering and LLM integration
  3. To build a complete GenAI application
  4. To practice project planning and execution

📖 Textbooks

  1. 1. OpenAI API Documentation
  2. 2. Anthropic API Documentation
DAGE201

Vector Databases & RAG Systems

📚 3 Credits 📅 Semester 2 📋 Prerequisite: DAGE101, DAGE103, DAGE105

Course Objectives

  1. To understand vector databases and embeddings
  2. To master semantic search and similarity matching
  3. To build Retrieval-Augmented Generation (RAG) systems
  4. To implement document processing and chunking
  5. To optimize RAG system performance
I

EMBEDDINGS AND VECTOR REPRESENTATIONS

Introduction to Embeddings – What are Embeddings? – Word Embeddings vs Sentence Embeddings – Embedding Models – Embedding Dimensions – Embedding Quality. Generating Embeddings – Using OpenAI Embeddings API – Using Open Source Embedding Models – Sentence Transformers – Embedding Generation Best Practices – Embedding Storage. Vector Similarity – Cosine Similarity – Euclidean Distance – Dot Product – Similarity Metrics Comparison – Choosing Similarity Metrics.

📝 Illustrative Problems

Generate embeddings for text; Calculate vector similarity; Compare embedding models; Store embeddings efficiently; Measure embedding quality.

II

VECTOR DATABASES

Introduction to Vector Databases – What are Vector Databases? – Vector Database vs Traditional Database – Use Cases for Vector Databases – Vector Database Architecture. Pinecone – Pinecone Overview – Setting Up Pinecone – Creating Indexes – Inserting Vectors – Querying Vectors – Pinecone Best Practices. Chroma – Chroma Overview – Installation and Setup – Creating Collections – Adding Documents – Querying – Chroma Features. Other Vector Databases – Weaviate – FAISS (Facebook AI Similarity Search) – Qdrant – Vector Database Comparison – Choosing Vector Database.

📝 Illustrative Problems

Set up vector database; Insert and query vectors; Compare vector databases; Optimize vector operations; Choose appropriate database.

III

RETRIEVAL-AUGMENTED GENERATION (RAG)

RAG Architecture – What is RAG? – Why RAG? – RAG Components: Retrieval, Augmentation, Generation – RAG Workflow – RAG vs Fine-tuning. Building RAG Systems – Document Processing – Text Chunking Strategies – Embedding Generation – Vector Storage – Retrieval Process – Context Assembly. LangChain RAG – LangChain RAG Components – Document Loaders – Text Splitters – Vector Stores – Retrievers – RAG Chains.

📝 Illustrative Problems

Build basic RAG system; Process documents; Implement retrieval; Integrate with LLM; Optimize RAG pipeline.

IV

ADVANCED RAG TECHNIQUES

Document Processing – PDF Processing – Web Scraping – Document Parsing – Text Extraction – Metadata Extraction. Chunking Strategies – Fixed-size Chunking – Semantic Chunking – Recursive Chunking – Overlapping Chunks – Chunk Size Optimization. Retrieval Optimization – Hybrid Search: Keyword + Semantic – Re-ranking Results – Retrieval Strategies – Context Window Management – Retrieval Quality.

📝 Illustrative Problems

Process various document types; Implement advanced chunking; Optimize retrieval; Improve RAG quality; Handle different document formats.

V

RAG SYSTEM OPTIMIZATION AND EVALUATION

RAG Performance Optimization – Reducing Latency – Improving Accuracy – Cost Optimization – Caching Strategies – Batch Processing. RAG Evaluation – Retrieval Metrics – Generation Quality – End-to-End Evaluation – A/B Testing RAG Systems – Evaluation Frameworks. Production RAG Systems – Scalability Considerations – Monitoring RAG Systems – Error Handling – Version Control – Best Practices.

📝 Illustrative Problems

Optimize RAG performance; Evaluate RAG systems; Monitor production RAG; Handle errors; Scale RAG applications.

📖 Textbooks

  1. 1. LangChain Documentation, "RAG Tutorials"
  2. 2. Pinecone Documentation
  3. 3. Harrison Chase, "LangChain for LLM Application Development", DeepLearning.AI
DAGE202

Vector Databases & RAG Systems Lab

📚 2 Credits 📅 Semester 2 📋 Prerequisite: DAGE201

Course Objectives

  1. To implement vector databases
  2. To build RAG systems
  3. To practice semantic search
  4. To optimize RAG performance

📖 Textbooks

  1. 1. LangChain Documentation
  2. 2. Pinecone Documentation
DAGE203

AI Agent Development with LangChain & CrewAI

📚 3 Credits 📅 Semester 2 📋 Prerequisite: DAGE201

Course Objectives

  1. To understand AI agent concepts and architecture
  2. To master LangChain framework for agent development
  3. To learn CrewAI for multi-agent systems
  4. To build autonomous AI agents
  5. To implement agent workflows and orchestration
I

AI AGENT FUNDAMENTALS

Introduction to AI Agents – What are AI Agents? – Agent vs LLM – Agent Architecture – Agent Components: Tools, Memory, Planning – Types of Agents: ReAct, Plan-and-Execute, Multi-Agent. Agent Capabilities – Tool Use – Function Calling – Web Search – Code Execution – Database Access – API Integration – Agent Limitations. Agent Design Patterns – Single Agent Systems – Multi-Agent Systems – Agent Hierarchies – Agent Collaboration – Agent Communication.

📝 Illustrative Problems

Design agent architecture; Choose agent type; Plan agent capabilities; Design agent interactions; Handle agent limitations.

II

LANGCHAIN AGENT FRAMEWORK

LangChain Overview – What is LangChain? – LangChain Components – Agents Module – Tools Module – Memory Module – Chains Module. Building Agents with LangChain – Agent Types: Zero-shot, ReAct, Plan-and-Execute – Creating Agents – Agent Tools – Agent Memory – Agent Execution. LangChain Tools – Built-in Tools – Custom Tools – Tool Wrappers – Tool Selection – Tool Execution – Error Handling.

📝 Illustrative Problems

Create LangChain agent; Add tools to agent; Implement agent memory; Execute agent tasks; Handle agent errors.

III

ADVANCED LANGCHAIN AGENTS

Agent Orchestration – Complex Agent Workflows – Sequential Agent Execution – Parallel Agent Execution – Conditional Logic – State Management. Agent Memory Systems – Conversation Memory – Buffer Memory – Summary Memory – Entity Memory – Memory Optimization. Custom Agent Development – Building Custom Agents – Custom Tools – Custom Chains – Agent Extensions – Best Practices.

📝 Illustrative Problems

Orchestrate agent workflows; Implement memory systems; Build custom agents; Extend agent capabilities; Optimize agent performance.

IV

CREWAI FOR MULTI-AGENT SYSTEMS

CrewAI Introduction – What is CrewAI? – CrewAI Architecture – Agents in CrewAI – Tasks in CrewAI – Crews in CrewAI – CrewAI vs LangChain. Building Multi-Agent Systems – Defining Agents – Creating Tasks – Forming Crews – Agent Roles – Task Assignment – Agent Collaboration. CrewAI Features – Agent Roles and Goals – Task Dependencies – Process Management – Output Parsing – CrewAI Best Practices.

📝 Illustrative Problems

Set up CrewAI; Define agent roles; Create tasks; Form crews; Execute multi-agent workflows; Handle agent collaboration.

V

PRODUCTION AGENT SYSTEMS

Agent Monitoring and Debugging – Monitoring Agent Execution – Debugging Agents – Logging Agent Actions – Performance Metrics – Error Tracking. Agent Optimization – Reducing Latency – Cost Optimization – Improving Accuracy – Caching Strategies – Batch Processing. Deploying Agent Systems – Deployment Strategies – Containerization – Cloud Deployment – API Endpoints – Scaling Agents – Security Considerations.

📝 Illustrative Problems

Monitor agent systems; Debug agent issues; Optimize agent performance; Deploy agents; Scale agent systems; Secure agent deployments.

📖 Textbooks

  1. 1. Harrison Chase, "LangChain for LLM Application Development", DeepLearning.AI
  2. 2. LangChain Documentation
  3. 3. CrewAI Documentation
DAGE204

AI Agent Development Lab

📚 2 Credits 📅 Semester 2 📋 Prerequisite: DAGE203

Course Objectives

  1. To implement AI agents using LangChain
  2. To build multi-agent systems with CrewAI
  3. To practice agent orchestration
  4. To optimize agent performance

📖 Textbooks

  1. 1. LangChain Documentation
  2. 2. CrewAI Documentation
DAGE205

Production GenAI Systems & MCP

📚 3 Credits 📅 Semester 2 📋 Prerequisite: DAGE201

Course Objectives

  1. To understand production GenAI system requirements
  2. To master Model Context Protocol (MCP)
  3. To implement monitoring and evaluation for GenAI
  4. To deploy and manage production GenAI applications
  5. To ensure reliability and scalability
I

PRODUCTION GENAI ARCHITECTURE

Production System Requirements – Scalability – Reliability – Performance – Cost Management – Security – Monitoring – Error Handling. GenAI System Architecture – API Gateway – Load Balancing – Caching Strategies – Rate Limiting – Fallback Mechanisms – Circuit Breakers. Deployment Patterns – Serverless Deployment – Container Deployment – Kubernetes Deployment – Edge Deployment – Hybrid Deployment.

📝 Illustrative Problems

Design production architecture; Plan scalability; Implement reliability; Handle errors; Optimize costs.

II

MODEL CONTEXT PROTOCOL (MCP)

Introduction to MCP – What is MCP? – Why MCP? – MCP Architecture – MCP Components – MCP Use Cases. MCP Implementation – Setting Up MCP – MCP Servers – MCP Clients – Context Management – Model Integration. MCP Best Practices – Context Organization – Context Retrieval – Context Updates – Performance Optimization – Security Considerations.

📝 Illustrative Problems

Set up MCP; Implement MCP servers; Integrate MCP clients; Manage context; Optimize MCP performance.

III

MONITORING AND EVALUATION

GenAI Monitoring – Monitoring Metrics – Latency Monitoring – Cost Monitoring – Quality Monitoring – Error Monitoring – User Feedback. Evaluation Frameworks – Evaluation Metrics – Human Evaluation – Automated Evaluation – A/B Testing – Evaluation Best Practices. Quality Assurance – Input Validation – Output Validation – Quality Checks – Bias Detection – Safety Checks.

📝 Illustrative Problems

Set up monitoring; Implement evaluation; Measure quality; Detect issues; Improve systems.

IV

RELIABILITY AND ERROR HANDLING

Error Handling Strategies – Retry Logic – Fallback Mechanisms – Graceful Degradation – Error Recovery – Error Notifications. Reliability Patterns – Redundancy – Failover – Health Checks – Circuit Breakers – Timeout Handling. Testing GenAI Systems – Unit Testing – Integration Testing – End-to-End Testing – Load Testing – Test Data Management.

📝 Illustrative Problems

Implement error handling; Build reliability; Test systems; Handle failures; Ensure uptime.

V

SCALING AND OPTIMIZATION

Scaling Strategies – Horizontal Scaling – Vertical Scaling – Auto-scaling – Load Distribution – Resource Management. Performance Optimization – Latency Optimization – Throughput Optimization – Cost Optimization – Caching Strategies – Batch Processing. Production Best Practices – Security Best Practices – Compliance Considerations – Documentation – Version Control – Rollback Strategies.

📝 Illustrative Problems

Scale systems; Optimize performance; Reduce costs; Implement security; Document systems.

📖 Textbooks

  1. 1. Model Context Protocol Documentation
  2. 2. OpenAI, "Production Best Practices", OpenAI Documentation
  3. 3. Anthropic, "Production Guide", Anthropic Documentation
DAGE206

Capstone Project 2

📚 6 Credits 📅 Semester 2 📋 Prerequisite: All Semester 2 courses

Course Objectives

  1. To integrate all Semester 2 concepts in a comprehensive GenAI project
  2. To build production-ready GenAI applications
  3. To implement RAG, agents, and MCP
  4. To deploy and manage production GenAI systems

📖 Textbooks

  1. 1. LangChain Documentation
  2. 2. CrewAI Documentation
  3. 3. Model Context Protocol Documentation