KnowYourBasics
Get Started
πŸŽ“ Acharya Nagarjuna University – Distance Education

Diploma in AI and ML Techniques

Industry-focused diploma bridging academic excellence with real-world skills

1 Year Duration
2 Semesters
35 Credits
12 Courses
πŸ“₯ Download Complete Syllabus (DOCX)
← Back to Distance Education Programs

Program Overview

Comprehensive training in ai and ml techniques with industry-standard tools and practices

🎯 Program Objectives

  • Industry-focused learning experiences
  • Hands-on projects and real-world applications
  • Modern tools and technologies
  • Professional skill development
  • Career-ready graduates

πŸ“š Learning Outcomes

  • Master core concepts and fundamentals
  • Apply advanced techniques and methodologies
  • Build production-ready applications
  • Work with industry-standard tools
  • Complete comprehensive capstone projects

πŸ› οΈ Technologies Covered

  • Modern programming languages and frameworks
  • Industry-standard tools and platforms
  • Cloud services and deployment
  • Best practices and methodologies
  • Production systems and MLOps

Course Structure

2 Semesters β€’ 12 Courses β€’ 35 Total Credits

S.No Course Code Course Title Credits
SEMESTER 1: Foundation & Fundamentals (16 Credits)
1 DAMT101 Python for Machine Learning 3
2 DAMT102 Python for Machine Learning Lab 2
3 DAMT103 Data Preprocessing & Feature Engineering 3
4 DAMT104 Data Preprocessing & Feature Engineering Lab 2
5 DAMT105 Supervised Learning Algorithms 3
6 DAMT106 Capstone Project 1 3
SEMESTER 2: Advanced & Production Systems (19 Credits)
7 DAMT201 Deep Learning Fundamentals 3
8 DAMT202 Deep Learning Fundamentals Lab 2
9 DAMT203 Computer Vision & NLP Applications 3
10 DAMT204 Computer Vision & NLP Applications Lab 2
11 DAMT205 MLOps & Production ML Systems 3
12 DAMT206 Capstone Project 2 6

Detailed Syllabus

Click on a course to jump to its detailed syllabus

DAMT101 DAMT102 DAMT103 DAMT104 DAMT105 DAMT106 DAMT201 DAMT202 DAMT203 DAMT204 DAMT205 DAMT206
DAMT101

Python for Machine Learning

πŸ“š 3 Credits πŸ“… Semester 1 πŸ“‹ Prerequisite: Basic Python knowledge

Course Objectives

  1. To introduce Python programming for machine learning
  2. To master essential Python libraries for ML
  3. To learn data manipulation and visualization
  4. To understand scikit-learn basics
  5. To build foundation for ML algorithms
I

PYTHON FUNDAMENTALS FOR ML

Python Basics Review – Variables and Data Types – Control Flow – Functions – List Comprehensions – Dictionary Operations – File Handling. NumPy for Numerical Computing – NumPy Arrays – Array Operations – Mathematical Operations – Broadcasting – Array Indexing and Slicing – NumPy Functions for ML. Pandas for Data Manipulation – Series and DataFrames – Reading Data – Data Selection – Data Filtering – Grouping Operations – Merging DataFrames.

πŸ“ Illustrative Problems

Perform array operations with NumPy; Manipulate data with pandas; Load and explore datasets; Filter and group data; Prepare data for ML.

II

DATA VISUALIZATION FOR ML

Matplotlib Basics – Creating Plots – Line Plots, Scatter Plots, Bar Charts – Customizing Plots – Subplots – Saving Figures. Seaborn for Statistical Visualization – Distribution Plots – Categorical Plots – Relationship Plots – Heatmaps – Advanced Visualizations. Visualization for ML – Feature Distribution Visualization – Correlation Visualization – Target Variable Analysis – Feature Relationships – Pre-ML Data Exploration.

πŸ“ Illustrative Problems

Create data visualizations; Analyze feature distributions; Visualize correlations; Explore target variables; Prepare visual reports.

III

SCIKIT-LEARN BASICS

Introduction to scikit-learn – scikit-learn API – Estimators and Predictors – Fit and Predict Pattern – Model Evaluation Basics. Data Splitting – Train-Test Split – Cross-Validation Concepts – Stratified Splitting – Time Series Splitting. Basic ML Workflow – Loading Data – Splitting Data – Training Model – Making Predictions – Evaluating Model.

πŸ“ Illustrative Problems

Use scikit-learn API; Split datasets; Train basic models; Make predictions; Evaluate models.

IV

DATA PREPARATION FOR ML

Handling Missing Values – Detecting Missing Values – Imputation Strategies – Dropping Missing Values – Handling Missing Data in scikit-learn. Feature Scaling – Why Scale Features? – Standardization – Normalization – Min-Max Scaling – When to Scale. Encoding Categorical Variables – Label Encoding – One-Hot Encoding – Ordinal Encoding – Encoding in scikit-learn – Handling Text Data.

πŸ“ Illustrative Problems

Handle missing values; Scale features; Encode categorical variables; Prepare complete dataset; Use preprocessing pipelines.

V

BASIC ML CONCEPTS

Supervised Learning Overview – What is Supervised Learning? – Regression vs Classification – Training Process – Model Generalization – Overfitting and Underfitting. Model Evaluation Basics – Accuracy Metrics – Confusion Matrix – Precision, Recall, F1-Score – Mean Squared Error – R-squared. Simple ML Models – Linear Regression Basics – Logistic Regression Basics – k-Nearest Neighbors – Model Comparison.

πŸ“ Illustrative Problems

Understand ML concepts; Evaluate models; Train simple models; Compare models; Interpret results.

πŸ“– Textbooks

  1. 1. Jake VanderPlas, "Python Data Science Handbook", O'Reilly Media, 2016
  2. 2. Andreas MΓΌller, Sarah Guido, "Introduction to Machine Learning with Python", O'Reilly Media, 2016
  3. 3. Sebastian Raschka, "Python Machine Learning", 3rd Edition, Packt Publishing, 2019
DAMT102

Python for Machine Learning Lab

πŸ“š 2 Credits πŸ“… Semester 1 πŸ“‹ Prerequisite: DAMT101

Course Objectives

  1. To practice Python for ML tasks
  2. To implement data manipulation workflows
  3. To create ML visualizations
  4. To use scikit-learn for ML

πŸ“– Textbooks

  1. 1. Jake VanderPlas, "Python Data Science Handbook", O'Reilly Media, 2016
  2. 2. Andreas MΓΌller, Sarah Guido, "Introduction to Machine Learning with Python", O'Reilly Media, 2016
DAMT103

Data Preprocessing & Feature Engineering

πŸ“š 3 Credits πŸ“… Semester 1 πŸ“‹ Prerequisite: DAMT101

Course Objectives

  1. To master data cleaning and preprocessing techniques
  2. To learn feature engineering methods
  3. To understand feature selection strategies
  4. To handle different data types and formats
  5. To build preprocessing pipelines
I

DATA CLEANING AND QUALITY

Data Quality Issues – Missing Values – Duplicate Records – Inconsistent Data – Outliers – Data Types – Data Quality Assessment. Handling Missing Data – Types of Missingness: MCAR, MAR, MNAR – Detection Methods – Imputation Strategies: Mean, Median, Mode, KNN Imputation – Advanced Imputation – When to Drop Missing Values. Outlier Detection and Treatment – What are Outliers? – Detection Methods: IQR, Z-score, Isolation Forest – Outlier Treatment: Removal, Capping, Transformation – Domain-Specific Outliers.

πŸ“ Illustrative Problems

Assess data quality; Handle missing values; Detect outliers; Treat outliers; Clean datasets.

II

FEATURE SCALING AND TRANSFORMATION

Why Feature Scaling? – Impact on ML Algorithms – Algorithms Requiring Scaling – Algorithms Not Requiring Scaling. Scaling Techniques – Standardization (Z-score Normalization) – Min-Max Scaling – Robust Scaling – Normalization – When to Use Each Method. Data Transformation – Log Transformation – Square Root Transformation – Box-Cox Transformation – Power Transformations – Handling Skewed Data.

πŸ“ Illustrative Problems

Scale features appropriately; Transform skewed data; Choose scaling method; Apply transformations; Optimize feature distributions.

III

ENCODING CATEGORICAL VARIABLES

Categorical Data Types – Nominal Variables – Ordinal Variables – High Cardinality Categorical – Handling Categorical Data. Encoding Techniques – Label Encoding – One-Hot Encoding – Ordinal Encoding – Target Encoding – Frequency Encoding – Binary Encoding. Text Data Encoding – Bag of Words – TF-IDF – Word Embeddings (Introduction) – Text Preprocessing – Handling Text Features.

πŸ“ Illustrative Problems

Encode categorical variables; Handle high cardinality; Encode text data; Choose encoding method; Optimize encoding.

IV

FEATURE ENGINEERING

Feature Engineering Concepts – What is Feature Engineering? – Creating New Features – Domain Knowledge – Feature Interactions – Polynomial Features. Temporal Features – Date and Time Features – Time-based Features – Cyclical Encoding – Lag Features – Rolling Statistics. Numerical Feature Engineering – Binning – Discretization – Feature Interactions – Ratio Features – Aggregated Features.

πŸ“ Illustrative Problems

Create temporal features; Engineer numerical features; Build feature interactions; Use domain knowledge; Generate new features.

V

FEATURE SELECTION

Feature Selection Importance – Curse of Dimensionality – Benefits of Feature Selection – Feature Selection vs Feature Extraction. Feature Selection Methods – Filter Methods: Correlation, Chi-square, Mutual Information – Wrapper Methods: Forward Selection, Backward Elimination – Embedded Methods: Lasso, Ridge, Tree-based – Feature Importance. Dimensionality Reduction (Introduction) – Principal Component Analysis (PCA) Basics – When to Use PCA – PCA Limitations – Other Dimensionality Reduction Techniques.

πŸ“ Illustrative Problems

Select features using filters; Use wrapper methods; Apply embedded methods; Reduce dimensionality; Optimize feature set.

πŸ“– Textbooks

  1. 1. Andreas MΓΌller, Sarah Guido, "Introduction to Machine Learning with Python", O'Reilly Media, 2016
  2. 2. AurΓ©lien GΓ©ron, "Hands-On Machine Learning", 3rd Edition, O'Reilly Media, 2022
DAMT104

Data Preprocessing & Feature Engineering Lab

πŸ“š 2 Credits πŸ“… Semester 1 πŸ“‹ Prerequisite: DAMT103

Course Objectives

  1. To implement data preprocessing techniques
  2. To practice feature engineering
  3. To build preprocessing pipelines
  4. To optimize feature sets

πŸ“– Textbooks

  1. 1. Andreas MΓΌller, Sarah Guido, "Introduction to Machine Learning with Python", O'Reilly Media, 2016
DAMT105

Supervised Learning Algorithms

πŸ“š 3 Credits πŸ“… Semester 1 πŸ“‹ Prerequisite: DAMT101, DAMT103

Course Objectives

  1. To understand supervised learning concepts
  2. To master regression algorithms
  3. To learn classification algorithms
  4. To evaluate model performance
  5. To compare different algorithms
I

LINEAR REGRESSION

Linear Regression Fundamentals – What is Linear Regression? – Simple Linear Regression – Multiple Linear Regression – Assumptions of Linear Regression – Cost Function – Gradient Descent. Implementing Linear Regression – Using scikit-learn – Training Linear Regression – Making Predictions – Interpreting Coefficients – Model Evaluation. Polynomial Regression – When to Use Polynomial Regression – Polynomial Features – Overfitting in Polynomial Regression – Regularization Concepts.

πŸ“ Illustrative Problems

Implement linear regression; Train regression models; Evaluate regression performance; Use polynomial regression; Handle overfitting.

II

LOGISTIC REGRESSION AND CLASSIFICATION BASICS

Logistic Regression – What is Logistic Regression? – Sigmoid Function – Decision Boundary – Binary Classification – Multi-class Classification – Using scikit-learn. Classification Metrics – Accuracy – Precision, Recall, F1-Score – Confusion Matrix – ROC Curve and AUC – Classification Report. Classification Algorithms Overview – k-Nearest Neighbors (k-NN) – Naive Bayes – Algorithm Comparison – When to Use Each Algorithm.

πŸ“ Illustrative Problems

Implement logistic regression; Build classification models; Calculate classification metrics; Compare algorithms; Choose appropriate algorithm.

III

DECISION TREES AND ENSEMBLE METHODS

Decision Trees – What are Decision Trees? – How Decision Trees Work – Splitting Criteria: Gini, Entropy – Tree Pruning – Overfitting in Trees – Using scikit-learn. Random Forests – What are Random Forests? – Bagging Concept – Random Forest Algorithm – Feature Importance – Hyperparameter Tuning – Advantages of Random Forests. Gradient Boosting – Boosting Concept – Gradient Boosting Machines – XGBoost Introduction – LightGBM Introduction – When to Use Boosting.

πŸ“ Illustrative Problems

Build decision trees; Implement random forests; Use gradient boosting; Tune hyperparameters; Compare ensemble methods.

IV

SUPPORT VECTOR MACHINES AND ADVANCED CLASSIFIERS

Support Vector Machines (SVM) – What are SVMs? – Maximum Margin Concept – Kernel Trick – Linear vs Non-linear SVMs – SVM Hyperparameters – Using scikit-learn. Advanced Classification – Neural Networks Basics (Introduction) – Algorithm Selection Guide – Ensemble Voting – Stacking Concepts. Model Evaluation and Validation – Cross-Validation – Stratified Cross-Validation – Learning Curves – Validation Curves – Bias-Variance Trade-off.

πŸ“ Illustrative Problems

Implement SVMs; Use different kernels; Evaluate models with CV; Analyze learning curves; Optimize model performance.

V

MODEL SELECTION AND OPTIMIZATION

Hyperparameter Tuning – What are Hyperparameters? – Grid Search – Random Search – Bayesian Optimization (Introduction) – scikit-learn Tools. Model Comparison – Comparing Multiple Models – Performance Metrics – Computational Cost – Interpretability – Choosing Best Model. Model Interpretation – Feature Importance – Model Coefficients – Partial Dependence Plots (Introduction) – Model Explainability Basics.

πŸ“ Illustrative Problems

Tune hyperparameters; Compare models; Interpret model results; Select best model; Optimize performance.

πŸ“– Textbooks

  1. 1. Andreas MΓΌller, Sarah Guido, "Introduction to Machine Learning with Python", O'Reilly Media, 2016
  2. 2. Sebastian Raschka, "Python Machine Learning", 3rd Edition, Packt Publishing, 2019
  3. 3. AurΓ©lien GΓ©ron, "Hands-On Machine Learning", 3rd Edition, O'Reilly Media, 2022
DAMT106

Capstone Project 1

πŸ“š 3 Credits πŸ“… Semester 1 πŸ“‹ Prerequisite: DAMT101, DAMT103, DAMT105

Course Objectives

  1. To apply Semester 1 concepts in a complete ML project
  2. To demonstrate mastery of data preprocessing and supervised learning
  3. To build and evaluate ML models
  4. To practice end-to-end ML workflow

πŸ“– Textbooks

  1. 1. Andreas MΓΌller, Sarah Guido, "Introduction to Machine Learning with Python", O'Reilly Media, 2016
  2. 2. AurΓ©lien GΓ©ron, "Hands-On Machine Learning", 3rd Edition, O'Reilly Media, 2022
DAMT201

Deep Learning Fundamentals

πŸ“š 3 Credits πŸ“… Semester 2 πŸ“‹ Prerequisite: DAMT105

Course Objectives

  1. To understand neural networks and deep learning concepts
  2. To master backpropagation and optimization algorithms
  3. To learn activation functions and network architectures
  4. To build deep learning models using TensorFlow/Keras
  5. To train and optimize neural networks
I

NEURAL NETWORKS FUNDAMENTALS

Introduction to Neural Networks – What are Neural Networks? – Biological Inspiration – Perceptron Model – Multi-Layer Perceptron (MLP) – Neural Network Architecture: Input Layer, Hidden Layers, Output Layer – Forward Propagation. Activation Functions – Why Activation Functions? – Sigmoid Function – Tanh Function – ReLU and Variants: Leaky ReLU, ELU, Swish – Activation Function Selection – Vanishing Gradient Problem. Neural Network Training – Loss Functions: MSE, Cross-Entropy – Cost Function – Gradient Descent – Learning Rate – Batch Processing – Epochs and Iterations.

πŸ“ Illustrative Problems

Build simple neural network; Implement forward propagation; Choose activation functions; Calculate loss; Train basic network.

II

BACKPROPAGATION AND OPTIMIZATION

Backpropagation Algorithm – What is Backpropagation? – Chain Rule – Computing Gradients – Backward Pass – Gradient Flow – Implementing Backpropagation. Optimization Algorithms – Gradient Descent Variants: Batch, Stochastic, Mini-batch – Momentum – RMSprop – Adam Optimizer – Learning Rate Scheduling – Adaptive Learning Rates. Regularization Techniques – Overfitting in Neural Networks – L1 and L2 Regularization – Dropout – Early Stopping – Data Augmentation – Batch Normalization.

πŸ“ Illustrative Problems

Implement backpropagation; Use different optimizers; Apply regularization; Prevent overfitting; Optimize training process.

III

DEEP LEARNING FRAMEWORKS

TensorFlow and Keras Introduction – TensorFlow Overview – Keras High-Level API – TensorFlow vs Keras – Installation and Setup – TensorFlow 2.x Features. Building Models with Keras – Sequential API – Functional API – Model Definition – Layer Types: Dense, Dropout, BatchNormalization – Compiling Models – Model Summary. Training Models – Model Training: fit() method – Validation Data – Callbacks: EarlyStopping, ModelCheckpoint, ReduceLROnPlateau – Training History – Monitoring Training.

πŸ“ Illustrative Problems

Set up TensorFlow/Keras; Build models using Sequential API; Use Functional API; Train models; Monitor training progress; Save models.

IV

ADVANCED NEURAL NETWORK ARCHITECTURES

Deep Networks – Deep vs Shallow Networks – Benefits of Depth – Challenges: Vanishing Gradients, Overfitting – Residual Connections – Skip Connections. Network Architectures – Feedforward Networks – Wide vs Deep Networks – Network Design Principles – Hyperparameter Tuning: Layers, Neurons, Learning Rate – Architecture Search Basics. Transfer Learning Concepts – What is Transfer Learning? – Pre-trained Models – Fine-tuning – Feature Extraction – Transfer Learning Benefits – When to Use Transfer Learning.

πŸ“ Illustrative Problems

Design deep architectures; Build wide networks; Implement skip connections; Apply transfer learning; Fine-tune pre-trained models.

V

MODEL EVALUATION AND DEPLOYMENT BASICS

Model Evaluation – Training vs Validation vs Test Sets – Evaluation Metrics for Deep Learning – Confusion Matrix – Classification Report – Regression Metrics – Model Comparison. Model Saving and Loading – Saving Models – Model Formats: H5, SavedModel – Loading Models – Model Versioning – Model Checkpointing. Introduction to Model Deployment – Deployment Options – Model Conversion – ONNX Format (Introduction) – Deployment Considerations – Model Serving Basics.

πŸ“ Illustrative Problems

Evaluate deep learning models; Save and load models; Compare model architectures; Prepare models for deployment; Convert model formats.

πŸ“– Textbooks

  1. 1. Ian Goodfellow, Yoshua Bengio, Aaron Courville, "Deep Learning", MIT Press, 2016
  2. 2. AurΓ©lien GΓ©ron, "Hands-On Machine Learning", 3rd Edition, O'Reilly Media, 2022
  3. 3. FranΓ§ois Chollet, "Deep Learning with Python", 2nd Edition, Manning Publications, 2021
DAMT202

Deep Learning Fundamentals Lab

πŸ“š 2 Credits πŸ“… Semester 2 πŸ“‹ Prerequisite: DAMT201

Course Objectives

  1. To gain hands-on experience with TensorFlow and Keras
  2. To implement neural networks from scratch
  3. To build and train deep learning models
  4. To experiment with different architectures and hyperparameters
  5. To evaluate and optimize model performance
DAMT203

Computer Vision & NLP Applications

πŸ“š 3 Credits πŸ“… Semester 2 πŸ“‹ Prerequisite: DAMT201

Course Objectives

  1. To understand computer vision fundamentals and CNNs
  2. To learn natural language processing basics
  3. To build image classification models
  4. To implement text analysis and sentiment analysis
  5. To apply deep learning to CV and NLP tasks
I

COMPUTER VISION FUNDAMENTALS

Introduction to Computer Vision – What is Computer Vision? – Image Representation – Pixels and Color Channels – Image Preprocessing – Common CV Tasks: Classification, Detection, Segmentation. Convolutional Neural Networks (CNNs) – Why CNNs for Images? – Convolution Operation – Filters and Kernels – Feature Maps – Convolution Layers – Pooling Layers: Max Pooling, Average Pooling – CNN Architecture. Building CNNs – CNN Layers: Conv2D, MaxPooling2D, Flatten – CNN Architecture Design – Building CNN with Keras – Training CNNs – Visualizing CNN Features.

πŸ“ Illustrative Problems

Preprocess images; Build basic CNN; Implement convolution layers; Design CNN architecture; Train image classifier.

II

ADVANCED CNN ARCHITECTURES

Popular CNN Architectures – LeNet – AlexNet – VGG – ResNet Concepts – Transfer Learning with Pre-trained CNNs – Using Pre-trained Models: VGG16, ResNet50. Image Augmentation – Why Data Augmentation? – Augmentation Techniques: Rotation, Scaling, Flipping, Cropping – Keras ImageDataGenerator – Augmentation Best Practices – Handling Small Datasets. Transfer Learning for Images – Loading Pre-trained Models – Feature Extraction – Fine-tuning – Freezing Layers – Transfer Learning Workflow – When to Use Transfer Learning.

πŸ“ Illustrative Problems

Use pre-trained CNNs; Apply image augmentation; Implement transfer learning; Fine-tune models; Optimize CNN performance.

III

NATURAL LANGUAGE PROCESSING FUNDAMENTALS

Introduction to NLP – What is NLP? – NLP Applications – Text Representation Challenges – Text Preprocessing – Tokenization – Text Cleaning. Text Preprocessing – Lowercasing – Removing Punctuation – Stop Word Removal – Stemming – Lemmatization – Handling Special Characters – Text Normalization. Text Representation – Bag of Words – TF-IDF – Word Embeddings Introduction – Word2Vec Concepts – Embedding Dimensions – Text Vectorization.

πŸ“ Illustrative Problems

Preprocess text data; Tokenize text; Create text representations; Generate word embeddings; Prepare text for ML models.

IV

DEEP LEARNING FOR NLP

Neural Networks for Text – Feedforward Networks for NLP – Embedding Layers – Sequence Models Introduction – RNN Concepts – LSTM Introduction – GRU Introduction. Sentiment Analysis – What is Sentiment Analysis? – Building Sentiment Classifiers – Using Pre-trained Embeddings – Text Classification with Deep Learning – Model Architecture for Sentiment Analysis. Text Classification – Binary Classification – Multi-class Classification – Multi-label Classification – Building Text Classifiers – Evaluating Text Models – Handling Imbalanced Data.

πŸ“ Illustrative Problems

Build text classification models; Implement sentiment analysis; Use embedding layers; Train NLP models; Evaluate text models.

V

PRACTICAL APPLICATIONS

Image Classification Project – End-to-End Image Classification – Data Collection and Preparation – Model Building – Training and Evaluation – Deployment Considerations. Sentiment Analysis Project – Building Sentiment Analyzer – Data Preprocessing – Model Development – Training and Tuning – Evaluation and Testing. Combining CV and NLP – Image Captioning Concepts – Visual Question Answering Introduction – Multimodal Learning Basics – Real-world Applications.

πŸ“ Illustrative Problems

Complete image classification project; Build sentiment analysis system; Integrate CV and NLP; Deploy models; Create end-to-end applications.

πŸ“– Textbooks

  1. 1. FranΓ§ois Chollet, "Deep Learning with Python", 2nd Edition, Manning Publications, 2021
  2. 2. JΓΌrgen Schmidhuber, "Deep Learning in Neural Networks: An Overview", 2015
  3. 3. Jurafsky & Martin, "Speech and Language Processing", 3rd Edition (Online)
DAMT204

Computer Vision & NLP Applications Lab

πŸ“š 2 Credits πŸ“… Semester 2 πŸ“‹ Prerequisite: DAMT203

Course Objectives

  1. To gain hands-on experience with computer vision tasks
  2. To implement NLP preprocessing and models
  3. To build image classification systems
  4. To develop sentiment analysis applications
  5. To work with real-world CV and NLP datasets
DAMT205

MLOps & Production ML Systems

πŸ“š 3 Credits πŸ“… Semester 2 πŸ“‹ Prerequisite: DAMT201, DAMT203

Course Objectives

  1. To understand MLOps principles and practices
  2. To learn model versioning and management
  3. To master model deployment strategies
  4. To implement monitoring and maintenance
  5. To build production-ready ML systems
I

MLOPS FUNDAMENTALS

Introduction to MLOps – What is MLOps? – MLOps vs DevOps – ML Lifecycle – Challenges in Production ML – MLOps Principles – MLOps Maturity Levels. ML Workflow – Data Collection – Data Preparation – Model Training – Model Evaluation – Model Deployment – Model Monitoring – Continuous Improvement. MLOps Tools and Platforms – MLflow Introduction – Kubeflow Concepts – TensorFlow Extended (TFX) Overview – Cloud ML Platforms – Tool Comparison.

πŸ“ Illustrative Problems

Design MLOps workflow; Set up MLOps environment; Choose MLOps tools; Plan ML lifecycle; Implement basic MLOps pipeline.

II

MODEL VERSIONING AND MANAGEMENT

Model Versioning – Why Version Models? – Model Versioning Strategies – Version Control for Models – Model Registry – Model Metadata – Model Lineage. MLflow for Model Management – MLflow Components: Tracking, Projects, Models, Registry – Logging Experiments – Model Registration – Model Serving – MLflow Workflow. Model Storage – Model Formats: H5, SavedModel, ONNX, Pickle – Model Storage Best Practices – Cloud Storage for Models – Model Archival – Retrieving Models.

πŸ“ Illustrative Problems

Version ML models; Use MLflow for tracking; Register models; Store models efficiently; Retrieve model versions; Track model lineage.

III

MODEL DEPLOYMENT

Deployment Strategies – Batch Inference – Real-time Inference – Edge Deployment – Cloud Deployment – On-premises Deployment – Deployment Options Comparison. Model Serving – REST API for Models – Flask/FastAPI for Model Serving – Model Endpoints – Request/Response Handling – Error Handling – Load Balancing. Containerization – Docker for ML Models – Creating Docker Images – Docker Compose – Container Orchestration Basics – Kubernetes Introduction – Container Best Practices.

πŸ“ Illustrative Problems

Deploy models as APIs; Create model endpoints; Containerize models; Serve models in production; Handle deployment errors; Scale model serving.

IV

MODEL MONITORING AND MAINTENANCE

Model Monitoring – Why Monitor Models? – Data Drift Detection – Concept Drift Detection – Performance Monitoring – Latency Monitoring – Resource Monitoring. Monitoring Metrics – Prediction Accuracy – Model Performance Metrics – Data Quality Metrics – System Metrics – Business Metrics – Alerting Thresholds. Model Retraining – When to Retrain? – Automated Retraining – Retraining Triggers – A/B Testing Models – Model Rollback – Continuous Learning Concepts.

πŸ“ Illustrative Problems

Set up model monitoring; Detect data drift; Monitor model performance; Implement alerts; Plan retraining strategy; Handle model degradation.

V

CI/CD FOR ML AND BEST PRACTICES

CI/CD for ML – Continuous Integration for ML – Continuous Deployment – Testing ML Models – Model Validation – Pipeline Automation – CI/CD Tools for ML. ML Pipeline Automation – Automated Data Pipeline – Automated Training Pipeline – Automated Deployment – Pipeline Orchestration – Error Handling in Pipelines – Pipeline Monitoring. Production Best Practices – Code Quality – Documentation – Security Considerations – Cost Optimization – Scalability – Disaster Recovery – Compliance.

πŸ“ Illustrative Problems

Implement CI/CD for ML; Automate ML pipelines; Test ML systems; Optimize costs; Scale systems; Ensure security; Document processes.

πŸ“– Textbooks

  1. 1. Mark Treveil, "Introducing MLOps", O'Reilly Media, 2020
  2. 2. Noah Gift, "Practical MLOps", O'Reilly Media, 2021
  3. 3. MLflow Documentation
DAMT206

Capstone Project 2

πŸ“š 6 Credits πŸ“… Semester 2 πŸ“‹ Prerequisite: All Semester 2 courses

Course Objectives

  1. To integrate all concepts learned throughout the diploma program
  2. To develop a comprehensive ML application from end to end
  3. To apply deep learning techniques to real-world problems
  4. To implement MLOps practices for production deployment
  5. To demonstrate professional ML development skills

πŸ“– Textbooks

  1. 1. Relevant course textbooks from all courses
  2. 2. Industry best practices documentation
  3. 3. MLOps and deployment guides
  4. 4. Project-specific research papers