Master cutting-edge Generative AI architectures, large-scale model training, and production-grade Agentic AI systems. From Transformer internals to multi-agent orchestration — everything you need to build and deploy real-world AI at the frontier.
Sat 3 hrs + Sun 3 hrs = 6 hrs/week | 72 Total Hours
📅 Cohort Start: May 2, 2026
From low-level transformer internals to production multi-agent systems — this is the deepest AI engineering program available on weekends.
Every week: Saturday (3 hrs) Training + Sunday (3 hrs) POC & Project.
FlashAttention-3 — memory-efficient attention for long contexts
Rotary Positional Embeddings (RoPE) — why they beat sinusoidal
State Space Models (SSMs) — Mamba architecture deep dive
Why SSMs are replacing standard attention for long-context windows
| Day 1: | Training + Demo |
| Day 2: | Training + POC Build |
| Deliverable: | Working UI App |
Pre-training at scale — data curation pipelines & quality filtering
Tokenizer optimization — BPE, SentencePiece, and vocabulary design
Distributed Training — DeepSpeed ZeRO stages & Megatron-LM tensor parallelism
Training stability — gradient clipping, loss spikes, and checkpointing strategies
| Day 1: | Training + Demo |
| Day 2: | Training + POC Build |
| Deliverable: | Working UI App |
Beyond Stable Diffusion — Rectified Flow and why it converges faster
ControlNet internals — conditioning mechanisms & adapter architecture
Latent Consistency Models (LCMs) — real-time generation in 4 steps
Flow Matching vs DDPM — mathematical intuition and practical tradeoffs
| Day 1: | Training + Demo |
| Day 2: | Training + POC Build |
| Deliverable: | Working UI App |
How models "see" — Vision Transformers (ViT) and patch embeddings
CLIP-style contrastive alignment — training vision-language models
Audio-native LLMs — Whisper architecture and speech-text alignment
VLM architectures — LLaVA, Flamingo, and cross-modal attention
| Day 1: | Training + Demo |
| Day 2: | Training + POC Build |
| Deliverable: | Working UI App |
QLoRA — 4-bit quantized LoRA for fine-tuning on consumer GPUs
DoRA (Weight-Decomposed Low-Rank Adaptation) — why it outperforms LoRA
Unsloth — 2x faster fine-tuning with memory optimization tricks
Evaluation — perplexity, ROUGE, and domain-specific benchmarks
| Day 1: | Training + Demo |
| Day 2: | Training + POC Build |
| Deliverable: | Working UI App |
RLHF pipeline — reward modeling, PPO, and preference datasets
Direct Preference Optimization (DPO) — why it's simpler than RLHF
KTO (Kahneman-Tversky Optimization) — aligning models for better reasoning
Constitutional AI & self-critique — Anthropic's alignment approach
| Day 1: | Training + Demo |
| Day 2: | Training + POC Build |
| Deliverable: | Working UI App |
GraphRAG — using knowledge graphs for structured retrieval (Microsoft GraphRAG)
Agentic RAG — model decides what to search, when, and how many times
Self-RAG — models that critique and re-rank their own retrieved context
HyDE & query rewriting — hypothetical document embeddings for better recall
| Day 1: | Training + Demo |
| Day 2: | Training + POC Build |
| Deliverable: | Working UI App |
Quantization techniques — AWQ, GGUF, and FP8/INT4 precision tradeoffs
vLLM — PagedAttention and continuous batching for high-throughput serving
TGI (Text Generation Inference) — production serving with HuggingFace
Edge deployment — running quantized models on-device vs cloud serving
| Day 1: | Training + Demo |
| Day 2: | Training + POC Build |
| Deliverable: | Working UI App |
Agentic loops — ReAct, Plan-and-Execute, and AutoGPT-style architectures
LangGraph — building stateful, cyclical agent workflows with persistence
Long-term memory — episodic, semantic, and procedural memory for agents
Tool-use & function calling — structured outputs and API orchestration
| Day 1: | Training + Demo |
| Day 2: | Training + POC Build |
| Deliverable: | Working UI App |
Multi-agent architectures — Coder, Critic, Manager agent patterns
State management across agents — shared memory and message passing
Agent Communication Protocols — structured handoffs and error recovery
CrewAI & AutoGen — frameworks for scalable multi-agent collaboration
| Day 1: | Training + Demo |
| Day 2: | Training + POC Build |
| Deliverable: | Working UI App |
Evaluation dimensions — correctness, faithfulness, groundedness, context relevance, toxicity, coherence & latency; designing a multi-dimensional scorecard that reflects real production SLAs
RAG evaluation pipelines — RAGAS metrics (faithfulness, answer relevance, context precision & recall), TruLens feedback functions, and citation accuracy measurement on a legal document Q&A system
Agent & task-specific evaluation — agent trajectory scoring (step correctness, goal completion rate, tool-call efficiency), code generation eval (functional correctness via unit tests, security scanning), and multi-turn conversation eval
Production eval infrastructure — BrainTrust, DeepEval, OpenAI Evals & PromptFoo; CI/CD quality gates that block deploys on regression; LLM-as-a-judge calibration; human-in-the-loop preference annotation (RLHF-style labelling workflows)
| Day 1: | Eval Frameworks |
| Day 2: | Eval Harness Build |
| Deliverable: | Eval Dashboard + CI Gate |
World Models — how AI learns to simulate environments (JEPA, Dreamer)
Test-Time Compute — scaling inference like OpenAI o1 (chain-of-thought search)
Capstone architecture review — final feedback and production readiness audit
Career pathways — AI Research Engineer, MLOps Lead, AI Product roles
| Day 1: | Future Trends + Review |
| Day 2: | Capstone Demo |
| Deliverable: | Deployed Agentic App |
12 weeks of frontier AI engineering — with live mentorship, 12 POC builds, and a production-grade capstone project.