A first-principles curriculum

The AI Engineer's Learning Path

A first principles,decision-driven curriculum that takes you from zero to production-ready in building with LLMs and generative AI.

8
Modules
34
Lessons
17
References
6
Provider Paths
Start with the Guides Browse Modules

Start with the problem.
Experience the gap.
Then build the solution.

First Principles, Not Recipes
Recipes, cookbooks and playbooks help when you have a foundation to plug them into. Build that foundation here, learn core concepts, discover the reasoning behind techniques, and the mechanics that make them work.
Measure Before You Build
You can't improve what you don't measure. Define what "good" looks like before writing any agent code, and save the vibes for toy projects.
Vendor-Agnostic Learning
Learn the principles, use the frameworks. OpenAI, Gemini, Anthropic, Hugging Face, Ollama, and GitHub Models are all supported paths. The concepts matter more than any specific implementation.
Build a System, Not a Demo
Each module in this course builds on top of the previous, leading you to an agent build that can be used in production and adapted to new use cases.

8 tailored modules. What you'll learn in this path.

These topics are intentionally ordered so we cut through the hype and focus on what matters, building on something you've already experienced. You'll learn things because you need to know them, and how to discern whether or not they're needed.

Module 00
Orientation
Before diving into any modules, get familiar with the landscape. These guides cover what AI engineering actually is, how this path is structured, how to choose a provider, and the hard rules that apply throughout.
9 guides
Module 01
Foundations of AI Engineering
Get oriented with the core concepts that underpin everything else in this path. Understand the landscape, the terminology, and the mental models you'll use throughout. This is the foundation the rest of the modules assume you have.
6 lessons
Module 02
Benchmark and Harness
Define what "good" looks like before you write any code. Learn to measure outcomes, design benchmarks that matter, and establish baselines you can actually improve against. Leave the vibes to toy projects.
3 lessons
Module 03
Agent and Tool Building
Build a tool-calling agent from scratch, then rebuild it with a framework so you understand what the framework actually does for you. Learn to make your tools portable with MCP so they aren't locked to a single runtime. The raw loop comes first because you need to know what you're abstracting away.
4 lessons
Module 04
Code Retrieval
Learn retrieval by building it four different ways. Each approach handles something the last one doesn't, from naive vector search to full context compilation. The goal isn't to always reach for the most complex option, but to know when each one is the right call.
5 lessons
Module 05
RAG and Grounded Answers
Retrieval on its own just finds things. This module connects retrieval to answer generation so your system can cite its sources and stay grounded in evidence. Learn to build context packs, route between retrieval modes, and know when the model has enough information to answer.
3 lessons
Module 06
Observability and Evals
If you can't see what your system is doing, you're just vibing. Build telemetry into your agent from the start, track costs and outcomes, then write evals that hold it accountable. This module ties measurement back to the benchmarks you built in Module 02.
5 lessons
Module 07
Orchestration and Memory
Learn to coordinate multiple agents, design specialists that handle specific tasks, and route work between them. Then tackle memory, both short-lived thread state and long-term persistence, so your system can remember what matters across conversations.
5 lessons
Module 08
Optimization
Premature optimization is a costly distraction. Until your system is stable and measured, tuning it is guesswork. Learn when distillation and fine-tuning actually make sense, and how to apply them without breaking what already works.
3 lessons

References to keep within reach

Lookup tables, schemas, decision frameworks, and terminology you'll return to throughout the path.

Every example runs on your provider

Direct provider APIs
OpenAI
The most widely documented API surface. If you already have an OpenAI platform key, start here.
Gemini
Google's direct model API. A strong free tier and native SDK make it a practical alternative to OpenAI.
Anthropic
Claude's direct API, or through AWS Bedrock. Strong structured output and tool-calling support.
Also supported through
Hugging Face
Routes to thousands of open models through one free account. A good path if you want to learn without spending.
Ollama
Run models locally or through Ollama Cloud. The best option for privacy, offline work, or predictable costs.
GitHub Models
Hosted inference through your GitHub account. Access multiple model publishers from a single API surface.

These are just the surfaces we interact with. The concepts are the focus, and your learning isn't gated behind any single vendor.

Get started orienting yourself to the AI space before starting any modules

Start with the guides to understand the approach, choose your provider, and set up your environment. Then Module 1 begins.