Module 3: Agent and Tool Building Building a Raw Tool Loop

Build a Raw Tool-Calling Loop

Up to now, the model has been answering questions from training data alone, and your baseline showed exactly how limited that is. In this lesson, we'll give the model tools it can use to actually read your codebase: list files, search text, and read file contents. The model will decide which tools to call, in what order, and when it has enough information to answer.

We're building this from scratch. No framework, no library. Just a Python loop that sends a message, checks whether the model wants to call a tool, executes it, sends the result back, and repeats. Understanding these raw mechanics is important because every agent framework is an abstraction over this same loop. If you understand the loop, you can debug any framework.

What you'll learn

  • Build a tool-calling control loop that lets a model interact with your codebase
  • Define tool schemas that tell the model what tools are available and what arguments they accept
  • Implement three core tools: list_files, search_text, and read_file
  • Apply the tool argument validation patterns from Security Basics
  • Run benchmark questions through your agent and compare results against your Module 2 baseline

Concepts

Agent: a system where a model decides which tools to call, observes the results, and iterates until a task is complete. The minimal definition: agent = model + tools + control loop. Everything else (state management, error handling, orchestration) builds on top of this.

Control loop: the code that manages the agent's cycle: send prompt → check for tool calls → execute tools → append results → repeat or finish. You own this loop. The model makes requests; your code decides whether and how to fulfill them.

Tool schema: a JSON description of a tool that tells the model what it does, what arguments it accepts, and what types those arguments are. The model uses this schema to decide when to call the tool and what arguments to provide. A clear schema reduces bad tool calls; a vague schema produces garbage arguments.

Tool result: the output your code returns to the model after executing a tool call. The model treats this as new context and decides what to do next: call another tool, call the same tool with different arguments, or produce a final answer.

Problem-to-Tool Map

Problem classSymptomCheapest thing to try firstTool or approach
Model can't inspect your repoAnswers are generic or hallucinated because the model has no access to your codeManual copy-paste of code into the promptGive the model tools to list, search, and read files
Tool calls have bad argumentsModel requests files that don't exist or passes wrong argument typesFix the tool schema descriptionsAdd argument validation with allowlists
Agent loops foreverModel keeps calling tools without converging on an answerSet a maximum iteration countAdd a hard stop after N tool rounds

Walkthrough

Project setup

We'll work in your anchor repository. Make sure your environment from Module 2 is active:

cd anchor-repo
source .venv/bin/activate

Create an agent/ directory for this module's code:

mkdir -p agent

By the end of this lesson, you'll have:

anchor-repo/
├── agent/
│   ├── tools.py          # Tool implementations
│   ├── schemas.py         # Tool schema definitions
│   ├── loop.py            # The raw control loop
│   └── run_benchmark.py   # Benchmark runner using the agent
├── harness/               # From Module 2
│   ├── runs/
│   └── ...
└── benchmark-questions.jsonl

Define your tools

Create three tools that let the model explore your codebase. Each tool is a plain Python function with input validation.

# agent/tools.py
"""Tools that let the model interact with the anchor repository."""
import os
import subprocess
from pathlib import Path

# All tool operations are restricted to this directory
REPO_ROOT = Path(".").resolve()

# Directories to exclude from search and listing — .venv, .git, __pycache__, etc.
EXCLUDED_DIRS = {".venv", ".git", "__pycache__", "node_modules", ".tox", ".mypy_cache"}


def _is_excluded(path: Path) -> bool:
    """Check whether a path falls inside an excluded directory.

    Args:
        path: Path relative to the repository root.

    Returns:
        bool: True when the path should be hidden from tool access.
    """
    return any(part in EXCLUDED_DIRS for part in path.parts)


def validate_path(path_str: str) -> Path:
    """Resolve and validate a repo-relative path for safe tool use.

    Args:
        path_str: Path provided by the model, relative to ``REPO_ROOT``.

    Returns:
        Path: A fully resolved path that stays inside the repository.

    Raises:
        ValueError: If the path escapes the repository or enters an excluded directory.
    """
    requested = (REPO_ROOT / path_str).resolve()
    try:
        requested.relative_to(REPO_ROOT)
    except ValueError:
        raise ValueError(f"Path '{path_str}' is outside the repository")
    if _is_excluded(requested.relative_to(REPO_ROOT)):
        raise ValueError(f"Path '{path_str}' is in an excluded directory")
    return requested


def list_files(glob_pattern: str = "**/*") -> str:
    """List repository files that match a glob pattern.

    Args:
        glob_pattern: Glob pattern to evaluate against files under ``REPO_ROOT``.

    Returns:
        str: A newline-delimited listing of matching files or a short status message.
    """
    matches = sorted(
        str(p.relative_to(REPO_ROOT))
        for p in REPO_ROOT.glob(glob_pattern)
        if p.is_file() and not _is_excluded(p.relative_to(REPO_ROOT))
    )
    if not matches:
        return f"No files matching '{glob_pattern}'"
    # Limit output to avoid flooding the context
    if len(matches) > 50:
        return "\n".join(matches[:50]) + f"\n... and {len(matches) - 50} more files"
    return "\n".join(matches)


def search_text(query: str, glob_pattern: str = None) -> str:
    """Search repository files for a text pattern.

    Args:
        query: Text pattern to pass to ``grep``.
        glob_pattern: Optional file glob to narrow the search scope.

    Returns:
        str: Matching lines with file and line context, or a status message.
    """
    exclude_args = []
    for d in EXCLUDED_DIRS:
        exclude_args.extend(["--exclude-dir", d])
    cmd = ["grep", "-rn", "--include=*.py"] + exclude_args + [query, "."]
    if glob_pattern:
        cmd = ["grep", "-rn", f"--include={glob_pattern}"] + exclude_args + [query, "."]
    try:
        result = subprocess.run(cmd, capture_output=True, text=True, timeout=10, cwd=REPO_ROOT)
        lines = result.stdout.strip().split("\n")
        if not lines or lines == [""]:
            return f"No matches for '{query}'"
        if len(lines) > 30:
            return "\n".join(lines[:30]) + f"\n... and {len(lines) - 30} more matches"
        return "\n".join(lines)
    except subprocess.TimeoutExpired:
        return "Search timed out"


def read_file(path: str, start_line: int = None, end_line: int = None) -> str:
    """Read a repository file, optionally trimming to a line range.

    Args:
        path: Repo-relative file path to read.
        start_line: Optional 1-based line number to start from.
        end_line: Optional inclusive line number to stop at.

    Returns:
        str: File contents or a short error/status message safe for the model to inspect.
    """
    file_path = validate_path(path)
    if not file_path.exists():
        return f"File not found: {path}"
    if not file_path.is_file():
        return f"Not a file: {path}"
    text = file_path.read_text()
    lines = text.split("\n")
    if start_line is not None or end_line is not None:
        start = max(0, (start_line or 1) - 1)
        end = end_line or len(lines)
        lines = lines[start:end]
    # Limit output size
    if len(lines) > 200:
        return "\n".join(lines[:200]) + f"\n... truncated ({len(lines)} total lines)"
    return "\n".join(lines)


# Registry for dispatching tool calls
TOOL_FUNCTIONS = {
    "list_files": list_files,
    "search_text": search_text,
    "read_file": read_file,
}

Notice the security patterns from Module 1: validate_path uses relative_to to prevent path traversal, outputs are size-limited to avoid flooding the model's context, and there's a timeout on the subprocess call.

Define tool schemas

Create the JSON schemas that tell the model what tools are available:

# agent/schemas.py
"""Tool schemas for the model API."""

TOOL_SCHEMAS = [
    {
        "type": "function",
        "function": {
            "name": "list_files",
            "description": "List files in the repository matching a glob pattern. Use this to explore the repo structure before reading specific files.",
            "parameters": {
                "type": "object",
                "properties": {
                    "glob_pattern": {
                        "type": "string",
                        "description": "Glob pattern to match files, e.g. '**/*.py', 'src/**/*.ts', 'tests/*'. Defaults to all files.",
                    }
                },
                "required": [],
            },
        },
    },
    {
        "type": "function",
        "function": {
            "name": "search_text",
            "description": "Search for a text pattern in repository files. Returns matching lines with file path and line number. Use this to find where something is defined or used.",
            "parameters": {
                "type": "object",
                "properties": {
                    "query": {
                        "type": "string",
                        "description": "Text to search for (passed to grep)",
                    },
                    "glob_pattern": {
                        "type": "string",
                        "description": "Optional file pattern to restrict search, e.g. '*.py', '*.md'",
                    },
                },
                "required": ["query"],
            },
        },
    },
    {
        "type": "function",
        "function": {
            "name": "read_file",
            "description": "Read the contents of a file in the repository. Optionally specify a line range to read a specific section.",
            "parameters": {
                "type": "object",
                "properties": {
                    "path": {
                        "type": "string",
                        "description": "File path relative to the repository root, e.g. 'src/main.py'",
                    },
                    "start_line": {
                        "type": "integer",
                        "description": "First line to read (1-based). Omit to start from the beginning.",
                    },
                    "end_line": {
                        "type": "integer",
                        "description": "Last line to read. Omit to read to the end.",
                    },
                },
                "required": ["path"],
            },
        },
    },
]

Having good tool descriptions is important. The model uses these descriptions to decide when to call a tool and what arguments to provide. "List files in the repository" is more useful than "List files." It tells the model what the tool operates on.

Build the control loop

This is the core of the agent. It's a loop that:

  1. Sends the conversation to the model with tool schemas
  2. Checks if the model wants to call a tool
  3. If yes: executes the tool, appends the result, and loops back to step 1
  4. If no: returns the model's final answer

Pick your provider for the complete agent/loop.py:

# agent/loop.py
"""Raw tool-calling control loop."""
import json
import sys
from openai import OpenAI
from agent.schemas import TOOL_SCHEMAS
from agent.tools import TOOL_FUNCTIONS

client = OpenAI()

SYSTEM_PROMPT = """You are a code assistant for this repository. Answer questions by using the available tools to explore the codebase.

Rules:
- Use list_files to understand the repo structure before diving into specific files.
- Use search_text to find where things are defined or used.
- Use read_file to examine specific code.
- Base your answers on what you find in the code, not on assumptions.
- If you can't find enough evidence, say so rather than guessing.
- When you have enough information, provide your answer with file references."""

MAX_TOOL_ROUNDS = 10


def run_agent(question: str, model: str = "gpt-4o-mini", verbose: bool = True) -> dict:
    """Run one question through the raw tool-calling control loop.

    Args:
        question: User question the agent should answer about the repository.
        model: Provider-specific model identifier to call.
        verbose: Whether to print each tool invocation while the loop runs.

    Returns:
        dict: Final answer text plus tool-call metadata and loop outcome details.
    """
    messages = [
        {"role": "system", "content": SYSTEM_PROMPT},
        {"role": "user", "content": question},
    ]
    tool_calls_made = []

    for round_num in range(MAX_TOOL_ROUNDS):
        response = client.chat.completions.create(
            model=model,
            messages=messages,
            tools=TOOL_SCHEMAS,
            temperature=0,
        )

        msg = response.choices[0].message

        if not msg.tool_calls:
            if verbose:
                print(f"  [{round_num + 1} rounds, {len(tool_calls_made)} tool calls]")
            return {
                "answer": msg.content,
                "tool_calls": tool_calls_made,
                "rounds": round_num + 1,
                "finish_reason": response.choices[0].finish_reason,
            }

        messages.append(msg)
        for call in msg.tool_calls:
            fn_name = call.function.name
            try:
                args = json.loads(call.function.arguments)
            except json.JSONDecodeError:
                args = {}

            if verbose:
                print(f"  Tool: {fn_name}({', '.join(f'{k}={v!r}' for k, v in args.items())})")

            if fn_name not in TOOL_FUNCTIONS:
                result = f"Rejected: '{fn_name}' is not a registered tool."
            else:
                try:
                    result = TOOL_FUNCTIONS[fn_name](**args)
                except (ValueError, TypeError) as e:
                    result = f"Validation error: {e}"
                except Exception as e:
                    result = f"Error: {e}"

            tool_calls_made.append({
                "tool": fn_name,
                "args": args,
                "result_preview": result[:200] if len(result) > 200 else result,
            })

            messages.append({
                "role": "tool",
                "tool_call_id": call.id,
                "content": result,
            })

    return {
        "answer": "Reached maximum tool rounds without a final answer.",
        "tool_calls": tool_calls_made,
        "rounds": MAX_TOOL_ROUNDS,
        "finish_reason": "max_rounds",
    }


if __name__ == "__main__":
    question = sys.argv[1] if len(sys.argv) > 1 else "What are the main modules in this repository?"
    print(f"Question: {question}\n")
    result = run_agent(question)
    print(f"\nAnswer:\n{result['answer']}")

Run it:

python -m agent.loop "What are the main modules in this repository?"

You should see the model calling list_files to explore the structure, possibly search_text to find specific patterns, and then producing an answer based on what it found. Watch the tool calls. This is the agent reasoning in real time.

Note that the Anthropic version includes its own tool schemas inline because Anthropic uses a different schema format (input_schema instead of parameters, no type: "function" wrapper). The OpenAI, Hugging Face, and Ollama versions all share agent/schemas.py.

Run your benchmark through the agent

Now the real test. Run your benchmark questions through the agent and compare against your Module 2 baseline:

# agent/run_benchmark.py
"""Run benchmark questions through the tool-calling agent."""
import json
import os
from datetime import datetime, timezone
from agent.loop import run_agent

RUN_ID = "agent-v1-" + datetime.now(timezone.utc).strftime("%Y-%m-%d-%H%M%S")
MODEL = "gpt-4o-mini"
PROVIDER = "openai"
BENCHMARK_FILE = "benchmark-questions.jsonl"
OUTPUT_FILE = f"harness/runs/{RUN_ID}.jsonl"
REPO_SHA = os.popen("git rev-parse --short HEAD").read().strip()

questions = []
with open(BENCHMARK_FILE) as f:
    for line in f:
        if line.strip():
            questions.append(json.loads(line))

print(f"Running {len(questions)} benchmark questions")
print(f"Run ID: {RUN_ID}")
print(f"Model: {MODEL}\n")

results = []
for i, q in enumerate(questions):
    print(f"[{i+1}/{len(questions)}] {q['category']}: {q['question'][:60]}...")
    result = run_agent(q["question"], model=MODEL, verbose=True)

    entry = {
        "run_id": RUN_ID,
        "question_id": q["id"],
        "question": q["question"],
        "category": q["category"],
        "answer": result["answer"],
        "model": MODEL,
        "provider": PROVIDER,
        "evidence_files": list(set(
            tc["args"].get("path", "") for tc in result["tool_calls"]
            if tc["tool"] == "read_file"
        )),
        "tools_called": [tc["tool"] for tc in result["tool_calls"]],
        "retrieval_method": "tool_calling",
        "grade": None,
        "failure_label": None,
        "grading_notes": "",
        "repo_sha": REPO_SHA,
        "timestamp": datetime.now(timezone.utc).isoformat(),
        "harness_version": "v0.2",
    }
    results.append(entry)
    print()

os.makedirs("harness/runs", exist_ok=True)
with open(OUTPUT_FILE, "w") as f:
    for entry in results:
        f.write(json.dumps(entry) + "\n")

print(f"Done. {len(results)} results saved to {OUTPUT_FILE}")
print("Next: grade these answers and compare against your baseline.")
python -m agent.run_benchmark

After grading (using the same grade_baseline.py from Module 2), compare the numbers:

python harness/summarize_run.py harness/runs/agent-v1-*-graded.jsonl

This is the first real comparison your harness enables. How much did tool access improve over the training-data-only baseline? Which categories improved most? Which failure labels shifted from missing_evidence to something more specific?

Exercises

  1. Build the three tools in agent/tools.py. Test each one independently before wiring them into the loop.
  2. Build the control loop in agent/loop.py. Run it against 3-5 questions manually and observe the tool call sequences.
  3. Run your full benchmark through the agent using run_benchmark.py.
  4. Grade at least 15 answers and compare against your Module 2 baseline. Which categories improved? What's the new failure distribution?
  5. Identify one question where the agent called the right tools but still got the wrong answer. What went wrong? Was it a tool output problem, a reasoning problem, or a context problem?

Reflection prompts

  • Which errors came from the model's reasoning (it had the right evidence but drew the wrong conclusion)?
  • Which came from the tool interface (bad arguments, missing tools, truncated output)?
  • Which came from oversized tool outputs flooding the context?
  • Which came from missing retrieval (the tools didn't surface the right code)?

Completion checkpoint

You have:

  • Three working tools (list_files, search_text, read_file) with input validation
  • A working control loop that runs to completion without infinite looping
  • A benchmark run graded and compared against your Module 2 baseline
  • An understanding of where tool access helps and where it's still insufficient

Connecting to the project

This raw loop is the mechanical foundation for everything in this module. In the next lesson, you'll rebuild it using a framework and see what the framework gives you (state management, easier composition) and what it hides (the loop mechanics you now understand).

The tools you built here will also evolve. In Module 4 they'll be joined by retrieval tools, and in the next two lessons they'll become portable MCP capabilities that any client can use.

What's next

Rebuilding with a Framework. Build it once by hand first; then the next lesson makes the framework tradeoffs legible because you know what the abstraction is carrying for you.

References

Start here

Build with this

Deep dive

Your Notes
GitHub Sync

Sync your lesson notes to a private GitHub Gist. If you have not entered a token yet, the sync button will open the GitHub token modal.

Glossary
API (Application Programming Interface)Foundational terms
A structured way for programs to communicate. In this context, usually an HTTP endpoint you call to interact with an LLM.
AST (Abstract Syntax Tree)Foundational terms
A tree representation of source code structure. Used by parsers like Tree-sitter to understand code as a hierarchy of functions, classes, and statements. You'll encounter this more deeply in the Code Retrieval module, but the concept appears briefly in retrieval fundamentals.
BM25 (Best Match 25)Foundational terms
A classical ranking function for keyword search. Scores documents by term frequency and inverse document frequency. Often competitive with or complementary to vector search.
ChunkingFoundational terms
Splitting a document into smaller pieces for indexing and retrieval. Chunk boundaries significantly affect retrieval quality. Split at the wrong place and your retrieval will return half a function or the end of one paragraph glued to the start of another.
Context engineeringFoundational terms
The discipline of selecting, packaging, and budgeting the information a model sees at inference time. Prompts, retrieved evidence, tool results, memory, and state are all parts of context. Context engineering is arguably the core skill of AI engineering. Bigger context windows are not a substitute for better context selection.
Context rotFoundational terms
Degradation of output quality caused by stale, noisy, or accumulated context. Symptoms include stale memory facts, conflicting retrieved evidence, bloated prompt history, and accumulated instructions that contradict each other. A form of technical debt in AI systems.
Context windowFoundational terms
The maximum number of tokens an LLM can process in a single request (input + output combined).
EmbeddingFoundational terms
A fixed-length numeric vector representing a piece of text. Used for similarity search: texts with similar meanings have nearby embeddings.
EndpointFoundational terms
A specific URL path that accepts requests and returns responses (e.g., POST /v1/chat/completions).
GGUFFoundational terms
A file format for quantized models used by llama.cpp and Ollama. When you see a model name like qwen2.5:7b-q4_K_M, the suffix indicates the quantization scheme. GGUF supports mixed quantization (different precision for different layers) and is the most common format for local inference.
HallucinationFoundational terms
When a model generates content that sounds confident but isn't supported by the evidence it was given, or fabricates details that don't exist. Not the same as "any wrong answer"; a model that misinterprets ambiguous instructions gave a bad answer but didn't hallucinate. Common causes: weak prompt, missing context, context rot, model limitation, or retrieval failure.
InferenceFoundational terms
Running a trained model to generate output from input. What happens when you call an API. Most AI engineering work is inference-time work: building systems around models, not training them. Use "inference," not "inferencing."
JSON (JavaScript Object Notation)Foundational terms
A lightweight text format for structured data. The lingua franca of API communication.
Lexical searchFoundational terms
Finding items by matching keywords or terms. Includes BM25, TF-IDF (Term Frequency–Inverse Document Frequency), and simple keyword matching. Returns exact term matches, not semantic similarity.
LLM (Large Language Model)Foundational terms
A neural network trained on large text corpora that generates text by predicting the next token. The core technology behind AI engineering; every tool, pattern, and pipeline in this curriculum runs on top of one.
MetadataFoundational terms
Structured information about a document or chunk (file path, language, author, date, symbol type). Used for filtering retrieval results.
Neural networkFoundational terms
A computing system loosely inspired by biological neurons, built from layers of mathematical functions that transform inputs into outputs. LLMs are a specific type of neural network (transformers) trained on text. You don't need to understand neural network internals to do AI engineering, but knowing the term helps when reading external resources.
Reasoning modelFoundational terms
A model optimized for complex multi-step planning, math, and logic (e.g., o3, o4-mini). Slower and more expensive but better on hard problems. Sometimes called "LRM" (large reasoning model), but "reasoning model" is the more consistent term across provider docs.
RerankingFoundational terms
A second-pass scoring step that re-orders retrieved results using a more expensive model. Improves precision after an initial broad retrieval.
SchemaFoundational terms
A formal description of the shape and types of a data structure. Used to validate inputs and outputs.
SLM (small language model)Foundational terms
A compact model (typically 1-7B parameters) that runs on consumer hardware with lower cost, latency, and better privacy (e.g., Phi, small Llama variants, Gemma). The right choice when privacy, offline operation, predictable cost, or low latency matter more than peak capability.
System promptFoundational terms
A special message that sets the model's behavior, role, and constraints for a conversation.
TemperatureFoundational terms
A parameter controlling output randomness. Lower values produce more deterministic output; higher values produce more varied output. Does not affect the model's intelligence.
TokenFoundational terms
The basic unit an LLM processes. Not a word. Tokens are sub-word fragments. "unhappiness" might be three tokens: "un", "happi", "ness". Token count determines cost and context window usage.
Top-kFoundational terms
The number of results returned from a retrieval query. "Top-5" means the five highest-scoring results.
Top-p (nucleus sampling)Foundational terms
An alternative to temperature for controlling output diversity. Selects from the smallest set of tokens whose cumulative probability exceeds p.
Vector searchFoundational terms
Finding items by proximity in embedding space (nearest neighbors). Returns "similar" results, not "exact match" results.
vLLM (virtual LLM)Foundational terms
An inference serving engine (not a model) that hosts open-weight models behind an OpenAI-compatible HTTP endpoint. Infrastructure layer, not model layer. Relevant when moving from hosted APIs to self-hosting.
WeightsFoundational terms
The learned parameters inside a model. Changed during training, fixed during inference.
Workhorse modelFoundational terms
A general-purpose LLM optimized for speed and broad capability (e.g., GPT-4o-mini, Claude Haiku, Gemini Flash). The default for most tasks. When someone says "LLM" without qualification, they usually mean this.
BaselineBenchmark and Harness terms
The first measured performance of your system on a benchmark. Everything else is compared against this. Without a baseline, you can't tell whether a change helped.
BenchmarkBenchmark and Harness terms
A fixed set of questions or tasks with known-good answers, used to measure system performance over time.
Run logBenchmark and Harness terms
A structured record (typically JSONL) of every system run: what input was given, what output was produced, what tools were called, how long it took, and what it cost. The raw data that evals, telemetry, and cost analysis are built from.
A2A (Agent-to-Agent protocol)Agent and Tool Building terms
An open protocol for peer-to-peer agent collaboration. Agents discover each other's capabilities and delegate or negotiate tasks as equals. Different from MCP (which connects agents to tools, not to other agents) and from handoffs (which transfer control within one system).
AgentAgent and Tool Building terms
A system where an LLM decides which tools to call, observes results, and iterates until a task is complete. Agent = model + tools + control loop.
Control loopAgent and Tool Building terms
The code that manages the agent's cycle: send prompt, check for tool calls, execute tools, append results, repeat or finish.
HandoffAgent and Tool Building terms
Passing control from one agent or specialist to another within an orchestrated system.
MCP (Model Context Protocol)Agent and Tool Building terms
An open protocol for exposing tools, resources, and prompts to AI applications in a standardized way. Connects agents to capabilities (tools and data), not to other agents.
Tool calling / function callingAgent and Tool Building terms
The model's ability to request execution of a specific function with structured arguments, rather than just generating text.
Context compilation / context packingCode Retrieval terms
The process of selecting and assembling the smallest useful set of evidence for a specific task. Not "dump everything retrieved into the prompt."
GroundingCode Retrieval terms
Tying model assertions to specific evidence. A grounded answer cites what it found; an ungrounded answer asserts without evidence.
Hybrid retrievalCode Retrieval terms
Combining multiple retrieval methods (e.g., vector search + keyword search + metadata filters) and merging or reranking the results.
Knowledge graphCode Retrieval terms
A data structure that stores entities and their relationships explicitly (e.g., "function A calls function B," "module X imports module Y"). Useful for traversal and dependency reasoning. One retrieval strategy among several, often overused when simpler metadata or adjacency tables would suffice.
RAG (Retrieval-Augmented Generation)Code Retrieval terms
A pattern where the model's response is grounded in retrieved external evidence rather than relying solely on its training data.
Symbol tableCode Retrieval terms
A mapping of code identifiers (functions, classes, variables) to their locations and metadata.
Tree-sitterCode Retrieval terms
An incremental parsing library that builds ASTs for source code. Used in this curriculum for code-aware chunking and symbol extraction.
Context packRAG and Grounded Answers terms
A structured bundle of evidence assembled for a specific task, with metadata about provenance, relevance, and token budget.
Evidence bundleRAG and Grounded Answers terms
A collection of retrieved items grouped for a specific sub-task, with enough metadata to evaluate whether the evidence is relevant and sufficient.
Retrieval routingRAG and Grounded Answers terms
Deciding which retrieval strategy or method to use for a given query. Different questions need different retrieval methods.
EvalObservability and Evals terms
A structured test that measures system quality. Not the same as training. Evals measure, they don't change the model.
Harness (AI harness / eval harness)Observability and Evals terms
The experiment and evaluation framework around your model or agent. It runs benchmark tasks, captures outputs, logs traces, grades results, and compares system versions. It turns ad hoc "try it and see" into repeatable, comparable experiments. Typically includes: input dataset, prompt and tool configuration, model/provider selection, execution loop, logging, grading, and artifact capture.
LLM-as-judgeObservability and Evals terms
Using a language model to evaluate or grade the output of another model or system. Useful for scaling evaluation beyond manual review, but requires rubric quality, judge consistency checks, and human spot-checking. Not a replacement for exact-match checks where they apply.
OpenTelemetry (OTel)Observability and Evals terms
An open standard for collecting and exporting telemetry data (traces, metrics, logs). Vendor-agnostic.
RAGASObservability and Evals terms
A specific eval framework for retrieval-augmented generation. Measures metrics like faithfulness, relevance, and context precision. One tool example, not a foundational concept. Learn the metrics first, then the tool.
SpanObservability and Evals terms
A single operation within a trace (e.g., one tool call, one retrieval query). Traces are made of spans.
TelemetryObservability and Evals terms
Structured data about system behavior: what happened, when, how long it took, what it cost. Includes traces, metrics, and events.
TraceObservability and Evals terms
A structured record of one complete run through the system, including all steps, tool calls, and decisions.
Long-term memoryOrchestration and Memory terms
Persistent facts that survive across conversations. Requires write policies to manage what gets stored, updated, or deleted.
OrchestrationOrchestration and Memory terms
Explicit control over how tasks are routed, delegated, and synthesized across multiple agents or specialists.
RouterOrchestration and Memory terms
A component that decides which specialist or workflow path to use for a given query.
SpecialistOrchestration and Memory terms
An agent or workflow tuned for a narrow task (e.g., "code search," "documentation lookup," "test generation"). Specialists are composed by an orchestrator.
Thread memoryOrchestration and Memory terms
Conversation state that persists within a single session or thread.
Workflow memoryOrchestration and Memory terms
Intermediate state that persists within a multi-step task but doesn't survive beyond the workflow's completion.
Catastrophic forgettingOptimization terms
When fine-tuning causes a model to lose capabilities it had before training. The model gets better at the fine-tuned task but worse at tasks it previously handled. PEFT methods like LoRA reduce this risk by freezing original weights.
DistillationOptimization terms
Training a smaller (student) model to reproduce the behavior of a larger (teacher) model on a specific task.
DPO (Direct Preference Optimization)Optimization terms
A method for preference-based model optimization that's simpler than RLHF, training the model directly on preference pairs without a separate reward model.
Fine-tuningOptimization terms
Updating a model's weights on task-specific data to change its behavior permanently. An umbrella term that includes SFT, instruction tuning, RLHF, DPO, and other techniques. See the fine-tuning landscape table in Lesson 8.3 for how these relate.
Full fine-tuningOptimization terms
Updating all of a model's parameters during training, as opposed to PEFT methods that update only a small subset. Requires significantly more GPU memory and compute. Produces the most thorough adaptation but carries higher risk of catastrophic forgetting.
Inference serverOptimization terms
Software (like vLLM or Ollama) that hosts a model and serves inference requests.
Instruction tuningOptimization terms
A specific application of SFT where the training data consists of instruction-response pairs. This is how base models become chat models: the technique is SFT, the data format is instructions. Not a separate technique from SFT.
LoRA (Low-Rank Adaptation)Optimization terms
A parameter-efficient fine-tuning method that trains small adapter matrices instead of updating all model weights. Dramatically reduces GPU memory and compute requirements.
Parameter countOptimization terms
The number of learned weights in a model, commonly expressed in billions (e.g., "7B" = 7 billion parameters). Determines memory requirements (roughly 2 bytes per parameter at FP16) and broadly correlates with capability, though training quality and architecture matter as much as size. See Model Selection and Serving for sizing guidance.
PEFT (Parameter-Efficient Fine-Tuning)Optimization terms
A family of methods (including LoRA) that fine-tune a small subset of parameters instead of the full model.
Preference optimizationOptimization terms
Training methods (RLHF, DPO) that use human or automated preference signals to improve model behavior. "This output is better than that output" rather than "this is the correct output."
QLoRA (Quantized LoRA)Optimization terms
LoRA applied to a quantized (compressed) base model. Further reduces memory requirements, enabling fine-tuning on consumer hardware.
QuantizationOptimization terms
Reducing the precision of model weights (e.g., FP16 → INT4) to shrink memory usage and increase inference speed at some quality cost. A 7B model at FP16 needs ~14 GB VRAM; quantized to 4-bit, it fits in ~4 GB. Common formats include GGUF (llama.cpp/Ollama), GPTQ and AWQ (vLLM/HuggingFace). See Model Selection and Serving for format details and tradeoffs.
OverfittingOptimization terms
When a model memorizes training examples instead of learning generalizable patterns. The model performs well on training data but poorly on new inputs. Detected by monitoring validation loss alongside training loss.
RLHF (Reinforcement Learning from Human Feedback)Optimization terms
A training method that uses human preference signals to improve model behavior through a reward model. More complex than DPO (requires training a separate reward model) but offers more control over the optimization objective.
SFT (Supervised Fine-Tuning)Optimization terms
Fine-tuning using input-output pairs where the desired output is known. The most common fine-tuning approach.
TRL (Transformer Reinforcement Learning)Optimization terms
A Hugging Face library for training language models with reinforcement learning, SFT, and other optimization methods.
Consumer chat appCross-cutting terms
The browser or desktop product meant for human conversation (ChatGPT, Claude, HuggingChat). Useful for experimentation, but not the same as API access.
Developer platformCross-cutting terms
The provider's API, billing, API-key, and developer-docs surface. This is what you need for this learning path.
Hosted APICross-cutting terms
The provider runs the model for you and you call it over HTTP.
Local inferenceCross-cutting terms
You run the model on your own machine.
ProviderCross-cutting terms
The company or service that hosts a model API you call from code.
Prompt cachingCross-cutting terms
Reusing computation from repeated prompt prefixes to reduce latency and cost on subsequent requests with the same prefix.
Rate limitingCross-cutting terms
Constraints on how many API requests you can make per unit of time. An operational concern that affects system design and cost.
Token budgetCross-cutting terms
The maximum number of tokens you allocate for a specific part of the context (e.g., "retrieval evidence gets at most 4K tokens"). A context engineering tool for preventing any single component from dominating the context window.