Saltar a contenido

AI Systems From Scratch - Beginner to OSS Contributor

From "I've heard of AI / LLMs" to "I can train a small model, fine-tune a transformer, build a small RAG app, evaluate it honestly, and submit a fix to an AI-adjacent OSS project."

Who this is for

  • You've finished Python From Scratch (or you're comfortable enough in Python to write small programs).
  • You've never trained a model, OR you've copy-pasted some PyTorch / Hugging Face code without really understanding what it does.

Soft prerequisite

Python comfort is mandatory - AI tooling lives in Python. If you can't write a function, walk a list, and read a stack trace, do Python From Scratch first.

You do not need a PhD in math. We use linear algebra at the level of "dot products and matrices"; we explain everything else as it appears.

What you'll need

  • A computer. A GPU helps a lot but isn't required for the first 8 pages. Free options for hands-on GPU work: Google Colab (free tier), Kaggle Notebooks, Lambda Labs (cheap), AWS / GCP (paid).
  • Python ≥3.10 (you set this up in the Python beginner path).
  • A text editor.
  • About 5 hours/week. Path is sized for 4-6 months.

Why AI systems

  • Biggest growth area in software. The job market and OSS activity around LLMs / ML infra is the most active it's ever been.
  • OSS is the heart of the field. PyTorch, Hugging Face, vLLM, llama.cpp, Ollama, LangChain - all open-source and welcoming.
  • The barrier is lower than it looks. Modern tooling lets you fine-tune real models with ~50 lines of code; serve them with another ~30.

How this path works

Same template as the other beginner paths: one concept per page, code first then walkthrough, exercise, Q&A, done recap.

We use PyTorch as the framework throughout - most popular, best ecosystem. Hugging Face Transformers for pre-trained models. vLLM / Ollama / llama.cpp for inference.

The pages

# Title What you'll know after
00 Introduction What we're doing and why
01 Setup Python + PyTorch + CUDA (or CPU) working
02 Tensors PyTorch's central data type
03 Linear algebra you actually need Dot products, matmul, gradients (intuitive)
04 Your first neural network A small MLP from scratch
05 Training loop Loss, optimizer, gradient descent
06 Inference and saving Loading a pretrained model, running it
07 Transformers and tokenization What an LLM actually does
08 Hugging Face Transformers Pre-trained models in 3 lines
09 Fine-tuning Adapt a model to your data (LoRA-friendly)
10 Retrieval-Augmented Generation Embeddings + vector DB + LLM
11 Evaluation The hardest part of ML done seriously
12 Serving models vLLM, Ollama, simple HTTP wrappers
13 Picking a project AI-OSS candidates
14 Anatomy of an AI OSS project Case study
15 Your first contribution Workflow + PR

Start with Introduction.

Print this path

Want to read offline or archive? Open the printable version - every section of this path concatenated into one page, styled for paper. Use your browser's Print → Save as PDF.