Skip to main content

0.3 AI Full-Stack Capability Map

AI Full-Stack Capability Map

Read the picture first. The course is one path:

tools -> Python -> data -> models -> LLM -> RAG -> Agent -> specialization/delivery

You do not need every detail now. Just remember:

If you are blocked by...Go back to...
running codetools and Python
messy inputsdata
unreliable answersevaluation and RAG
uncontrolled actionsAgent traces and permissions

The Seven Layers

LayerCourse chaptersFirst visible evidenceDeeper question
Tools1A reproducible project folder and Git historyCan another person rerun it?
Python2Small scripts with clear inputs and outputsIs the code readable, typed, and testable?
Data3Clean tables, charts, and notesDo you know where the data is wrong or biased?
Models4-6Trained or inspected model experimentsWhat metric would change your decision?
LLM7Prompt, tokens, embeddings, Transformer intuitionWhich behavior comes from data, decoding, or context?
RAG8Retrieval trace and answer evaluationDid the answer use the right evidence?
Agent9Tool traces, permissions, memory notes, deployment notesWhat can fail when users, files, and actions are real?
Specialization / delivery10-12 and electivesVision/NLP/multimodal demos, exported assets, deployment notesWhich domain constraints change the product decision?

The course is not a pile of topics. It is a debugging stack. When an AI application behaves badly, the cause may live several layers below the feature you are looking at.

Main Line And Expansion Tracks

Use Chapters 1-9 as the default main line. After Chapter 9, you should be able to build a small LLM/RAG/Agent project with evidence, logs, and a safety boundary.

Then choose Chapters 10-12 by product need:

NeedChooseWhy
Images, cameras, OCR, detection, segmentationChapter 10 Computer VisionThe output is visual: labels, boxes, masks, text, or video events
Text labels, extraction, summaries, linguistic evaluationChapter 11 NLPThe output is a text task with labels, fields, spans, or generated text
Images, PDFs, audio, video, creative assets, multimodal RAGChapter 12 Multimodal/AIGCThe workflow mixes modalities and needs source, prompt, review, and export records
Deployment, advanced Python, classic ML depthElectivesThe main project needs a specific engineering or algorithmic side skill

How To Use The Map

Before starting a project, mark the highest-risk layer. For example, a PDF question-answering app usually fails first in data cleaning and retrieval, not in the chat UI. An automation agent usually fails first in tool permissions, state, and evaluation, not in the prompt wording.

During each chapter, keep one artifact that proves the layer works. Screenshots are useful, but logs, README commands, small datasets, metric tables, and failure notes are stronger because they help you debug later.

Optional background: if you want the history behind these layers, skim the 15-stage AI development map.

Next, choose a learning path.