8.0 Learning Checklist: LLM Application Development and RAG
Use this page as a printable checklist. If you need the full explanation, return to the Chapter 8 entry page.

Two-Hour First Pass
| Time box | Do this | Stop when you can say |
|---|---|---|
| 20 min | Read the RAG application loop on the entry page | "A RAG answer should be tied to retrieved evidence." |
| 25 min | Run the tiny RAG script | "I can inspect top-k chunks before trusting the answer." |
| 25 min | Skim 8.1 RAG basics and document processing | "Chunk size, overlap, and metadata affect retrieval and citations." |
| 25 min | Skim 8.3 API practice and tool/function calling | "An LLM app needs request, response, error, and retry paths." |
| 25 min | Read the debugging ladder | "I can separate document, retrieval, generation, citation, and ops failures." |
Required Evidence
| Evidence | Minimum version |
|---|---|
chunks.jsonl | 5-10 chunks with id, source, text, and version |
retrieval_logs.jsonl | query, top-k chunk IDs, score, and source for each test question |
eval_questions.csv | at least 10 fixed questions with expected source or answer points |
failure_cases.md | at least three failures labeled as document, chunking, retrieval, generation, citation, or deploy |
rag_config.md | chunk size, overlap, top-k, rerank choice, prompt version |
rag_app_workshop_output.txt | output from 8.5.6 Hands-on: Full Chapter 8 RAG App Workshop |
README.md | run command, sample question, cited answer, evaluation result, next fix |
Quality Gates
| Gate | Pass condition |
|---|---|
| Citation | Every factual answer cites a chunk, source, and version. |
| Empty retrieval | System refuses to answer when evidence is missing. |
| Regression eval | Same questions run before and after each chunking, retrieval, reranking, or prompt change. |
| Operations | Logs include query, top-k, prompt version, latency, token cost, and failure label. |
Exit Questions
- Can you explain why RAG is different from asking a longer Prompt?
- Can you show which document chunks were retrieved for a question?
- Can you explain why chunk metadata is necessary for citations and debugging?
- Can you handle empty retrieval with a no-answer response instead of a guess?
- Can you compare two RAG versions using the same evaluation questions?
If the answer is yes, move to Chapter 9. Chapter 9 will upgrade the system from answer generation to Agents that can plan, call tools, and recover from failures.