Skip to main content

8.0 Learning Checklist: LLM Application Development and RAG

Use this page as a printable checklist. If you need the full explanation, return to the Chapter 8 entry page.

RAG portfolio evidence pack

Two-Hour First Pass

Time boxDo thisStop when you can say
20 minRead the RAG application loop on the entry page"A RAG answer should be tied to retrieved evidence."
25 minRun the tiny RAG script"I can inspect top-k chunks before trusting the answer."
25 minSkim 8.1 RAG basics and document processing"Chunk size, overlap, and metadata affect retrieval and citations."
25 minSkim 8.3 API practice and tool/function calling"An LLM app needs request, response, error, and retry paths."
25 minRead the debugging ladder"I can separate document, retrieval, generation, citation, and ops failures."

Required Evidence

EvidenceMinimum version
chunks.jsonl5-10 chunks with id, source, text, and version
retrieval_logs.jsonlquery, top-k chunk IDs, score, and source for each test question
eval_questions.csvat least 10 fixed questions with expected source or answer points
failure_cases.mdat least three failures labeled as document, chunking, retrieval, generation, citation, or deploy
rag_config.mdchunk size, overlap, top-k, rerank choice, prompt version
rag_app_workshop_output.txtoutput from 8.5.6 Hands-on: Full Chapter 8 RAG App Workshop
README.mdrun command, sample question, cited answer, evaluation result, next fix

Quality Gates

GatePass condition
CitationEvery factual answer cites a chunk, source, and version.
Empty retrievalSystem refuses to answer when evidence is missing.
Regression evalSame questions run before and after each chunking, retrieval, reranking, or prompt change.
OperationsLogs include query, top-k, prompt version, latency, token cost, and failure label.

Exit Questions

  • Can you explain why RAG is different from asking a longer Prompt?
  • Can you show which document chunks were retrieved for a question?
  • Can you explain why chunk metadata is necessary for citations and debugging?
  • Can you handle empty retrieval with a no-answer response instead of a guess?
  • Can you compare two RAG versions using the same evaluation questions?

If the answer is yes, move to Chapter 9. Chapter 9 will upgrade the system from answer generation to Agents that can plan, call tools, and recover from failures.