Skip to main content

7.2.1 LLM Overview Roadmap: Capability, Cost, Product Fit

LLM overview is not a model-name list. It helps you decide what a large model can do, what it costs, and when prompting, RAG, Agent, or fine-tuning is a better route.

Look at the Capability Stack First

LLM overview chapter relationship diagram

Large model capability stack and application ecosystem diagram

RouteUse when...
promptthe model already knows enough and task is simple
RAGprivate or changing knowledge must be cited
Agentthe model must use tools or take steps
fine-tuningbehavior/style/format needs repeated adaptation

Run One Route Decision

request = {
"needs_private_docs": True,
"needs_tool_action": False,
"needs_repeated_style": False,
}

if request["needs_tool_action"]:
route = "Agent"
elif request["needs_private_docs"]:
route = "RAG"
elif request["needs_repeated_style"]:
route = "fine-tuning"
else:
route = "prompt"

print("recommended_route:", route)

Expected output:

recommended_route: RAG

LLM route decision run result map

This is not a full architecture decision. It is the habit: choose the smallest route that solves the actual product need.

Learn in This Order

OrderReadWhat to keep
17.2.2 Development Historywhy scaling and instruction tuning mattered
27.2.3 Core Conceptscontext, tokens, temperature, latency, cost
37.2.4 Industry Landscapemodel/provider selection notes
47.2.5 LLM Call Workbenchone request/response record

Pass Check

You pass this roadmap when you can explain one model choice in terms of capability, context, cost, latency, data privacy, and route fit.