LLM cost optimization evolved from ad-hoc tricks to systematic infrastructure. Semantic caching stores common responses. LLMLingua compresses prompts 20x while preserving effectiveness. Intelligent routing picks optimal models per task. Organizations report 30-50% cost reduction. The unglamorous math that determines whether your agent runs profitably or burns budget.