KV Caching in LLMs: A Guide for Developers (machinelearningmastery.com)

Language models generate text one token at a time, reprocessing the entire sequence at each step.