TIL: Context Compaction with Google ADK 1.16
· 9 min read
Long agent conversations accumulate thousands of tokens. After 100+ exchanges, sending the entire history to the model becomes expensive and slow. Context Compaction fixes this by intelligently summarizing old interactions.
In one sentence: Context Compaction automatically summarizes older conversation events using an LLM, reducing token usage while preserving conversational context.
