How I cut multi-agent LLM costs by 70% using prompt caching, artifact minification, and smarter execution order — and what the AWAF assessment found along the way.