Optimization
3 posts
I Tested LLM Prompt Caching With Anthropic and OpenAI
Experiments testing prompt caching with actual API calls, measuring cache hits, token counts, and costs across Anthropic Claude and OpenAI GPT.
Where Did All the RAM Go? Memory Profiling with Memray
High CPU isn''t the only performance issue. Learn how Memray helps track memory leaks and excessive allocation in your Python library to optimize usage.
Finding the Slowdown: Profiling Python Code with Pyinstrument
Your benchmark says a function is slow, but why? Profilers like Pyinstrument help you pinpoint exactly where your Python code is spending its time.