Since Ollama has prompt caching now (right?), it would be great if the utilization of the cache could be returned in requests. E.g. the OpenAI-compatible API could be extended with the new [`usage/prompt_tokens_details/cached_tokens`](https://platform.openai.com/docs/guides/prompt-caching). A similar field in the Ollama API would also be great.