Skip to content

Track and potentially improve Open AI cache use #250

@sdirix

Description

@sdirix

Open AI has an automatic internal prompt cache which can be influences a bit. Within the responses, it reports how many tokens were a cache hit.

See https://platform.openai.com/docs/guides/prompt-caching

ToDo:

  • Track cache hits when using OpenAI
  • The tracking stats and tracking GUI needs to be aligned with the cache tracking already done for Anthropic. Caching also has different effects on price and rate limits than it has for Anthropic.
    • no extra costs for caching, still affects rate limits,
  • Check whether we can improve cache hits when using OpenAI

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions