We're utilizing LiteLLM Proxy to create budgets, rate limits, observability of LLM calls etc (eventually a/b testing) and also offer a wider selection of models than are available in the native extensions.
LiteLLM Proxy is compatible with the OpenAI interface, so to achieve our goals for at scale deployment for our organization, it's a critical component to us that offers significantly more flexibility than our current options with clay.