{"status":"ok","service":"llm-cache-proxy","cache":"cf-cache-api","providers":["openai","openrouter","gemini"]}