Azure GPT-4: no wait / Azure GPT-4 32k: no wait / Azure GPT-4 Turbo: no wait
for cooders
pass: hatchling
{ "uptime": 14671067, "endpoints": { "azure": "https://cooders.veryscrappy.moe/proxy/azure/openai" }, "proompts": 27015, "tookens": "96.45m", "proomptersNow": 0, "azureKeys": 6, "azure-gpt4-turbo": { "usage": "56.87m tokens", "activeKeys": 0, "revokedKeys": 2, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "azure-gpt4": { "usage": "35.25m tokens", "activeKeys": 0, "revokedKeys": 2, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "azure-gpt4-32k": { "usage": "4.34m tokens", "activeKeys": 0, "revokedKeys": 2, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "config": { "gatekeeper": "proxy_key", "maxIpsAutoBan": "true", "textModelRateLimit": "4", "imageModelRateLimit": "4", "maxContextTokensOpenAI": "128000", "maxContextTokensAnthropic": "0", "maxOutputTokensOpenAI": "4096", "maxOutputTokensAnthropic": "400", "allowAwsLogging": "false", "promptLogging": "false", "tokenQuota": { "turbo": "0", "gpt4": "0", "gpt4-32k": "0", "gpt4-turbo": "0", "dall-e": "0", "claude": "0", "claude-opus": "0", "gemini-pro": "0", "mistral-tiny": "0", "mistral-small": "0", "mistral-medium": "0", "mistral-large": "0", "aws-claude": "0", "aws-claude-opus": "0", "azure-turbo": "0", "azure-gpt4": "0", "azure-gpt4-32k": "0", "azure-gpt4-turbo": "0", "azure-dall-e": "0" }, "allowOpenAIToolUsage": "true", "allowImagePrompts": "false" }, "build": "[ci] b1062dc (main@khanon/oai-reverse-proxy)" }