GPT-3.5 Turbo: no wait / GPT-4: no wait / GPT-4 32k: no wait / GPT-4 Turbo: no wait / Claude (Sonnet): no wait / Claude (Opus): no wait / Gemini Pro: no wait / Mistral 7B: no wait / Mixtral Small: no wait / Mistral Medium: no wait / Mistral Large: no wait / Azure GPT-3.5 Turbo: no wait / Azure GPT-4: no wait
{ "uptime": 52981, "endpoints": { "openai": "https://whore-proxy.ru/proxy/openai", "openai2": "https://whore-proxy.ru/proxy/openai/turbo-instruct", "anthropic": "https://whore-proxy.ru/proxy/anthropic", "google-ai": "https://whore-proxy.ru/proxy/google-ai", "mistral-ai": "https://whore-proxy.ru/proxy/mistral-ai", "aws": "https://whore-proxy.ru/proxy/aws/claude", "azure": "https://whore-proxy.ru/proxy/azure/openai" }, "proompts": 7396, "tookens": "99.82m", "proomptersNow": 7, "openaiKeys": 20, "openaiOrgs": 12, "anthropicKeys": 3, "google-aiKeys": 10, "mistral-aiKeys": 3, "awsKeys": 1, "azureKeys": 5, "turbo": { "usage": "2.2k tokens", "activeKeys": 5, "revokedKeys": 9, "overQuotaKeys": 6, "trialKeys": 0, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "gpt4-turbo": { "usage": "4.80m tokens", "activeKeys": 5, "overQuotaKeys": 2, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "gpt4": { "usage": "1.12m tokens", "activeKeys": 5, "overQuotaKeys": 2, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "gpt4-32k": { "usage": "32.0k tokens", "activeKeys": 5, "overQuotaKeys": 2, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "claude": { "usage": "3.40m tokens", "activeKeys": 2, "revokedKeys": 0, "overQuotaKeys": 1, "trialKeys": 0, "prefilledKeys": 1, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "claude-opus": { "usage": "90.47m tokens", "activeKeys": 2, "revokedKeys": 0, "overQuotaKeys": 1, "trialKeys": 0, "prefilledKeys": 1, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "gemini-pro": { "usage": "0 tokens", "activeKeys": 10, "revokedKeys": 0, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "mistral-tiny": { "usage": "0 tokens", "activeKeys": 2, "revokedKeys": 1, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "mistral-small": { "usage": "0 tokens", "activeKeys": 2, "revokedKeys": 0, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "mistral-medium": { "usage": "0 tokens", "activeKeys": 2, "revokedKeys": 0, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "mistral-large": { "usage": "0 tokens", "activeKeys": 2, "revokedKeys": 0, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "azure-gpt4": { "usage": "0 tokens", "activeKeys": 0, "revokedKeys": 4, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "azure-turbo": { "usage": "0 tokens", "activeKeys": 1, "revokedKeys": 0, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "config": { "gatekeeper": "user_token", "maxIpsAutoBan": "true", "textModelRateLimit": "3", "imageModelRateLimit": "2", "maxContextTokensOpenAI": "0", "maxContextTokensAnthropic": "0", "maxOutputTokensOpenAI": "1024", "maxOutputTokensAnthropic": "2048", "allowAwsLogging": "false", "promptLogging": "false", "tokenQuota": { "turbo": "0", "gpt4": "0", "gpt4-32k": "0", "gpt4-turbo": "0", "dall-e": "0", "claude": "0", "claude-opus": "0", "gemini-pro": "0", "mistral-tiny": "0", "mistral-small": "0", "mistral-medium": "0", "mistral-large": "0", "aws-claude": "0", "aws-claude-opus": "0", "azure-turbo": "0", "azure-gpt4": "0", "azure-gpt4-32k": "0", "azure-gpt4-turbo": "0", "azure-dall-e": "0" }, "allowOpenAIToolUsage": "true", "allowImagePrompts": "true" }, "build": "6dabc82 (main@khanon/oai-reverse-proxy)" }