superdup95 / Su

GPT-3.5 Turbo: no wait / GPT-4: no wait / GPT-4 32k: no wait / GPT-4 Turbo: no wait / Claude (Sonnet): no wait / Claude (Opus): no wait / Gemini Pro: no wait / Mistral 7B: no wait / Mixtral Small: no wait / Mistral Medium: no wait / Mistral Large: no wait / AWS Claude (Sonnet): no wait / AWS Claude (Opus): no wait / DALL-E: no wait

Server Greeting

Recent DALL-E Generations

A large supermarket ablaze with bright, intense flames leaping from the windows and roof. Smoke billows out, filling the night sky, mingling with the twinkling stars. The flames lick the signboard, the lettering distorted by the heat. Silhouetted against this fiery background are hoses from fire trucks, their water arcs attempting to douse the conflagration. Nearby trees sway under the thermal updraft, their leaves illuminated by the flickering firelight. Despite the chaos, a sense of urgency and professional dedication pervades as firefighters, both men and women of diverse descents, bravely battle the inferno.

View all recent images


Service Info

{
  "uptime": 665358,
  "endpoints": {
    "openai": "https://superdup95-su.hf.space/su/openai",
    "openai2": "https://superdup95-su.hf.space/su/openai/turbo-instruct",
    "openai-image": "https://superdup95-su.hf.space/su/openai-image",
    "anthropic": "https://superdup95-su.hf.space/su/anthropic",
    "google-ai": "https://superdup95-su.hf.space/su/google-ai",
    "mistral-ai": "https://superdup95-su.hf.space/su/mistral-ai",
    "aws": "https://superdup95-su.hf.space/su/aws/claude"
  },
  "proompts": 1235,
  "tookens": "19.01m",
  "proomptersNow": 0,
  "openaiKeys": 46,
  "openaiOrgs": 45,
  "anthropicKeys": 3,
  "google-aiKeys": 19,
  "mistral-aiKeys": 3,
  "awsKeys": 1,
  "dall-e": {
    "usage": "4.0k tokens",
    "activeKeys": 23,
    "overQuotaKeys": 18,
    "proomptersInQueue": 0,
    "estimatedQueueTime": "no wait"
  },
  "turbo": {
    "usage": "1.7k tokens",
    "activeKeys": 23,
    "revokedKeys": 5,
    "overQuotaKeys": 18,
    "trialKeys": 0,
    "proomptersInQueue": 0,
    "estimatedQueueTime": "no wait"
  },
  "gpt4-turbo": {
    "usage": "1.94m tokens",
    "activeKeys": 23,
    "overQuotaKeys": 8,
    "proomptersInQueue": 0,
    "estimatedQueueTime": "no wait"
  },
  "gpt4": {
    "usage": "0 tokens",
    "activeKeys": 23,
    "overQuotaKeys": 8,
    "proomptersInQueue": 0,
    "estimatedQueueTime": "no wait"
  },
  "gpt4-32k": {
    "usage": "12.8k tokens",
    "activeKeys": 3,
    "overQuotaKeys": 3,
    "proomptersInQueue": 0,
    "estimatedQueueTime": "no wait"
  },
  "claude": {
    "usage": "0 tokens",
    "activeKeys": 1,
    "revokedKeys": 1,
    "overQuotaKeys": 1,
    "trialKeys": 0,
    "prefilledKeys": 0,
    "proomptersInQueue": 0,
    "estimatedQueueTime": "no wait"
  },
  "claude-opus": {
    "usage": "0 tokens",
    "activeKeys": 1,
    "revokedKeys": 1,
    "overQuotaKeys": 1,
    "trialKeys": 0,
    "prefilledKeys": 0,
    "proomptersInQueue": 0,
    "estimatedQueueTime": "no wait"
  },
  "gemini-pro": {
    "usage": "44.0k tokens",
    "activeKeys": 19,
    "revokedKeys": 0,
    "proomptersInQueue": 0,
    "estimatedQueueTime": "no wait"
  },
  "mistral-tiny": {
    "usage": "0 tokens",
    "activeKeys": 3,
    "revokedKeys": 0,
    "proomptersInQueue": 0,
    "estimatedQueueTime": "no wait"
  },
  "mistral-small": {
    "usage": "0 tokens",
    "activeKeys": 3,
    "revokedKeys": 0,
    "proomptersInQueue": 0,
    "estimatedQueueTime": "no wait"
  },
  "mistral-medium": {
    "usage": "0 tokens",
    "activeKeys": 3,
    "revokedKeys": 0,
    "proomptersInQueue": 0,
    "estimatedQueueTime": "no wait"
  },
  "mistral-large": {
    "usage": "0 tokens",
    "activeKeys": 3,
    "revokedKeys": 0,
    "proomptersInQueue": 0,
    "estimatedQueueTime": "no wait"
  },
  "aws-claude": {
    "usage": "0 tokens",
    "activeKeys": 1,
    "revokedKeys": 0,
    "sonnetKeys": 1,
    "haikuKeys": 1,
    "proomptersInQueue": 0,
    "estimatedQueueTime": "no wait"
  },
  "aws-claude-opus": {
    "usage": "17.01m tokens",
    "activeKeys": 1,
    "revokedKeys": 0,
    "proomptersInQueue": 0,
    "estimatedQueueTime": "no wait"
  },
  "config": {
    "gatekeeper": "proxy_key",
    "textModelRateLimit": "6",
    "imageModelRateLimit": "3",
    "maxContextTokensOpenAI": "0",
    "maxContextTokensAnthropic": "0",
    "maxOutputTokensOpenAI": "2048",
    "maxOutputTokensAnthropic": "4096",
    "rejectMessage": "This content violates /aicg/'s acceptable use policy.",
    "allowAwsLogging": "false",
    "promptLogging": "false",
    "tokenQuota": {
      "turbo": "0",
      "gpt4": "0",
      "gpt4-32k": "0",
      "gpt4-turbo": "0",
      "dall-e": "0",
      "claude": "0",
      "claude-opus": "0",
      "gemini-pro": "0",
      "mistral-tiny": "0",
      "mistral-small": "0",
      "mistral-medium": "0",
      "mistral-large": "0",
      "aws-claude": "0",
      "aws-claude-opus": "0",
      "azure-turbo": "0",
      "azure-gpt4": "0",
      "azure-gpt4-32k": "0",
      "azure-gpt4-turbo": "0",
      "azure-dall-e": "0"
    },
    "allowOpenAIToolUsage": "false",
    "allowImagePrompts": "true"
  },
  "build": "23d4a78 (main@yae-miko/oai-reverse-proxy)"
}