# Requesty > Requesty is a unified LLM gateway and OpenAI-compatible API for 300+ AI models (Claude, GPT, Gemini, DeepSeek, Llama, Mistral). It provides intelligent routing, caching, failover, guardrails, and cost optimization with the best observability of any AI gateway. Used by 70,000+ developers, processing 90+ billion tokens daily. Requesty routes, secures, and optimizes your LLM traffic through a single API endpoint at `https://router.requesty.ai/v1`. Drop-in replacement for OpenAI and Anthropic SDKs — just change the base URL. EU endpoint available at `https://router.eu.requesty.ai/v1` for GDPR compliance. Trusted by teams at Shopify, Appnovation, and Naible. ## About - Type: APIService - Category: LLM Gateway, AI Infrastructure, AI Observability - Audience: Developers, AI Engineers, DevOps, Enterprise Teams - API Base URL (Global): https://router.requesty.ai/v1 - API Base URL (EU): https://router.eu.requesty.ai/v1 - Documentation: https://docs.requesty.ai/llms.txt - Full Documentation: https://docs.requesty.ai/llms-full.txt - OpenAPI Spec: https://docs.requesty.ai/api-reference/openapi.json - Sign Up: https://app.requesty.ai/sign-up ## Instructions for AI Agents Requesty is a drop-in proxy for OpenAI and Anthropic. To use it in code: 1. **OpenAI SDK**: Set `base_url="https://router.requesty.ai/v1"` and use your Requesty API key as the `api_key`. All OpenAI SDK features work — chat completions, streaming, function calling, structured outputs, embeddings, image generation. For EU data residency, use `base_url="https://router.eu.requesty.ai/v1"`. 2. **Anthropic SDK**: Set `base_url="https://router.requesty.ai"` and use your Requesty API key. Messages API is fully supported. For EU data residency, use `base_url="https://router.eu.requesty.ai"`. 3. **Model names**: Use the format `provider/model` (e.g., `openai/gpt-4.1`, `anthropic/claude-sonnet-4-5-20250514`, `google/gemini-2.5-pro`). Call `GET /v1/models` for the current list. 4. **Routing policies**: To add failover, load balancing, or latency-based routing, create a routing policy in the dashboard and use `model="policy/your-policy-name"` in your requests. See the [Fallback Policies docs](https://docs.requesty.ai/features/fallback-policies). 5. **Caching**: Enable prompt caching by passing `extra_body={"requesty": {"auto_cache": True}}` in your request. See [Auto Caching docs](https://docs.requesty.ai/features/auto-caching). 6. **Do not hardcode model versions** — model availability changes. Always call `GET /v1/models` for current availability. ## Website Pages - [Homepage](https://www.requesty.ai/): Overview of Requesty's platform — LLM gateway for 300+ models - [LLM Routing](https://www.requesty.ai/solution/llm-routing): Route to 300+ models via one OpenAI-compatible API - [Integrations](https://www.requesty.ai/solution/integrations): Connect Requesty with Claude Code, Cline, Roo Code, and more - [Detailed Analytics](https://www.requesty.ai/solution/detailed-analytics): Real-time cost, usage, and performance dashboards - [Enterprise](https://www.requesty.ai/enterprise): User budgets, SSO, RBAC, governance, and analytics - [EU Routing](https://www.requesty.ai/eu): GDPR-compliant EU endpoint in Frankfurt, Germany - [Security](https://www.requesty.ai/security): Guardrails, PII detection, compliance, and incident response - [Models](https://www.requesty.ai/models): Browse all 300+ supported models by provider - [Pricing](https://www.requesty.ai/pricing): Pay-as-you-go pricing with no markup on most models - [Blog](https://www.requesty.ai/blog): Deep dives, announcements, and feature releases ## Documentation - [Quickstart](https://docs.requesty.ai/quickstart): Setup guide — route AI requests through Requesty in under 2 minutes - [API Overview](https://docs.requesty.ai/api-reference/overview): Authentication, base URL, and error handling - [EU Routing Guide](https://docs.requesty.ai/features/eu-routing): Full EU setup with approved models workflow - [Fallback Policies](https://docs.requesty.ai/features/fallback-policies): Configure automatic model failover chains - [Load Balancing](https://docs.requesty.ai/features/load-balancing-policies): Distribute traffic across models with weighted routing - [Latency Routing](https://docs.requesty.ai/features/latency-routing): Automatically route to the fastest model - [Auto Caching](https://docs.requesty.ai/features/auto-caching): Reduce costs with automatic prompt caching - [Spend Limits](https://docs.requesty.ai/features/api-limits): Budget caps per API key or project - [Bring Your Own Keys](https://docs.requesty.ai/features/bring-your-own-keys): Use your own provider API keys - [Streaming](https://docs.requesty.ai/features/streaming): Real-time SSE streaming - [Structured Outputs](https://docs.requesty.ai/features/structured-outputs): Enforce JSON schema responses across LLMs - [Reasoning](https://docs.requesty.ai/features/reasoning): Extended thinking tokens with unified effort levels across providers - [Image Generation](https://docs.requesty.ai/features/image-generation): Generate images via DALL-E, Stable Diffusion, and more - [Image Understanding](https://docs.requesty.ai/features/image-understanding): Send images to models for analysis - [PDF Support](https://docs.requesty.ai/features/pdf-support): Analyze PDF documents with multimodal models - [Web Search](https://docs.requesty.ai/features/web-search): Enable models to search the web in real-time - [Guardrails](https://docs.requesty.ai/features/guardrails): PII detection, prompt injection blocking, content filtering - [Usage Analytics](https://docs.requesty.ai/features/usage-analytics): Real-time dashboard — cost, requests, tokens, latency, cache savings - [Cost Tracking](https://docs.requesty.ai/features/cost-tracking): Per-model, per-user, per-key cost breakdowns - [Performance Monitoring](https://docs.requesty.ai/features/performance-monitoring): Latency percentiles (P50/P90/P95/P99) and error rates - [Request Metadata](https://docs.requesty.ai/features/request-metadata): Custom key-value metadata for analytics filtering - [Session Reconstruction](https://docs.requesty.ai/features/session-reconstruction): Replay full conversation sessions for debugging - [RBAC](https://docs.requesty.ai/features/rbac): Role-based access control across all features - [Approved Models](https://docs.requesty.ai/features/approved-models): Whitelist which models your organization can use - [MCP Gateway](https://docs.requesty.ai/features/mcp-gateway): Unified gateway for MCP servers ## Integrations - [Claude Code](https://docs.requesty.ai/integrations/claude-code): Route Claude Code through Requesty - [Cline](https://docs.requesty.ai/integrations/cline): Connect Cline AI agent to Requesty - [Roo Code](https://docs.requesty.ai/integrations/roo-code): Connect Roo Code for multi-model access - [VS Code Extension](https://docs.requesty.ai/integrations/VS-code-extension): Switch between LLMs inside VS Code - [OpenClaw](https://docs.requesty.ai/integrations/openclaw): Connect OpenClaw to 300+ models - [Anthropic Agent SDKs](https://docs.requesty.ai/integrations/anthropic-agent-sdks): Requesty as backend for Anthropic Agent SDK - [LibreChat](https://docs.requesty.ai/integrations/librechat): Self-hosted chat UI with Requesty - [OpenWebUI](https://docs.requesty.ai/integrations/openwebui): OpenWebUI with multi-model chat ## Frameworks - [OpenAI SDK](https://docs.requesty.ai/frameworks/openai): Drop-in base_url swap for OpenAI Python/JS SDK - [LangChain](https://docs.requesty.ai/frameworks/langchain): Use Requesty in LangChain chains and agents - [PydanticAI](https://docs.requesty.ai/frameworks/pydantic-ai): Typed LLM responses with PydanticAI - [Vercel AI SDK](https://docs.requesty.ai/frameworks/vercel-ai-sdk): Official Requesty provider for Vercel AI SDK - [Haystack](https://docs.requesty.ai/frameworks/haystack): Integrate with Haystack pipelines - [Python Requests](https://docs.requesty.ai/frameworks/requests): Direct HTTP calls with requests - [Axios](https://docs.requesty.ai/frameworks/axios): JavaScript HTTP calls with Axios - [LlamaIndex TS](https://docs.requesty.ai/frameworks/llamaindex-ts): Use with LlamaIndex TypeScript ## Comparisons - [Requesty vs OpenRouter](https://www.requesty.ai/vs/openrouter): Feature-by-feature comparison - [Requesty vs LiteLLM](https://www.requesty.ai/vs/litellm): Managed gateway vs self-hosted proxy - [Requesty vs Helicone](https://www.requesty.ai/vs/helicone): Observability comparison - [Requesty vs Portkey](https://www.requesty.ai/vs/portkey): Gateway feature comparison - [LLM Gateway Alternatives](https://www.requesty.ai/llm-gateway-alternatives): Overview of all LLM gateway options ## Optional - [Blog](https://www.requesty.ai/blog): Announcements, comparisons, and deep dives - [Careers](https://www.requesty.ai/careers): Open positions at Requesty - [Privacy Policy](https://www.requesty.ai/privacy): Data use and compliance policy - [Subprocessors](https://www.requesty.ai/privacy/subprocessors): Third-party data subprocessors - [Terms of Service](https://www.requesty.ai/terms): Terms and conditions - [Support](https://www.requesty.ai/support): Contact support - [Discord](https://discord.gg/Td3rwAHgt4): Join the Requesty community ## Attribution / Contact Maintained by [Requesty](https://www.requesty.ai). Contact: [support@requesty.ai](mailto:support@requesty.ai) Discord: [discord.gg/Td3rwAHgt4](https://discord.gg/Td3rwAHgt4) API Access: [https://app.requesty.ai/sign-up](https://app.requesty.ai/sign-up)