Librechat + Requesty

Mar 11, 2025

Try the Requesty Router and get $6 free credits 🔀

Join the discord

LibreChat is a powerful open-source web UI that serves as your one-stop interface for AI agents and models. In this post, we’ll show you how to pair LibreChat with Requesty—a single API platform that connects you to 150+ AI models (like GPT-4, Claude, and more). You’ll see how easy it is to start for free, set up a Docker-based LibreChat instance, and unlock a whole world of free AI experimentation.

Table of Contents

  1. Why Combine LibreChat & Requesty?

  2. Getting Started with Docker & LibreChat

    • Prerequisites

    • The LibreChat .env

  3. Adding Requesty to Your Setup

    • One .env File to Rule Them All

    • Requesty Config Snippet

  4. Running Docker Compose

  5. Exploring AI Agents in LibreChat

  6. FAQ & Troubleshooting

  7. Wrap-Up

  8. Why Combine LibreChat & Requesty? LibreChat is an easy-to-deploy, open-source chat interface for all your AI interactions. It’s user-friendly, offers multiple tabs, and supports advanced features like code highlighting and conversation history. But what if you want to access more than just one or two default models?

Enter Requesty—the one API platform that routes your prompts to 150+ AI models. With Requesty, you can:

  • Start for free with $6 in sign-up credits.

  • Enjoy a “one API key” approach for GPT, Claude, Google G-Flan, and many more.

  • Seamlessly switch between AI providers without rewriting config files.

  • Monitor usage, costs, and logs in a single dashboard.

The synergy is clear: LibreChat gives you a polished UI, while Requesty unlocks a huge library of AI agents in the background.

  1. Getting Started with Docker & LibreChat

Prerequisites

  • Docker & Docker Compose: Make sure you have them installed on your machine.

  • Git: If you plan to clone the LibreChat repository.

The LibreChat .env

After pulling the LibreChat project (usually via Git), look for a file named .env or .env.example. This is where LibreChat keeps its core configuration.

Out of the box, .env might already have some basic settings. We’ll expand it to let Requesty seamlessly route your prompts.

  1. Adding Requesty to Your Setup

One .env File to Rule Them All

The beauty here is that you’ll only need to add a few lines to your existing .env to connect LibreChat to any AI agent behind Requesty’s router.

Requesty Config Snippet

Just copy and paste this snippet into your LibreChat .env file:


Where do you get your API key?

  1. Go to Requesty’s Router Page.

  2. Under “Manage API Keys,” click Create API Key and name it something like librechat-key.

  3. Copy that key and replace the placeholder in the snippet above.

  4. Running Docker Compose After saving your updated .env, run:

That’s it! Docker will spin up LibreChat with your new environment variables in place. You’ll see logs showing that LibreChat recognized the CONFIG_PATH and REQUESTY_KEY.

Once the containers initialize, open your browser and go to:

(Your port may vary depending on your Docker settings.)

  1. Exploring AI Agents in LibreChat When LibreChat loads, you’ll notice new model options that weren’t there before. That’s because Requesty automatically imports 150+ models into your chat interface—no separate API keys, no complex config changes.

  • Test it out: Type “Hello, world!” in your chat.

  • Switch models: Change the model from GPT-based to Claude or Anthropic with a simple dropdown selection.

  • Manage cost & usage: Head to Requesty’s Dashboard to see how many tokens you used, how much it cost, and logs of your conversations (only if logging is enabled).

  1. FAQ & Troubleshooting Q: What if I see “Invalid API Key” errors?
    A: Double-check you’ve pasted your Requesty key correctly, with no extra spaces.

Q: Can I disable logs for privacy?
A: Yes. In Requesty’s dashboard, you can toggle logs on or off for each API key.

Q: Is this free?
A: You can get started for free with $6 in Requesty credits, which cover your initial usage. After that, you’ll pay per token. LibreChat itself is 100% open-source.

Q: My Docker containers won’t start.
A: Check Docker logs for port conflicts or YAML syntax issues. Make sure your .env variables are spelled correctly.

  1. Wrap-Up You now have a fully-featured AI assistant that can access any model you want, all through a single .env config. LibreChat keeps the UI simple, while Requesty connects you to everything from GPT-4o to Deepseek-R1, Anthropic Claude-3-7-sonnet.

  • No more juggling multiple API keys.

  • No more complicated Docker scripts.

  • One API platform for all your AI agents.

Ready to see what you can build? Jump into your newly configured LibreChat, spin up your favorite model, and experience free AI interactions at scale. If you have any questions, join our Discord or visit Requesty’s Docs.

Happy chatting—and welcome to the future of open-source AI!

Follow us on

© Requesty Ltd 2025