**MiniMax-01 Is Now Live on Requesty Router! **Easily Integrate MiniMax-Text-01 with Cline, OpenWebUI, and More
Overview
MiniMax has open-sourced the MiniMax-01 family of models—MiniMax-Text-01 and MiniMax-VL-01. These models offer advanced performance thanks to a novel Lightning Attention mechanism, scaling to handle up to 4 million tokens. Now, you can integrate these models directly with Requesty Router, which uses OpenAI-compatible endpoints and a single API credential. This seamless setup makes it easy to try MiniMax alongside other large language models through platforms like Cline and OpenWebUI.
Key MiniMax-01 Features
- **Lightning Attention **MiniMax-01 incorporates a linear attention approach in most layers, with occasional SoftMax layers for added expressiveness. This combination allows for high-speed handling of very long contexts.
- **Scalability **Featuring 456 billion parameters (45.9 billion active per inference), MiniMax-01 is designed for commercial-scale deployments, matching top-tier global models on core benchmarks.
- **Ultra-Long Context (Up to 4 Million Tokens) **Long contexts are crucial for advanced AI agent workflows, whether you’re maintaining a single agent’s memory or enabling multi-agent collaboration.
- **Open Source + Competitive Pricing **Model weights are available via GitHub, and API rates are set at $0.2 per million input tokens and $1.1 per million output tokens via the MiniMax Platform.
Integrating MiniMax-01 with Requesty Router
Requesty Router consolidates over 50 language models under one API key. If you already use Requesty, simply add the MiniMax endpoint to your existing setup:
- Obtain Access
- Sign up for the MiniMax Open Platform or visit their GitHub for open-source access.
- Add to Requesty Router
- In your Requesty Router admin panel, create a new OpenAI-compatible connection.
- Paste your Requesty API key, set the endpoint for MiniMax, and save your configuration.
- Use in Cline or OpenWebUI
- Cline: Point your coding workflow or AI assistant tasks to the newly added MiniMax endpoint.
- OpenWebUI: Switch between MiniMax-01 and other models in real time to compare responses on the same query.
Why Combine MiniMax-01 with Requesty Router?
- **One Key, Multiple Models **Eliminate the need to juggle several credentials. Use a single Requesty key to access MiniMax-01, GPT, Claude, and many others.
- **Flexible Workflows **Cline and OpenWebUI users can try out advanced coding or chat tasks with MiniMax, then switch to another model if needed.
- **Unified Billing **Track usage and manage your budget in one place. Gain better visibility into costs across different models.
- **Seamless Deployment **Requesty follows the OpenAI-style API standard, making it simple to integrate MiniMax-01 into your existing projects and pipelines.
Getting Started
- **Read the Tech Report **See the MiniMax-01 Report for performance details and architecture insights.
- **Clone the Repos **Check out the open-source code and weights on MiniMax’s GitHub.
- **Configure Requesty Router **Enable MiniMax-01 as a new endpoint. Make sure to keep your single API key ready for immediate use.
- Start Experimenting
- Code Generation: Use Cline with MiniMax-Text-01 to generate or debug complex scripts.
- Multimodal Projects: Run image-based tasks with MiniMax-VL-01 in the same environment.
- Side-by-Side Comparisons: In OpenWebUI, run the same prompt against multiple models and evaluate results.
Conclusion
With MiniMax-01 now open-sourced and accessible through Requesty Router, you have an efficient way to experiment with ultra-long context models. Integrate MiniMax-Text-01 or MiniMax-VL-01 into your AI workflows with a single key—no extra setup. Whether you’re building a single AI agent with deep memory or orchestrating multi-agent environments, MiniMax-01 offers robust performance at scale.
Try MiniMax-01 Today

