| title | description |
|---|---|
Rate Limits |
Understanding ModelsLab API rate limits, request queuing, and concurrency limits for Free, Starter, Pro, and Enterprise subscription plans. |
Our API uses request queue limits to manage server load and ensure optimal performance. The limits vary based on your subscription plan:
**5 queued API requests**Perfect for individual developers and small projects getting started with our APIs.
Ideal for growing businesses and applications with moderate usage requirements.
Designed for enterprise applications and high-volume usage scenarios.
Request queuing ensures that API calls are processed sequentially in a controlled manner. Here's what you need to know:
- Sequential Processing: Requests are processed one after another in queue order
- Queue Management: New requests are added to the queue and processed when previous ones complete
- Per Account: Limits apply to your entire account, not per API endpoint
- Real-time: The limit is enforced in real-time as requests come in
When you reach your queue limit:
- Queue Full: Additional requests are rejected with a rate limit error
- Sequential Processing: Requests are processed one after another in queue order
- FIFO Order: Requests are processed in First-In-First-Out order
- Automatic Processing: Queued requests are automatically processed as previous ones complete
When you hit rate limits, you'll receive an HTTP 429 status code with details about the limit:
{
"status": "error",
"message": "Rate limit exceeded. Maximum 5 queued requests allowed.",
"retry_after": 30
}If you need higher queue limits:
- Log in to your ModelsLab account
- Navigate to the billing section
- Select a higher tier plan
- New limits take effect immediately
Need help with rate limits or want to discuss custom solutions?
- Documentation: Check our API Reference for detailed endpoint information
- Support: Contact us at support@modelslab.com
- Discord: Join our Discord community for real-time help