Skip to main content
Rate limits help ensure fair usage and optimal performance for all users. Each subscription plan has specific limits for queued API requests.

Request Queue Limits

Our API uses request queue limits to manage server load and ensure optimal performance. The limits vary based on your subscription plan:

Pay as you go plan

5 queued API requestsPerfect for individual developers and small projects getting started with our APIs.

Standard plan

10 queued API requestsIdeal for growing businesses and applications with moderate usage requirements.

Unlimited Premium Plan

15 queued API requestsDesigned for enterprise applications and high-volume usage scenarios.

How Request Queuing Works

Request queuing ensures that API calls are processed sequentially in a controlled manner. Here’s what you need to know:

Understanding the Limits

  • Sequential Processing: Requests are processed one after another in queue order
  • Queue Management: New requests are added to the queue and processed when previous ones complete
  • Per Account: Limits apply to your entire account, not per API endpoint
  • Real-time: The limit is enforced in real-time as requests come in

Example Scenarios

If you have 5 requests in the queue and try to make a 6th request, it will be rejected until one of the queued requests completes and is processed.
You can have up to 10 requests in the queue. Any additional requests will be rejected until queue space becomes available.
Up to 15 requests can be queued, providing the highest capacity for enterprise applications.

Request Queuing

When you reach your queue limit:
  1. Queue Full: Additional requests are rejected with a rate limit error
  2. Sequential Processing: Requests are processed one after another in queue order
  3. FIFO Order: Requests are processed in First-In-First-Out order
  4. Automatic Processing: Queued requests are automatically processed as previous ones complete

Best Practices

Optimizing for Rate Limits

1

Monitor Your Queue

Keep track of your queued requests to avoid hitting limits unexpectedly.
2

Implement Retry Logic

Add exponential backoff retry logic for better handling of rate limit scenarios.
3

Batch Operations

When possible, combine multiple operations into single requests to reduce queue usage.
4

Upgrade When Needed

Consider upgrading your plan if you consistently hit queue limits.

Error Handling

When you hit rate limits, you’ll receive an HTTP 429 status code with details about the limit:
json
{
  "status": "error",
  "message": "Rate limit exceeded. Maximum 5 queued requests allowed.",
  "retry_after": 30
}

Upgrading Your Plan

If you need higher queue limits:
  1. Log in to your ModelsLab account
  2. Navigate to the billing section
  3. Select a higher tier plan
  4. New limits take effect immediately
Need more than 15 queued requests? Contact our enterprise team for custom solutions tailored to your specific needs.

Frequently Asked Questions

Additional requests are rejected with a rate limit error. You’ll need to wait for queued requests to complete before making new ones.
No, queue limits are enforced in real-time, not on a daily basis. They’re based on current queue size, not total requests per day.
No, the queue limit applies to your entire account across all API endpoints.

Support

Need help with rate limits or want to discuss custom solutions?