How It Works

  1. Your primary model gets the request.
  2. If it fails (timeout, error, etc.), the router immediately tries the next model.
  3. This continues until one model delivers the results you need.

How Does This Help?

  • No more stalled workflows
  • Higher success rates
  • Optimized costs - you only pay for what works

Get Started

  1. Go to API Keys Page
  2. Add a Fallback Policy (see screenshot)

Here’s an example setup:

  • Policy Name:

    Deepseek-R1

  • Fallback Chain:

    deepseek/reasoner → nebius/DeepSeek-R1 → deepinfra/DeepSeek-R1

Make sure each model in your Fallback Chain is compatible with your request parameters (e.g., context length). If a particular model cannot handle the request’s requirements, the policy will move on to the next model without additional warnings.