Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.requesty.ai/llms.txt

Use this file to discover all available pages before exploring further.

Pi is an AI coding agent that runs locally on your machine. Using the Requesty integration, you can:
  • Access 300+ models from OpenAI, Anthropic, Google, Mistral, and many other providers through one API key.
  • Track and manage your spend in a single location.
  • Apply fallback policies, load balancing, and latency routing to keep your agent responsive.

Prerequisites

  • Pi installed on your machine. See pi.dev for installation instructions.
  • A Requesty API key from the API Keys page.

Configuration

1

Install Pi

Follow the instructions at pi.dev to download and install Pi on your machine.
2

Configure the models file

Create or edit the models configuration file at ~/.pi/agent/models.json and add Requesty as a provider:
{
  "providers": {
    "requesty": {
      "name": "Requesty",
      "baseUrl": "https://router.requesty.ai/v1",
      "api": "openai-completions",
      "apiKey": "rqsty-sk-...."
    }
  }
}
Replace rqsty-sk-.... with your Requesty API key from the API Keys page.
3

Install the Requesty extension

Install the Requesty extension for Pi by running:
pi install git:github.com/requestyai/pi-requesty
4

Sync models

Run the following command inside Pi to sync the available Requesty models:
/requesty-models-sync
This pulls the latest model catalog from Requesty so you can select any approved model from your organization.
5

Reload the configuration

Reload Pi to pick up the new provider and models:
/reload

Selecting a model

After the sync completes, run the /model command inside Pi to choose a model from the list you synced. You should be able to see both discrete models, e.g. anthropic/claude-opus-4-7 and your custom policies, e.g. policy/opus-europe.

Verifying the integration

Run a quick prompt inside Pi to confirm that requests are routed through Requesty. Then open the Requesty analytics dashboard to verify the request was logged. Spend, latency, and token counts should appear within seconds.

References

Last modified on May 11, 2026