14 KiB
Requesty.ai Integration - User Migration Guide
Welcome to Requesty!
This guide will help you start using Requesty.ai with agentic-flow to access 300+ AI models with significant cost savings.
Quick Start (3 Steps)
1. Get Your Requesty API Key
- Visit Requesty.ai
- Sign up or log in to your account
- Navigate to API Keys section
- Click Generate New Key
- Copy your API key (starts with
requesty-)
2. Set Your API Key
Option A: Environment Variable
export REQUESTY_API_KEY="requesty-xxxxxxxxxxxxx"
Option B: .env File
# Add to .env file
REQUESTY_API_KEY=requesty-xxxxxxxxxxxxx
3. Run Your First Command
npx agentic-flow --agent coder \
--task "Create a hello world function" \
--provider requesty
That's it! You're now using Requesty with 300+ models.
Why Use Requesty?
| Feature | Benefit |
|---|---|
| 300+ Models | Access OpenAI, Anthropic, Google, Meta, DeepSeek, and more |
| 80% Cost Savings | Significantly cheaper than direct API calls |
| Auto-Routing | Intelligent model selection based on your needs |
| Built-in Caching | Reduce redundant API calls |
| Analytics Dashboard | Track usage, costs, and performance |
| Zero Downtime | Automatic failover and load balancing |
Usage Examples
Basic Chat Completion
npx agentic-flow --agent coder \
--task "Explain async/await in JavaScript" \
--provider requesty
Uses default model: openai/gpt-4o-mini
Specify a Model
npx agentic-flow --agent researcher \
--task "Research quantum computing trends" \
--provider requesty \
--model "anthropic/claude-3.5-sonnet"
Enable Streaming
npx agentic-flow --agent coder \
--task "Write a detailed tutorial on React hooks" \
--provider requesty \
--model "openai/gpt-4o" \
--stream
Use FREE Models
# Google Gemini 2.5 Flash - Completely FREE!
npx agentic-flow --agent coder \
--task "Create a REST API with Express" \
--provider requesty \
--model "google/gemini-2.5-flash"
Cost-Optimized Models
# DeepSeek Chat V3 - Only $0.03 per 1M tokens
npx agentic-flow --agent coder \
--task "Build a calculator function" \
--provider requesty \
--model "deepseek/deepseek-chat-v3"
Premium Quality Models
# GPT-4o - Top-tier quality
npx agentic-flow --agent architect \
--task "Design a microservices architecture" \
--provider requesty \
--model "openai/gpt-4o"
Recommended Models
For General Coding
| Model | Cost/1M Tokens | Speed | Quality | Best For |
|---|---|---|---|---|
openai/gpt-4o-mini |
$0.03 | Fast | Good | Quick tasks, debugging |
deepseek/deepseek-chat-v3 |
$0.03 | Fast | Good | Cost-effective coding |
qwen/qwen-2.5-coder-32b |
$0.05 | Fast | Excellent | Specialized coding |
openai/gpt-4o |
$0.50 | Medium | Excellent | Complex architecture |
For Research & Analysis
| Model | Cost/1M Tokens | Speed | Quality | Best For |
|---|---|---|---|---|
google/gemini-2.5-flash |
FREE | Very Fast | Good | General research |
anthropic/claude-3.5-sonnet |
$0.60 | Medium | Excellent | Deep analysis |
openai/gpt-4-turbo |
$1.00 | Medium | Excellent | Complex reasoning |
google/gemini-2.5-pro |
$0.10 | Fast | Very Good | Large context tasks |
For Cost Optimization
| Model | Cost/1M Tokens | Savings vs Claude 3.5 | Quality |
|---|---|---|---|
google/gemini-2.5-flash |
FREE | 100% | Good |
deepseek/deepseek-chat-v3 |
$0.03 | 95% | Good |
meta-llama/llama-3.3-8b |
$0.02 | 97% | Good |
openai/gpt-4o-mini |
$0.03 | 95% | Very Good |
Configuration Options
Environment Variables
# Required
export REQUESTY_API_KEY="requesty-xxxxxxxxxxxxx"
# Optional
export REQUESTY_BASE_URL="https://router.requesty.ai/v1" # Custom base URL
export USE_REQUESTY="true" # Force Requesty provider
export COMPLETION_MODEL="openai/gpt-4o-mini" # Default model
export PROXY_PORT="3000" # Proxy server port
CLI Flags
--provider requesty # Use Requesty provider
--model "model-id" # Specify model
--stream # Enable streaming
--temperature 0.7 # Creativity (0.0-1.0)
--max-tokens 4096 # Maximum output length
--verbose # Detailed logging
Migration from Other Providers
From Anthropic Direct
Before:
export ANTHROPIC_API_KEY="sk-ant-xxxxx"
npx agentic-flow --agent coder --task "Create function"
After:
export REQUESTY_API_KEY="requesty-xxxxx"
npx agentic-flow --agent coder \
--task "Create function" \
--provider requesty \
--model "anthropic/claude-3.5-sonnet"
Benefits:
- 80% cost savings
- Same Claude quality
- Access to 300+ other models
From OpenRouter
Before:
export OPENROUTER_API_KEY="sk-or-xxxxx"
npx agentic-flow --agent coder \
--task "Create function" \
--provider openrouter \
--model "meta-llama/llama-3.1-8b-instruct"
After:
export REQUESTY_API_KEY="requesty-xxxxx"
npx agentic-flow --agent coder \
--task "Create function" \
--provider requesty \
--model "meta-llama/llama-3.3-70b-instruct"
Benefits:
- 200 more models (300 vs 100)
- Built-in analytics dashboard
- Auto-routing and caching
From Google Gemini
Before:
export GOOGLE_GEMINI_API_KEY="xxxxx"
npx agentic-flow --agent coder \
--task "Create function" \
--provider gemini
After:
export REQUESTY_API_KEY="requesty-xxxxx"
npx agentic-flow --agent coder \
--task "Create function" \
--provider requesty \
--model "google/gemini-2.5-flash"
Benefits:
- Access to OpenAI, Anthropic, DeepSeek models
- Unified billing and analytics
- Model fallback support
Advanced Usage
Use with Claude Code
Terminal 1 - Start Proxy
npx agentic-flow proxy --provider requesty --port 3000
Terminal 2 - Configure Claude Code
export ANTHROPIC_BASE_URL="http://localhost:3000"
export ANTHROPIC_API_KEY="sk-ant-proxy-dummy-key"
export REQUESTY_API_KEY="requesty-xxxxx"
claude
Now Claude Code will use Requesty models!
Auto-Start Proxy
# One command - proxy + Claude Code
npx agentic-flow claude-code --provider requesty "Create a React app"
Model Optimization
# Let agentic-flow choose the best model automatically
npx agentic-flow --agent coder \
--task "Build API endpoint" \
--optimize \
--priority cost \
--provider requesty
Priorities:
quality- Best results (Claude, GPT-4o)balanced- Good quality + cost (DeepSeek, Gemini)cost- Cheapest options (FREE tier models)speed- Fastest responses (Gemini Flash)
Troubleshooting
Issue: "REQUESTY_API_KEY required"
Solution:
# Check if API key is set
echo $REQUESTY_API_KEY
# If empty, set it
export REQUESTY_API_KEY="requesty-xxxxx"
Issue: "Invalid API key"
Solution:
- Verify your API key starts with
requesty- - Check for typos or extra spaces
- Generate a new key at https://app.requesty.ai
- Make sure key is active (not revoked)
Issue: "Rate limit exceeded"
Solution:
# Wait and retry (auto-retry is built-in)
# Or upgrade your Requesty tier
# Or use a different model temporarily
Issue: "Model not found"
Solution:
# Check model ID format: <provider>/<model-name>
# Example: openai/gpt-4o-mini (correct)
# gpt-4o-mini (incorrect - missing provider)
# Verify model exists at https://app.requesty.ai/model-list
Issue: Proxy won't start
Solution:
# Check if port 3000 is already in use
lsof -i :3000
# Use a different port
PROXY_PORT=8080 npx agentic-flow proxy --provider requesty --port 8080
Issue: Response is slow
Solution:
# Use faster models
--model "google/gemini-2.5-flash" # Fastest
--model "openai/gpt-4o-mini" # Fast
--model "deepseek/deepseek-chat-v3" # Fast + cheap
# Enable streaming for perceived speed
--stream
Issue: Tool calling not working
Solution:
# Some older models don't support tools
# Use known tool-compatible models:
--model "openai/gpt-4o-mini" # ✓ Tools
--model "anthropic/claude-3.5-sonnet" # ✓ Tools
--model "google/gemini-2.5-flash" # ✓ Tools
--model "deepseek/deepseek-chat-v3" # ✓ Tools
# Avoid older models like:
--model "mistralai/mistral-7b-instruct" # ✗ No tools (emulation used)
Cost Comparison
Agentic Flow Task: "Create a REST API with Express.js"
| Provider | Model | Tokens Used | Cost | Savings |
|---|---|---|---|---|
| Anthropic Direct | claude-3.5-sonnet | 5,000 | $0.0150 | Baseline |
| Requesty | anthropic/claude-3.5-sonnet | 5,000 | $0.0030 | 80% |
| Requesty | openai/gpt-4o-mini | 5,000 | $0.00015 | 99% |
| Requesty | google/gemini-2.5-flash | 5,000 | $0.0000 | 100% |
| Requesty | deepseek/deepseek-chat-v3 | 5,000 | $0.00015 | 99% |
Real Savings Example:
- 100 tasks/day with Claude 3.5 Sonnet
- Direct: $1.50/day = $45/month
- Requesty: $0.30/day = $9/month
- Savings: $36/month (80%)
Best Practices
1. Start with Free Tier
# Test Requesty with FREE models first
--model "google/gemini-2.5-flash"
2. Use Right Model for Task
# Simple tasks → cheap models
--model "deepseek/deepseek-chat-v3"
# Complex tasks → premium models
--model "openai/gpt-4o"
# Research → large context models
--model "google/gemini-2.5-pro"
3. Enable Streaming for UX
# Always stream for user-facing tasks
--stream
4. Monitor Costs
Visit Requesty Dashboard to:
- Track token usage
- Monitor spending
- Set budget alerts
- Compare model costs
5. Use Model Optimizer
# Let agentic-flow choose the best model
--optimize --priority balanced
FAQ
Q: Do I need both ANTHROPIC_API_KEY and REQUESTY_API_KEY?
A: No, only REQUESTY_API_KEY is needed when using --provider requesty.
Q: Can I use Requesty and Anthropic together?
A: Yes! Use --provider requesty for some tasks and --provider anthropic for others.
Q: Does Requesty work with all agentic-flow features?
A: Yes! Tool calling, streaming, MCP servers, and all agents work with Requesty.
Q: Is my data secure with Requesty?
A: Yes. Requesty follows industry-standard security practices. Check their privacy policy.
Q: Can I use Requesty with Claude Code/Cursor?
A: Yes! Use proxy mode:
npx agentic-flow proxy --provider requesty
Q: How do I get support?
A:
- Requesty support: support@requesty.ai
- Agentic Flow issues: https://github.com/ruvnet/agentic-flow/issues
Q: Are there usage limits?
A: Yes, limits depend on your Requesty tier. Free tier has lower limits. Upgrade for higher limits.
Q: Can I use custom models?
A: If your custom model is available on Requesty's platform, yes! Check their model catalog.
Q: Does Requesty support vision/image models?
A: Yes! Some models support vision:
--model "openai/gpt-4o" # Supports vision
--model "google/gemini-2.5-pro" # Supports vision
Q: How does caching work?
A: Requesty automatically caches similar requests to reduce costs. No configuration needed.
Model Catalog
Full Model List
Visit Requesty Model Library for the complete catalog of 300+ models.
Popular Models Quick Reference
# OpenAI
openai/gpt-4o # Premium quality, $0.50/1M tokens
openai/gpt-4o-mini # Fast, cost-effective, $0.03/1M tokens
openai/gpt-4-turbo # High quality, $1.00/1M tokens
openai/gpt-3.5-turbo # Legacy, cheap, $0.05/1M tokens
# Anthropic
anthropic/claude-3.5-sonnet # Best reasoning, $0.60/1M tokens
anthropic/claude-3-opus # Premium, $1.50/1M tokens
anthropic/claude-3-sonnet # Balanced, $0.30/1M tokens
anthropic/claude-3-haiku # Fast, $0.08/1M tokens
# Google
google/gemini-2.5-pro # Large context, $0.10/1M tokens
google/gemini-2.5-flash # FREE tier, fast
# DeepSeek
deepseek/deepseek-chat-v3 # Cost-optimized, $0.03/1M tokens
deepseek/deepseek-coder # Coding-focused, $0.03/1M tokens
# Meta/Llama
meta-llama/llama-3.3-70b-instruct # Open source, $0.10/1M tokens
meta-llama/llama-3.3-8b-instruct # Fast, cheap, $0.02/1M tokens
# Qwen
qwen/qwen-2.5-coder-32b-instruct # Coding expert, $0.05/1M tokens
# Mistral
mistralai/mistral-large # European alternative, $0.20/1M tokens
Getting Help
Documentation
- Requesty Docs: https://docs.requesty.ai
- Agentic Flow Docs: https://github.com/ruvnet/agentic-flow
Support Channels
- Email: support@requesty.ai
- Discord: Requesty Discord
- GitHub Issues: https://github.com/ruvnet/agentic-flow/issues
Community
- Share tips and tricks
- Report bugs
- Request new features
Next Steps
- Get Your API Key - https://app.requesty.ai
- Try Free Models -
google/gemini-2.5-flash - Test Premium Models -
anthropic/claude-3.5-sonnet - Monitor Usage - Check Requesty dashboard
- Optimize Costs - Use
--optimizeflag
Happy Coding with Requesty! 🚀
Save money, access 300+ models, and build amazing AI applications with agentic-flow + Requesty.ai.