tasq/node_modules/agentic-flow/docs/archived/RELEASE-SUMMARY-v1.1.14-beta.1.md

8.6 KiB

v1.1.14-beta.1 Release Summary

Date: 2025-10-05 Status: PUBLISHED & LIVE NPM: https://www.npmjs.com/package/agentic-flow/v/1.1.14-beta.1 GitHub: https://github.com/ruvnet/agentic-flow/releases/tag/v1.1.14-beta.1


🎉 Major Achievement

Fixed critical OpenRouter proxy bug that was causing 100% failure rate!

From 0% success80% success (8 out of 10 models working)


Installation & Usage

Install Beta Version

# Install globally
npm install -g agentic-flow@beta

# Or use with npx (no installation needed)
npx agentic-flow@beta --help

Quick Start

# List available agents
npx agentic-flow@beta --list

# Run with Anthropic (default)
npx agentic-flow@beta --agent coder --task "Write Python hello world"

# Run with OpenRouter (99% cost savings!)
npx agentic-flow@beta --agent coder --task "Write Python hello world" \
  --provider openrouter --model "openai/gpt-4o-mini"

# Run with Grok 4 Fast (FREE!)
npx agentic-flow@beta --agent coder --task "Write Python hello world" \
  --provider openrouter --model "x-ai/grok-4-fast"

Verified Working

NPX Command

$ npx agentic-flow@beta --version
agentic-flow v1.1.14-beta.1

$ npx agentic-flow@beta --help
🤖 Agentic Flow v1.1.14-beta.1 - AI Agent Orchestration with OpenRouter Support
[Full help output shown]

$ npx agentic-flow@beta --agent coder --task "hello world" --provider anthropic
✅ Completed! [Working perfectly]

OpenRouter Models (8/10 = 80%)

Model Status Time Cost/M Tokens Use Case
openai/gpt-4o-mini 7s $0.15 Best value
openai/gpt-3.5-turbo 5s $0.50 Fastest
meta-llama/llama-3.1-8b-instruct 14s $0.06 Open source
anthropic/claude-3.5-sonnet 11s $3.00 Highest quality
mistralai/mistral-7b-instruct 6s $0.25 Fast & efficient
google/gemini-2.0-flash-exp 6s Free Free tier
x-ai/grok-4-fast 8s Free #1 most popular!
z-ai/glm-4.6 5s Varies Fixed in this release

Known Issues (2/10)

Model Issue Workaround
meta-llama/llama-3.3-70b-instruct Intermittent timeout Use llama-3.1-8b instead
x-ai/grok-4 Too slow (60s+) Use grok-4-fast instead

💰 Cost Savings

Comparison vs Claude Direct API

Model Cost vs Claude ($3/M) Savings
GPT-4o-mini $0.15/M $2.85 95%
Llama 3.1 8B $0.06/M $2.94 98%
Mistral 7B $0.25/M $2.75 92%
GPT-3.5-turbo $0.50/M $2.50 83%
Grok 4 Fast Free $3.00 100%
Gemini 2.0 Flash Free $3.00 100%

Average savings: ~94% across all working models


🔧 What Was Fixed

Critical Bug

TypeError: anthropicReq.system?.substring is not a function

Root Cause:

  • Anthropic API allows system field to be string OR array of content blocks
  • Claude Agent SDK sends it as array (for prompt caching)
  • Proxy assumed string only → called .substring() on array → crash
  • Result: 100% failure rate

Solution:

// Before (BROKEN)
interface AnthropicRequest {
  system?: string;
}

// After (FIXED)
interface AnthropicRequest {
  system?: string | Array<{ type: string; text?: string; [key: string]: any }>;
}

// Safe extraction logic
if (typeof anthropicReq.system === 'string') {
  originalSystem = anthropicReq.system;
} else if (Array.isArray(anthropicReq.system)) {
  originalSystem = anthropicReq.system
    .filter(block => block.type === 'text' && block.text)
    .map(block => block.text)
    .join('\n');
}

📊 Testing Results

Regression Tests

  • Anthropic Direct: No regressions
  • Google Gemini: No regressions
  • OpenRouter: Fixed from 0% → 80%

MCP Tools

  • All 15 tools working through OpenRouter proxy
  • File operations validated (Write, Read, Bash)
  • Tool format conversion working (Anthropic ↔ OpenAI)

Performance

  • GPT-3.5-turbo: 5s (fastest)
  • Mistral 7B: 6s
  • Gemini 2.0 Flash: 6s
  • GPT-4o-mini: 7s
  • Grok 4 Fast: 8s
  • Claude 3.5 Sonnet: 11s
  • Llama 3.1 8B: 14s

📖 Documentation

Technical Details

Quick Reference

  • 66+ specialized agents available
  • 111 MCP tools for coordination
  • 4 providers: Anthropic, OpenRouter, Gemini, ONNX
  • 400+ models via OpenRouter
  • Zero breaking changes - fully backward compatible

🚀 Example Usage

Basic Code Generation

# With Anthropic (highest quality)
npx agentic-flow@beta --agent coder --task "Create REST API with Express"

# With OpenRouter GPT-4o-mini (best value)
npx agentic-flow@beta --agent coder --task "Create REST API with Express" \
  --provider openrouter --model "openai/gpt-4o-mini"

# With Grok 4 Fast (free!)
npx agentic-flow@beta --agent coder --task "Create REST API with Express" \
  --provider openrouter --model "x-ai/grok-4-fast"

Multi-Agent Workflows

# Research task with cheaper model
npx agentic-flow@beta --agent researcher \
  --task "Research best practices for microservices" \
  --provider openrouter --model "openai/gpt-3.5-turbo"

# Code review with high-quality model
npx agentic-flow@beta --agent reviewer \
  --task "Review my authentication code" \
  --provider openrouter --model "anthropic/claude-3.5-sonnet"

# Testing with fast model
npx agentic-flow@beta --agent tester \
  --task "Create Jest tests for my API" \
  --provider openrouter --model "mistralai/mistral-7b-instruct"

Configuration

# Interactive wizard
npx agentic-flow@beta config

# Set OpenRouter API key
npx agentic-flow@beta config set OPENROUTER_API_KEY "sk-or-..."

# List configuration
npx agentic-flow@beta config list

🐛 Reporting Issues

This is a beta release - please test and report any issues:

GitHub Issues: https://github.com/ruvnet/agentic-flow/issues

When reporting, please include:

  • Model being used
  • Task description
  • Error message (if any)
  • Output received
  • Expected behavior

🔄 Upgrade Path

From v1.1.13 → v1.1.14-beta.1

Changes:

  • OpenRouter proxy now functional (was 100% broken)
  • No breaking changes to API
  • All existing code continues to work
  • New: 8 OpenRouter models now available

Migration:

# Update to beta
npm install agentic-flow@beta

# Or use npx (always gets latest)
npx agentic-flow@beta [commands]

Rollback if needed:

npm install agentic-flow@1.1.13

🎯 Next Steps

Before Stable Release (v1.1.14)

  1. User beta testing feedback
  2. Test DeepSeek models with proper API keys
  3. Debug Llama 3.3 70B timeout issue
  4. Test streaming responses
  5. Performance benchmarking
  6. Additional model validation

Future Enhancements (v1.2.0)

  1. Auto-detect best model for task
  2. Automatic failover between models
  3. Model capability detection
  4. Streaming response support
  5. Cost optimization features
  6. Performance metrics dashboard

📈 Success Metrics

Before v1.1.14-beta.1

  • OpenRouter success rate: 0% (100% failure)
  • Working models: 0
  • Cost savings: Not available
  • User complaints: High

After v1.1.14-beta.1

  • OpenRouter success rate: 80% (8/10 working)
  • Working models: 8
  • Cost savings: Up to 99%
  • MCP tools: All 15 working
  • Most popular model: Working (Grok 4 Fast)

Release Checklist

  • Core bug fixed (anthropicReq.system)
  • 10 models tested (8 working)
  • Popular models validated (Grok 4 Fast)
  • MCP tools working (all 15)
  • File operations confirmed
  • No regressions in baseline providers
  • Documentation complete
  • Changelog updated
  • Package version updated
  • TypeScript build successful
  • Git tag created
  • NPM published with beta tag
  • GitHub release created
  • npx command verified
  • User communication prepared

🙏 Credits

Debugging time: ~4 hours Lines changed: ~50 Models tested: 10 Success rate: 80% Impact: Unlocked 400+ models via OpenRouter

Built with: Claude Code


Ready for production after beta testing! 🚀