v3/@claude-flow/providers/README.md
Multi-LLM Provider System for Claude Flow V3 - unified interface for Anthropic, OpenAI, Google, Cohere, Ollama, and RuVector with intelligent load balancing, automatic failover, and cost optimization.
claude-3-5-sonnet-20241022, claude-3-5-sonnet-latestclaude-3-opus-20240229claude-3-sonnet-20240229claude-3-haiku-20240307gpt-4o, gpt-4o-minigpt-4-turbo, gpt-4gpt-3.5-turboo1-preview, o1-mini, o3-minigemini-2.0-flashgemini-1.5-pro, gemini-1.5-flashgemini-procommand-r-plus, command-rcommand-light, commandllama3.2, llama3.1mistral, mixtralcodellama, phi-4deepseek-codernpm install @claude-flow/providers
import { createProviderManager } from '@claude-flow/providers';
// Create provider manager with multiple providers
const manager = await createProviderManager({
providers: [
{
provider: 'anthropic',
apiKey: process.env.ANTHROPIC_API_KEY!,
model: 'claude-3-5-sonnet-latest',
},
{
provider: 'openai',
apiKey: process.env.OPENAI_API_KEY!,
model: 'gpt-4o',
},
],
loadBalancing: {
enabled: true,
strategy: 'cost-based',
},
fallback: {
enabled: true,
maxAttempts: 2,
},
cache: {
enabled: true,
ttl: 300000,
maxSize: 1000,
},
});
// Make a completion request
const response = await manager.complete({
messages: [
{ role: 'user', content: 'Hello, how are you?' }
],
maxTokens: 100,
});
console.log('Response:', response.content);
console.log('Provider:', response.provider);
console.log('Cost:', response.cost?.totalCost);
import { ProviderManager, createProviderManager } from '@claude-flow/providers';
// Create and initialize
const manager = await createProviderManager(config);
// Or manually initialize
const manager = new ProviderManager(config);
await manager.initialize();
// Complete a request
const response = await manager.complete(request, preferredProvider?);
// Stream completion
for await (const event of manager.streamComplete(request, preferredProvider?)) {
if (event.type === 'content') {
process.stdout.write(event.delta?.content || '');
}
}
// Health check all providers
const health = await manager.healthCheck();
health.forEach((result, provider) => {
console.log(`${provider}: ${result.healthy ? 'OK' : 'FAIL'}`);
});
// Estimate costs across providers
const estimates = await manager.estimateCost(request);
estimates.forEach((estimate, provider) => {
console.log(`${provider}: $${estimate.estimatedCost.total.toFixed(4)}`);
});
// Get aggregated usage
const usage = await manager.getUsage('day');
console.log(`Total cost: $${usage.cost.total}`);
// Get provider metrics
const metrics = manager.getMetrics();
metrics.forEach((m, provider) => {
console.log(`${provider}: ${m.latency}ms avg, ${m.errorRate * 100}% errors`);
});
// List available providers
const providers = manager.listProviders();
// Get specific provider
const anthropic = manager.getProvider('anthropic');
// Clear cache
manager.clearCache();
// Shutdown
manager.destroy();
import {
AnthropicProvider,
OpenAIProvider,
GoogleProvider,
CohereProvider,
OllamaProvider,
RuVectorProvider,
} from '@claude-flow/providers';
// Create provider directly
const anthropic = new AnthropicProvider({
config: {
provider: 'anthropic',
apiKey: process.env.ANTHROPIC_API_KEY!,
model: 'claude-3-5-sonnet-latest',
temperature: 0.7,
maxTokens: 1000,
},
});
await anthropic.initialize();
const response = await anthropic.complete({
messages: [{ role: 'user', content: 'Hello!' }],
});
// Stream response
for await (const event of anthropic.streamComplete({
messages: [{ role: 'user', content: 'Tell me a story' }],
})) {
if (event.type === 'content') {
process.stdout.write(event.delta?.content || '');
}
}
const manager = await createProviderManager({
providers: [...],
loadBalancing: {
enabled: true,
strategy: 'round-robin', // or 'least-loaded', 'latency-based', 'cost-based'
},
});
| Strategy | Description | Best For |
|---|---|---|
round-robin | Rotate through providers | Even distribution |
least-loaded | Use provider with lowest load | High throughput |
latency-based | Use fastest provider | Low latency |
cost-based | Use cheapest provider | Cost optimization |
const manager = await createProviderManager({
providers: [...],
fallback: {
enabled: true,
maxAttempts: 3, // Try up to 3 providers
},
});
// Events for monitoring
manager.on('fallback_success', ({ originalProvider, fallbackProvider, attempts }) => {
console.log(`Fallback from ${originalProvider} to ${fallbackProvider}`);
});
manager.on('fallback_exhausted', ({ originalProvider, attempts }) => {
console.error(`All ${attempts} fallback attempts failed`);
});
const manager = await createProviderManager({
providers: [...],
cache: {
enabled: true,
ttl: 300000, // 5 minutes
maxSize: 1000, // Max cached responses
},
});
// Cache is keyed by: messages + model + temperature + maxTokens
// Identical requests return cached responses
// Clear cache when needed
manager.clearCache();
const manager = await createProviderManager({
providers: [...],
costOptimization: {
enabled: true,
maxCostPerRequest: 0.10, // Max $0.10 per request
},
});
// Request with cost constraints
const response = await manager.complete({
messages: [...],
costConstraints: {
maxCost: 0.05,
preferredModels: ['gpt-4o-mini', 'claude-3-haiku-20240307'],
},
});
// Get cost breakdown
console.log('Cost:', {
prompt: response.cost?.promptCost,
completion: response.cost?.completionCost,
total: response.cost?.totalCost,
currency: response.cost?.currency,
});
const response = await manager.complete({
messages: [
{ role: 'user', content: 'What is the weather in San Francisco?' }
],
tools: [
{
type: 'function',
function: {
name: 'get_weather',
description: 'Get weather for a location',
parameters: {
type: 'object',
properties: {
location: { type: 'string', description: 'City name' },
unit: { type: 'string', enum: ['celsius', 'fahrenheit'] },
},
required: ['location'],
},
},
},
],
toolChoice: 'auto',
});
// Check for tool calls
if (response.toolCalls) {
for (const call of response.toolCalls) {
console.log('Tool:', call.function.name);
console.log('Args:', JSON.parse(call.function.arguments));
}
}
// Async iterator streaming
for await (const event of manager.streamComplete(request)) {
switch (event.type) {
case 'content':
process.stdout.write(event.delta?.content || '');
break;
case 'tool_call':
console.log('Tool call:', event.delta?.toolCall);
break;
case 'error':
console.error('Error:', event.error);
break;
case 'done':
console.log('\nUsage:', event.usage);
console.log('Cost:', event.cost);
break;
}
}
// Image input (OpenAI, Anthropic, Google)
const response = await manager.complete({
messages: [
{
role: 'user',
content: [
{ type: 'text', text: 'What is in this image?' },
{ type: 'image', imageUrl: 'https://example.com/image.jpg' },
// or imageBase64: 'base64-encoded-image'
],
},
],
});
import type {
// Provider types
LLMProvider,
LLMModel,
LLMProviderConfig,
ILLMProvider,
ProviderCapabilities,
// Message types
LLMMessage,
LLMContentPart,
LLMToolCall,
LLMTool,
// Request/Response
LLMRequest,
LLMResponse,
LLMStreamEvent,
// Manager types
ProviderManagerConfig,
LoadBalancingStrategy,
// Status types
HealthCheckResult,
ProviderStatus,
CostEstimate,
UsageStats,
// Error types
LLMProviderError,
RateLimitError,
AuthenticationError,
ModelNotFoundError,
ProviderUnavailableError,
} from '@claude-flow/providers';
import {
LLMProviderError,
RateLimitError,
AuthenticationError,
isLLMProviderError,
isRateLimitError,
} from '@claude-flow/providers';
try {
const response = await manager.complete(request);
} catch (error) {
if (isRateLimitError(error)) {
console.log(`Rate limited. Retry after: ${error.retryAfter}ms`);
} else if (isLLMProviderError(error)) {
console.log(`Provider error: ${error.code}`);
console.log(`Retryable: ${error.retryable}`);
}
}
| Error | Code | HTTP | Retryable |
|---|---|---|---|
RateLimitError | RATE_LIMIT | 429 | Yes |
AuthenticationError | AUTHENTICATION | 401 | No |
ModelNotFoundError | MODEL_NOT_FOUND | 404 | No |
ProviderUnavailableError | PROVIDER_UNAVAILABLE | 503 | Yes |
# API Keys
ANTHROPIC_API_KEY=sk-ant-...
OPENAI_API_KEY=sk-...
GOOGLE_API_KEY=...
COHERE_API_KEY=...
# Ollama (local)
OLLAMA_BASE_URL=http://localhost:11434
# Optional overrides
ANTHROPIC_BASE_URL=https://api.anthropic.com
OPENAI_BASE_URL=https://api.openai.com/v1
{
provider: 'anthropic',
apiKey: process.env.ANTHROPIC_API_KEY!,
model: 'claude-3-5-sonnet-latest',
temperature: 0.7,
maxTokens: 4096,
timeout: 60000,
retryAttempts: 3,
}
{
provider: 'openai',
apiKey: process.env.OPENAI_API_KEY!,
model: 'gpt-4o',
temperature: 0.7,
maxTokens: 4096,
frequencyPenalty: 0,
presencePenalty: 0,
}
{
provider: 'google',
apiKey: process.env.GOOGLE_API_KEY!,
model: 'gemini-1.5-pro',
temperature: 0.7,
maxTokens: 8192,
topK: 40,
}
{
provider: 'cohere',
apiKey: process.env.COHERE_API_KEY!,
model: 'command-r-plus',
temperature: 0.7,
maxTokens: 4000,
}
{
provider: 'ollama',
apiUrl: 'http://localhost:11434',
model: 'llama3.2',
temperature: 0.7,
maxTokens: 4096,
}
| Metric | Target |
|---|---|
| Provider initialization | <500ms |
| Request routing | <5ms |
| Cache lookup | <1ms |
| Health check | <2s |
import { createProviderManager } from '@claude-flow/providers';
import { createUnifiedSwarmCoordinator } from '@claude-flow/swarm';
// Create provider manager
const providers = await createProviderManager({
providers: [
{ provider: 'anthropic', apiKey: '...', model: 'claude-3-5-sonnet-latest' },
{ provider: 'openai', apiKey: '...', model: 'gpt-4o' },
],
loadBalancing: { enabled: true, strategy: 'cost-based' },
});
// Use with swarm coordinator
const coordinator = createUnifiedSwarmCoordinator({
// Agents can use provider manager for LLM calls
extensions: {
providers,
},
});
MIT