How It Works
Configure Provider in Portal
Go to Settings > LLM Providers and add a provider with its credentials (API key, AWS credentials, etc).
Auto-Discover Models
Runflow automatically discovers available models for your provider and shows them in the model picker.
Provider Helpers
The SDK exports a helper function for each provider type:ModelProvider object that tells the runtime which provider and model to use:
Supported Providers
OpenAI
sk-...)
Popular models: gpt-4o, gpt-4o-mini, gpt-4-turbo, o1, o1-mini, o3-mini
Anthropic (Claude)
sk-ant-...)
Popular models: claude-sonnet-4-20250514, claude-3-5-sonnet-20241022, claude-3-5-haiku-20241022, claude-3-opus-20240229
AWS Bedrock
Use Claude, Titan, Llama, and other models through your AWS account — no separate API keys needed, billing goes through AWS.accessKeyId, secretAccessKey, and optionally region)
Popular models: anthropic.claude-3-5-sonnet-20241022-v2:0, anthropic.claude-3-haiku-20240307-v1:0, amazon.titan-text-express-v1, meta.llama3-70b-instruct-v1:0
Groq
Ultra-fast inference for open-source models.gsk_...)
Popular models: llama-3.3-70b-versatile, llama-3.1-8b-instant, mixtral-8x7b-32768, gemma2-9b-it
Google Gemini
AIza...)
Popular models: gemini-2.5-flash, gemini-2.5-pro, gemini-2.0-flash
Azure OpenAI
Use OpenAI models hosted on your Azure subscription. Configure this provider in the portal with your Azure endpoint and deployment name. In code, useopenai() with providerName pointing to your Azure configuration:
endpoint and deploymentName)
Custom (OpenAI-Compatible)
Connect any OpenAI-compatible API — Ollama, LiteLLM, vLLM, LM Studio, or any other provider that follows the OpenAI API format.baseUrl)
Use cases: Self-hosted models, private deployments, specialized inference endpoints
Named Provider Configurations
If you have multiple configurations of the same provider type (e.g., separate OpenAI keys for dev and production), useproviderName to target a specific one:
- Environment isolation: Different API keys for dev/staging/production
- Cost control: Route expensive calls through a specific key with budget limits
- Regional routing: Target specific AWS regions for Bedrock
Using with LLM Standalone
All providers work with direct LLM calls (no agent needed):Model Discovery
When you add a provider in the portal, Runflow can auto-discover available models by querying the provider’s API. Discovered models include metadata like:- Maximum context window size
- Streaming support
- Tool/function calling support
- Vision/multimodal support
- Cost per 1K tokens (input/output)
Next Steps
Agents
Create agents with any provider
LLM Standalone
Direct LLM calls without agents
Custom Memory Provider
Build your own memory backend
Streaming
Real-time streaming responses