For iOS Developers: If your company already pays for AWS Bedrock or Google Cloud AI, this guide shows how to use those models in Xcode with Alex Sidebar instead of buying separate API keys.
What is LiteLLM?
LiteLLM is an open-source proxy that translates between the OpenAI API format and 100+ different AI providers. Alex Sidebar can work with enterprise AI services that don’t support the OpenAI API format.LiteLLM is what Alex Sidebar uses internally for model connections, making it a well-tested solution for enterprise deployments. Current stable version: v1.73.6-stable (June 2025)
Why Use LiteLLM?
Use Your Company's AI
If your company uses AWS Bedrock or Google Cloud AI, LiteLLM lets you access those models through Alex Sidebar
Data Never Leaves Your Infrastructure
Your code stays within your company’s cloud. No data goes to Alex Sidebar servers
Track Costs by Project
See exactly how much each project costs. Set budgets and get alerts
One Interface for All Models
Switch between Claude 4 on Bedrock, Gemini 2.5 on Vertex, or GPT-4 on Azure without changing code
Quick Start
1
Install LiteLLM
Choose your deployment method:Option 1: pip install (simplest)Option 2: Docker (recommended for production)
2
Configure Your Providers
Create a
config.yaml
file in your LiteLLM directory:3
Connect Alex Sidebar
In Alex Sidebar, add a custom model pointing to your LiteLLM proxy:
- Open Settings → Models → Custom Models
- Click “Add New Model”
- Configure:
- Model ID: Your model name from config.yaml (e.g.,
claude-4-sonnet
) - Base URL: Your LiteLLM URL +
/v1
(e.g.,https://litellm.company.com/v1
) - API Key: Your LiteLLM proxy key (if configured)
- Model ID: Your model name from config.yaml (e.g.,
Provider-Specific Setup
Amazon Bedrock
- Setup
- Authentication
- Ensure your AWS credentials are configured on the LiteLLM server
- Enable the models you need in the AWS Bedrock console
- Add to your LiteLLM config:
Google Vertex AI
- Setup
- Authentication
- Enable the Vertex AI API in your GCP project
- Set up authentication (service account recommended)
- Add to your LiteLLM config:
Azure OpenAI
- Setup
- Authentication
Configure Azure OpenAI with latest models:
Advanced Features
Reasoning and Thinking Capabilities
Enable advanced reasoning for supported models:Multimodal Support
Configure models for text, image, audio, and video:MCP Gateway Integration
Enable Model Context Protocol for enhanced tool use:Team Configuration
For team accounts, you can override all Alex Sidebar model endpoints:- Go to Alex Sidebar Admin Portal
- Navigate to Models tab
- Add your LiteLLM proxy URL as Base URL for each model type
- All team members automatically use your proxy
All AI requests from your team go through your infrastructure. You control the data and costs.
Advanced Configuration
Load Balancing with Fallbacks
Distribute requests across multiple model deployments with intelligent routing:Cost Tracking and Budget Management
Enable comprehensive cost tracking:Security and Rate Limiting
Secure your LiteLLM deployment with advanced controls:Vector Store Integration
Connect to knowledge bases and vector stores:Monitoring & Observability
LiteLLM v1.73.6 provides enhanced monitoring capabilities:Performance Metrics
- 2x Higher RPS: Enhanced aiohttp transport for improved performance
- 50ms Median Latency: Optimized for high-throughput applications
- Multi-instance Rate Limiting: Accurate rate limiting across deployments
Dashboard Features
Real-time Monitoring
Troubleshooting
Connection refused error
Connection refused error
- Verify LiteLLM is running and accessible
- Check firewall rules and security groups
- Ensure you’re using the correct URL format with
/v1
suffix - For Docker: Check port mapping and container status
Authentication errors
Authentication errors
- Bedrock: Verify AWS credentials, IAM permissions, and model access
- Vertex: Check service account permissions and project settings
- Azure: Ensure API keys and resource endpoints are correct
- Verify master key matches if configured
Model not found or deprecated
Model not found or deprecated
- Check model name matches exactly in config.yaml
- Verify the model is enabled in your cloud provider console
- Update to latest model versions (e.g., claude-4 instead of claude-3)
- Check region/location settings and model availability
High latency or rate limiting
High latency or rate limiting
- Enable aiohttp transport:
USE_AIOHTTP_TRANSPORT=True
- Implement load balancing across multiple deployments
- Adjust
max_parallel_requests
and rate limiting settings - Consider regional deployment distribution
Cost tracking issues
Cost tracking issues
- Ensure database connection is properly configured
- Check that
track_cost_callback: true
is set - Verify model pricing information is up to date
- Review spend logs retention settings
Common Use Cases for iOS Teams
Scenario 1: Company Uses AWS with Latest Models
Your company has AWS Bedrock with Claude 4 models. Instead of buying Anthropic API keys:- Deploy LiteLLM v1.73.6-stable on an EC2 instance
- Configure it to use your Bedrock Claude 4 models with reasoning capabilities
- Developers connect Alex Sidebar to your LiteLLM endpoint
- All costs go to your AWS bill with detailed tracking
Scenario 2: Multi-Cloud Model Testing
Test latest models across providers without changing code:Scenario 3: Development vs Production with New Models
Scenario 4: Enterprise Security and Compliance
Enterprise Features (LiteLLM v1.73.6+)
SCIM Integration
Automatic user provisioning from your identity provider:- Okta, Azure AD, OneLogin support
- Automatic team creation and user assignment
- Deprovisioning when users are removed
Advanced Analytics
- Team and tag-based usage tracking
- Daily spend analysis by model and user
- Session grouping and analysis
- Audit logs for compliance
Enhanced Security
- Vector store permissions by team/user
- MCP server access controls
- IP allowlisting and rate limiting
- End-to-end encryption options
Next Steps
- Review LiteLLM’s official documentation for detailed configuration options
- Check the Proxy UI to monitor costs and usage with the new dashboard
- Explore Vector Store integration for RAG applications
- Join the Alex Sidebar Discord for help with enterprise setups
- Contact team@alexcodes.app for business support and enterprise features
LiteLLM v1.73.6-stable gives you control over your AI infrastructure with the latest models and enterprise-grade features, working seamlessly with all Alex Sidebar capabilities.