AI Providers
Alma supports multiple AI providers, giving you flexibility in choosing models and managing costs.
Supported Providers
| Provider | Models | Streaming | Tools | Vision |
|---|---|---|---|---|
| OpenAI | GPT-4o, GPT-4, GPT-3.5 | ✅ | ✅ | ✅ |
| Anthropic | Claude 3.5, Claude 3 | ✅ | ✅ | ✅ |
| Google Gemini | Gemini Pro, Gemini Flash | ✅ | ✅ | ✅ |
| DeepSeek | DeepSeek Chat, DeepSeek Coder | ✅ | ✅ | ❌ |
| Azure OpenAI | All Azure-hosted models | ✅ | ✅ | ✅ |
| OpenRouter | 100+ models | ✅ | ✅ | Varies |
| Custom | Any OpenAI-compatible API | ✅ | Varies | Varies |
Adding a Provider
- Open Settings (
Cmd+,/Ctrl+,) - Navigate to Providers
- Click Add Provider
- Select the provider type
- Enter your API key and configuration
- Click Save
Provider Configuration
Each provider has these common settings:
- Name - A display name for the provider
- API Key - Your authentication key
- Base URL (optional) - Custom endpoint URL
- Enabled - Toggle to enable/disable the provider
Managing Models
After adding a provider, Alma automatically fetches available models. You can:
- Fetch Models - Click to refresh the model list from the provider
- Enable/Disable Models - Toggle which models appear in the model selector
- Add Custom Models - Manually add model IDs not in the fetched list
Multiple Providers
You can configure multiple providers of the same type (e.g., two OpenAI accounts with different API keys). Each will appear separately in the model selector.
Testing Providers
After configuration, test your provider:
- Click the Test button in provider settings
- Alma sends a simple request to verify the connection
- If successful, you'll see a confirmation message
Best Practices
API Key Security
- Alma stores API keys locally and encrypts them
- Never share your API keys
- Use separate keys for development and production
Cost Management
- Use different models for different tasks (e.g., GPT-3.5 for simple queries)
- Monitor usage through provider dashboards
- Consider OpenRouter for pay-per-use pricing across providers
Performance
- Providers closer to your location typically have lower latency
- Some models support caching for faster responses
- Use streaming for better perceived performance
