Summary
Add native support for Together AI platform as a first-class provider in the Copilot SDK. This includes seamless integration with Together AI's diverse model catalog, real-time inference API, and enterprise-grade features like batch processing and usage monitoring.
Problem / Use Case
- Growing Together AI Adoption: More teams are migrating to Together AI for cost-effective, open-source model inference at scale
- Feature Parity: Currently, accessing Together AI requires manual API integration without built-in fallback chains or routing strategies (e.g., falling back to open-source models when proprietary APIs are rate-limited)
- Consistency Gaps: No standardized way to handle Together AI alongside other providers (OpenAI, Anthropic, etc.) in a single SDK
Proposed Solution
1. Together AI Provider Implementation
- Add
TogetherAIProvider class with full support for:
- Text Generation: 70+ open-source models (Llama, Mistral, Falcon, etc.)
- Batch Processing API: Asynchronous batch job submission for large-scale processing
- Embeddings: Via Together AI embeddings endpoint
- Streaming: Real-time token streaming with proper error handling
2. Model Registry Integration
- Register Together AI models in the SDK's model registry with metadata:
- Input/output token limits
- Pricing per 1M tokens
- Context window size
- Quantization levels available
3. Fallback & Routing Compatibility
- Enable fallback chains across Together AI models (e.g., fallback from Llama-70B to Llama-7B on rate limits)
- Support routing strategies (priority, round-robin) specifically optimized for Together AI's load balancing
4. Authentication & Configuration
- Environment variable support:
TOGETHER_API_KEY
- Configurable base URL for self-hosted Together inference endpoints
- Request timeout and retry logic aligned with Together's SLA
Alternatives Considered
- Using Together AI's SDK directly - Lacks centralized error handling and routing across multiple providers
- Generic HTTP client - Would require duplicating error handling and rate limit logic
- Async wrapper layer - Insufficient for seamless integration with existing fallback/routing infrastructure
Additional Context
- Together AI API Reference: https://docs.together.ai/reference
- Community request for multi-provider support with open-source models
- Competitive advantage: Better support for cost-sensitive applications## Summary
Summary
Add native support for Together AI platform as a first-class provider in the Copilot SDK. This includes seamless integration with Together AI's diverse model catalog, real-time inference API, and enterprise-grade features like batch processing and usage monitoring.
Problem / Use Case
Proposed Solution
1. Together AI Provider Implementation
TogetherAIProviderclass with full support for:2. Model Registry Integration
3. Fallback & Routing Compatibility
4. Authentication & Configuration
TOGETHER_API_KEYAlternatives Considered
Additional Context