-
Notifications
You must be signed in to change notification settings - Fork 111
Open
Description
Feature Summary
Add helm chart support for deploying LiteLLM proxy as part of the Texera Kubernetes deployment. This enables AI agent functionality by providing a unified LLM API gateway that can route requests to various LLM providers (Anthropic, OpenAI, etc.).
The LiteLLM proxy acts as a central point for:
- Managing API keys for multiple LLM providers
- Providing a unified OpenAI-compatible API for all models
- Tracking usage and spend across different models
- Enabling the access control service to interact with LLM models for agent features
Proposed Solution or Design
The implementation includes:
-
New Helm Templates:
litellm-config.yaml- ConfigMap for LiteLLM configurationlitellm-deployment.yaml- Deployment for the LiteLLM proxy containerlitellm-secret.yaml- Secret for API keys (Anthropic, OpenAI) and database credentialslitellm-service.yaml- ClusterIP service for internal accesspostgresql-litellm-persistence.yaml- PV/PVC for LiteLLM's PostgreSQL database
-
Chart Dependencies:
- Add postgresql-litellm as an aliased postgresql dependency for LiteLLM's persistence
-
Service Integration:
- Update access-control-service deployment to include LITELLM_MASTER_KEY and LITELLM_BASE_URL environment variables
- Add ExternalName services for cross-namespace access
- Add ingress paths for
/api/modelsand/api/chatendpoints
-
Configuration:
- Add
litellmsection in values.yaml with configurable image, resources, API keys, and persistence settings - Add
postgresql-litellmsection for the dedicated LiteLLM database
- Add
Impact / Priority
(P2)Medium – useful enhancement
Affected Area
- Deployment / Infrastructure
Metadata
Metadata
Assignees
Labels
No labels