Deploy and manage AI API proxies with GitOps-ready Kubernetes Operators. Automate scaling, self-healing, and lifecycle management for production AI workloads.
Declarative configuration with automatic reconciliation and self-healing
Kubernetes-native API for managing AI proxy configurations
Complete example of deploying an AI proxy with the operator
# AI Proxy Configuration apiVersion: ai-proxy.io/v1 kind: AIProxyConfig metadata: name: openai-config spec: provider: openai apiEndpoint: https://api.openai.com/v1 authSecretRef: name: openai-api-key key: api-key rateLimit: requestsPerSecond: 100 burst: 50 cacheConfig: enabled: true ttl: "5m" --- # AI Proxy Deployment apiVersion: ai-proxy.io/v1 kind: AIProxyDeployment metadata: name: ai-gateway spec: replicas: 3 image: ai-proxy:1.0.0 configRef: openai-config resources: requests: cpu: "500m" memory: "512Mi" limits: cpu: "2000m" memory: "2Gi" hpa: minReplicas: 3 maxReplicas: 20 targetCPUUtilization: 70
The operator handles all aspects of the proxy lifecycle
Deploy the operator using kubectl or Helm
# Install using kubectl kubectl apply -f https://github.com/ai-proxy/operator/releases/download/v1.0.0/crds.yaml kubectl apply -f https://github.com/ai-proxy/operator/releases/download/v1.0.0/operator.yaml # Or using Helm helm repo add ai-proxy https://charts.ai-proxy.io helm install ai-proxy-operator ai-proxy/operator --namespace ai-proxy-system --create-namespace # Verify installation kubectl get pods -n ai-proxy-system kubectl get crd | grep ai-proxy