machine_learning/ml_infrastructure/inference-server-performance/server/trtis_service.yaml (25 lines of code) (raw):

apiVersion: v1 kind: Service metadata: labels: name: inference-server name: inference-server namespace: default spec: #externalTrafficPolicy: Cluster ports: - name: http-inference-server port: 8000 protocol: TCP targetPort: 8000 - name: grpc-inference-server port: 8001 protocol: TCP targetPort: 8001 - name: metrics-inference-server port: 8002 protocol: TCP targetPort: 8002 selector: app: inference-server sessionAffinity: None type: ClusterIP