machine_learning/ml_infrastructure/inference-server-performance/server/trtis_service.yaml (25 lines of code) (raw):
apiVersion: v1
kind: Service
metadata:
labels:
name: inference-server
name: inference-server
namespace: default
spec:
#externalTrafficPolicy: Cluster
ports:
- name: http-inference-server
port: 8000
protocol: TCP
targetPort: 8000
- name: grpc-inference-server
port: 8001
protocol: TCP
targetPort: 8001
- name: metrics-inference-server
port: 8002
protocol: TCP
targetPort: 8002
selector:
app: inference-server
sessionAffinity: None
type: ClusterIP