The guide shows how to train model and create InferenceService for the trained model for on-prem cluster.
Refer to the document to create Persistent Volume (PV) and Persistent Volume Claim (PVC), the PVC will be used to store model.
Follow the mnist example guide to train a mnist model and store it to PVC. The InferenceService is deployed in the notebook example by Kubeflow Fairing
that uses kfserving
SDK. If you want to apply the InferenceService via kubectl by using the YAML format as below, no need to run the deployment step in the notebook. In this example, the relative path of model will be ./export/
on the PVC.
Update the ${PVC_NAME} to the created PVC name in the mnist-pvc.yaml
and apply:
kubectl apply -f mnist-pvc.yaml
Expected Output
$ inferenceservice.serving.kubeflow.org/mnist-sample configured
$ kubectl get inferenceservice
NAME URL READY DEFAULT TRAFFIC CANARY TRAFFIC AGE
mnist-sample http://mnist-sample.kubeflow.example.com/v1/models/mnist-sample True 100 1m