Name: values.FullName,
Namespace: ips_namespace,
Labels: map[string]string{
- "controller-tools.k8s.io": "1.0",
"app": values.FullName,
},
},
ifsv.Spec.Predictor.Tensorflow = &api_v1beta1.TFServingSpec{
PredictorExtensionSpec: api_v1beta1.PredictorExtensionSpec{
StorageURI: &values.StorageURI,
+ RuntimeVersion: &values.RuntimeVersion,
Container: core_v1.Container{
Image: values.Image,
- Ports: []core_v1.ContainerPort{
- {
- Name: "h2c",
- ContainerPort: 9000,
- Protocol: "TCP",
- },
- },
},
},
}
StorageURI: &values.StorageURI,
Container: core_v1.Container{
Image: values.Image,
- Ports: []core_v1.ContainerPort{
- {
- Name: "h2c",
- ContainerPort: 9000,
- Protocol: "TCP",
- },
- },
},
},
}
Resources string `yaml:"resources"`
RICServiceAccountName string `yaml:"ric_serviceaccount_name"`
StorageURI string `yaml:"storageUri"`
+ RuntimeVersion string `yaml:"runtimeVersion"`
Image string `yaml:"image"`
ResourceVersion string
CanaryTrafficPercent int64
data["engine"] = c.config.InferenceService.Engine
data["storageUri"] = c.config.InferenceService.StorageURI
+ data["runtimeVersion"] = c.config.InferenceService.RuntimeVersion
//data["resources"] = c.config.
data["max_replicas"] = c.config.InferenceService.MaxReplicas
}
type InferenceService struct {
- Engine string `json:"engine"`
- StorageURI string `json:"storage_uri"`
- ApiVersion string `json:"api_version"`
- MinReplicas int32 `json:"min_replicas"`
- MaxReplicas int32 `json:"max_replicas"`
+ Engine string `json:"engine"`
+ StorageURI string `json:"storage_uri"`
+ RuntimeVersion string `json:"runtime_version"`
+ ApiVersion string `json:"api_version"`
+ MinReplicas int32 `json:"min_replicas"`
+ MaxReplicas int32 `json:"max_replicas"`
}
"sa_name": "default",
"inferenceservice": {
"engine": "tensorflow",
- "storage_uri": "s3://mlpipeline/artifacts/sample-rl-pipeline-bbzh4/sample-rl-pipeline-bbzh4-949264085/sample-training-saved-model.tgz",
+ "storage_uri": "<Model URL>",
+ "runtime_version": "2.5.1",
"api_version": "serving.kubeflow.org/v1beta1",
"min_replicas": 1,
"max_replicas": 1
"required": [
"engine",
"storage_uri",
+ "runtime_version",
"api_version",
"min_replicas",
"max_replicas"
"s3://mlpipeline/artifacts/sample-rl-pipeline-bbzh4/sample-rl-pipeline-bbzh4-949264085/sample-training-saved-model.tgz"
]
},
+ "runtime_version": {
+ "$id": "#/properties/inferenceservice/properties/runtime_version",
+ "type": "string",
+ "title": "The runtime version Schema",
+ "default": "",
+ "examples": [
+ "2.5.1"
+ ]
+ },
"api_version": {
"$id": "#/properties/inferenceservice/properties/api_version",
"type": "string",
}
}
-
\ No newline at end of file
+
}
type InferenceServiceProperty struct {
- Engine StringProperty `json:"engine"`
- StorageURI StringProperty `json:"storage_uri"`
- ApiVersion StringProperty `json:"api_version"`
- MinReplicas IntegerProperty `json:"min_replicas"`
- MaxReplicas IntegerProperty `json:"max_replicas"`
+ Engine StringProperty `json:"engine"`
+ StorageURI StringProperty `json:"storage_uri"`
+ RuntimeVersion StringProperty `json:"runtime_version"`
+ ApiVersion StringProperty `json:"api_version"`
+ MinReplicas IntegerProperty `json:"min_replicas"`
+ MaxReplicas IntegerProperty `json:"max_replicas"`
}
type StringProperty struct {