!LuUSGaeArTeoOgUpwk:matrix.org

kubeflow-kfserving

433 Members
2 Servers

Load older messages


SenderMessageTime
14 Oct 2021
@_slack_kubeflow_U02HA8R0NQ7:matrix.org_slack_kubeflow_U02HA8R0NQ7 joined the room.03:21:00
@_slack_kubeflow_UFVUV2UFP:matrix.orgDan Sun Suresh Nakkiran I do not see any reason why we can't go from 0.5.0 to 0.7.0 as long as the version supports v1beta1 API and you are still using the same version of istiod/knative 05:02:24
@_slack_kubeflow_UFVUV2UFP:matrix.orgDan Sun actually to be precise if your istio/knative version meets the min requirement of kserve 0.7 then the migration path from 0.5 to 0.7 should work in my mind, but we will test out to make sure 05:14:18
@_slack_kubeflow_UFVUV2UFP:matrix.orgDan Sun Peilun Li what's your istio/knative version used along side with KFS 0.5 05:15:35
@_slack_kubeflow_UFVUV2UFP:matrix.orgDan Sun
In reply toundefined
(edited) ... KFS 0.5 => ... KFS 0.5?
05:15:39
@_slack_kubeflow_U02AF95BSUF:matrix.orgSuresh Nakkiran Dan Sun In quick install we have istio 1.6.2.. 06:21:41
@_slack_kubeflow_UFVUV2UFP:matrix.orgDan Sun we should have istio 1.9 there 06:37:30
@_slack_kubeflow_U0104H1616Z:matrix.orgiamlovingit
In reply to@_slack_kubeflow_U02HTJ9GLKT:matrix.org
Hi, I have installed Kubeflow 1.4 from manifests and I tried to deploy the sklearn-iris predictor sample and I am getting the following error: Any ideas on a way forward? Many thanks 🙏
Riley Green Hi, can you show kfserving logs here?
08:02:04
@_slack_kubeflow_U0113251MGS:matrix.orgPeilun Li We are currently on K8S 1.19, Istio 1.8, Knative 0.17 -- Definitely some work on our end to upgrade to Istio 1.9 and Knative 0.23+ before we upgrade kserve to 0.6/0.7 16:42:16
15 Oct 2021
@_slack_kubeflow_U0253V66WCT:matrix.orgYulong joined the room.02:42:51
@_slack_kubeflow_U0253V66WCT:matrix.orgYulong Hi, guys. I am using kfserving v0.6.0 on kubeflow v1.3.0. I was trying to use replace or patch function to apply a new version of TensorFlow model, but it it seems that after running the code, the InferenceService and pod didn’t changed.
predictor_2 = V1beta1PredictorSpec(tensorflow=V1beta1TFServingSpec(storage_uri=storage_uri,
                                                                         image=image),
                                        min_replicas=1,
                                        max_replicas=3,
                                        canary_traffic_percent=50,
                                       )

isvc_2 = V1beta1InferenceService(api_version="serving.kubeflow.org/v1beta1",
                          kind=constants.KFSERVING_KIND,
                          metadata=client.V1ObjectMeta(name=predictor_name, namespace=namespace),
                          spec=V1beta1InferenceServiceSpec(predictor=predictor_2))

kfs.replace(name=predictor_name, inferenceservice=isvc_2, namespace=namespace)
(base) jovyan@tensorboard-normal-0:~$ k get InferenceService
NAME     URL                                    READY   PREV   LATEST   PREVROLLEDOUTREVISION   LATESTREADYREVISION              AGE
demo   http://demo.test-ns.example.com          True           100                              demo-predictor-default-5sn57   8m59s
What should I do to make it right?
02:52:37
@_slack_kubeflow_U0253V66WCT:matrix.orgYulong (edited) Hi, guys. I am ... => Hi, guys. I am ... 02:52:50
@_slack_kubeflow_U0253V66WCT:matrix.orgYulong By the way, if I use different storage_uri , it will replace or patch successfully, but we are using TFX. Pusher, a component of TFX, saves trained-model under a folder, e.g., ./models/1, /models/2, … we need to specify the storage_uri to the ./models level for loading saved model of TF. In this case, kfserving won’t do replace or patch. Does any possible way to deal with it? 06:40:20
@_slack_kubeflow_UFVUV2UFP:matrix.orgDan Sun
In reply to@_slack_kubeflow_U0253V66WCT:matrix.org
By the way, if I use different storage_uri , it will replace or patch successfully, but we are using TFX. Pusher, a component of TFX, saves trained-model under a folder, e.g., ./models/1, /models/2, … we need to specify the storage_uri to the ./models level for loading saved model of TF. In this case, kfserving won’t do replace or patch. Does any possible way to deal with it?
Yulong that's by design, if the storage uri is not changed then nothing is deployed as we want to make storage immutable. But I think in this case you can add an annotation to indicate the version change then InferenceService will redeploy and load the correct version
13:02:29
@_slack_kubeflow_UFVUV2UFP:matrix.orgDan Sun
In reply to@_slack_kubeflow_UFVUV2UFP:matrix.org
Yulong that's by design, if the storage uri is not changed then nothing is deployed as we want to make storage immutable. But I think in this case you can add an annotation to indicate the version change then InferenceService will redeploy and load the correct version
Any change should be surfaced on inference service yaml which is the whole point of declarative deployment
13:03:40
@_slack_kubeflow_U027LHY3610:matrix.orgVedant Padwal joined the room.13:49:25
@_slack_kubeflow_U027LHY3610:matrix.orgVedant Padwal changed their display name from jordan sumitomo to Vedant Padwal.18:55:44
@_slack_kubeflow_U027LHY3610:matrix.orgVedant Padwal changed their profile picture.18:55:45
@_slack_kubeflow_U027LHY3610:matrix.orgVedant Padwal Hi Everyone, The KServe v0.7 release blog is out!, please check it out at https://kserve.github.io/website/blog/articles/2021-10-11-KServe-0.7-release/ 18:55:51
@_slack_kubeflow_U02JP9VPHH6:matrix.orgKevin Hu joined the room.21:04:14
@_slack_kubeflow_U02JP9VPHH6:matrix.orgKevin Hu changed their display name from _slack_kubeflow_U02JP9VPHH6 to Kevin Hu.21:15:48
@_slack_kubeflow_U02JP9VPHH6:matrix.orgKevin Hu set a profile picture.21:15:51
16 Oct 2021
@_slack_kubeflow_U01JXCGKCLS:matrix.orgNagaraj Janardhana joined the room.04:26:21
@_slack_kubeflow_U02CK0DRH38:matrix.orgMarek Opęchowski joined the room.07:37:37
@_slack_kubeflow_UD4QSK9EW:matrix.orgAndrey Velichkevich joined the room.08:38:23
18 Oct 2021
@_slack_kubeflow_U0253V66WCT:matrix.orgYulong
In reply to@_slack_kubeflow_UFVUV2UFP:matrix.org
Any change should be surfaced on inference service yaml which is the whole point of declarative deployment
Dan Sun Thank you for your kind response. That is a possible solution to make it.
03:09:54
@_slack_kubeflow_U025D42L7J9:matrix.org_slack_kubeflow_U025D42L7J9 joined the room.06:58:59
@_slack_kubeflow_U0140TXAN4T:matrix.org_slack_kubeflow_U0140TXAN4T joined the room.08:05:42
@_slack_kubeflow_U0241BFPK6J:matrix.orgmetafeather joined the room.12:09:05
@_slack_kubeflow_U02HQ5HFBLP:matrix.org_slack_kubeflow_U02HQ5HFBLP joined the room.14:48:45

Show newer messages


Back to Room ListRoom Version: 6