-
Notifications
You must be signed in to change notification settings - Fork 120
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: storage phase 1 for inference service reconciler #56
Conversation
@njhill @pvaneck Quick question, is the In Kserve, the pytorch server assumes both model and graph .mar files are under the In Kserve triton server, it assumes the path is pointing to a folder with structure similar to tensorflow's savedmodel format. I think the |
Storage *StorageSpec `json:"storage,omitempty"` | ||
} | ||
|
||
type StorageSpec struct { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can we discuss this? we do not want to maintain two versions of InferenceService
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm going to create the same spec update to the kserve repo once we agreed on how to handle some of the current edge cases on modelmesh storage. The long term goal is to move the modelmesh serving code as part of the kserve reconciler.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We have a conversation on this and the |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks @Tomcli, I made a few comments inline
secretKey := inferenceService.ObjectMeta.Annotations[v1beta1.SecretKeyAnnotation] | ||
if schemaPath, ok := inferenceService.ObjectMeta.Annotations[v1beta1.SchemaPathAnnotation]; ok { | ||
p.Spec.SchemaPath = &schemaPath | ||
} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There appears to be duplication here with below. It would be best to move the annotation checks into the ProcessInferenceServiceStorage function. And only need to check for the annotations if the corresponding explicit fields aren't set.
Re |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks @Tomcli, sorry for the delay in re-reviewing!
thanks @njhill for your reviews |
config/default/kustomization.yaml
Outdated
@@ -94,3 +94,4 @@ resources: | |||
- ../crd | |||
- ../rbac | |||
- ../manager | |||
namespace: modelmesh-serving |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This looks unrelated to the PR?
@Tomcli are you sure we want to include the new May be better to omit it until the other kserve changes are ready? The PR could still be merged, I don't think the new logic will break anything when used with the existing CRD... the storage spec will just appear empty. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hey @Tomcli sorry I noticed a few more things
thanks @njhill for your reviews. I updated based on your comments |
Thanks @njhill I addressed the pointer changes. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks @Tomcli
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: njhill, Tomcli The full list of commands accepted by this bot can be found here. The pull request process is described here
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
/lgtm |
Motivation
rebase #32 to the new inference service reconciler for model mesh
For Storage Spec details, please refer to the design doc: https://docs.google.com/document/d/1rYNh93XNMRp8YaR56m-_zK3bqpZhhGKgpVbLPPsi-Ow/edit#
Additional storages/parameters support will come in phase 2.
Modifications
Result