-
Notifications
You must be signed in to change notification settings - Fork 1.2k
Insights: kserve/kserve
Overview
Could not load contribution data
Please try again later
5 Pull requests merged by 5 people
-
Stop and resume a model [Raw Deployment]
#4455 merged
Jun 11, 2025 -
[Bug] Fixes error in trace logging
#4514 merged
Jun 11, 2025 -
Fix: do not update poetry dependency when install hf cpu deps
#4516 merged
Jun 11, 2025 -
Fix pss restricted warnings
#4327 merged
Jun 6, 2025 -
Initial segregation of the storage module from KServe SDK
#4391 merged
Jun 5, 2025
2 Pull requests opened by 2 people
-
Moved webhdfs dependencies to pyproject.toml file
#4518 opened
Jun 11, 2025 -
[WIP] Add the option to configure knative ns from values file
#4519 opened
Jun 12, 2025
4 Issues closed by 3 people
-
Huggingface Server trace logging throw error
#4515 closed
Jun 11, 2025 -
one question about custom runtime, thanks very much.
#4512 closed
Jun 8, 2025 -
Support for Multiple ContainerStorageContainers
#4361 closed
Jun 5, 2025 -
Make storage initializer install only what is needed for it to run
#3489 closed
Jun 5, 2025
5 Issues opened by 4 people
-
Unified LLM Inference Service API and disaggregated p/d serving support
#4520 opened
Jun 12, 2025 -
Can't create Serverless InferenceService when knative is installed in non-default namespace
#4517 opened
Jun 11, 2025 -
Container fails to start with CUDA version error when using `kserve/huggingfaceserver:latest-gpu`
#4513 opened
Jun 10, 2025 -
Predictor config is required but marked as Optional in Model class
#4511 opened
Jun 5, 2025 -
Predictor health check - support custom preprocess function that mutates model name and version
#4510 opened
Jun 5, 2025
25 Unresolved conversations
Sometimes conversations happen on old items that aren’t yet closed. Here is a list of all the Issues and Pull Requests with unresolved conversations.
-
feat: support remote storage URI injection for serving runtimes
#4492 commented on
Jun 11, 2025 • 5 new comments -
ci: PR style check
#4499 commented on
Jun 12, 2025 • 0 new comments -
Allow support with latest xgboost models
#4493 commented on
Jun 8, 2025 • 0 new comments -
Huggingface ARM64 Support & Refactor multi-platform build workflows
#4480 commented on
Jun 12, 2025 • 0 new comments -
4380 - Inference logging to blob storage
#4473 commented on
Jun 12, 2025 • 0 new comments -
fix: Allow CA bundle path without config map
#4451 commented on
Jun 8, 2025 • 0 new comments -
Add id2label support, fix CUDA bug in return_probabilities
#4444 commented on
Jun 11, 2025 • 0 new comments -
fix: 4439 - update telepresence-setup.sh to adopt tooling changes
#4440 commented on
Jun 12, 2025 • 0 new comments -
Refactor LocalModelNodeAgent
#4431 commented on
Jun 12, 2025 • 0 new comments -
feat: refactor storage initializer resources configuration
#4411 commented on
Jun 11, 2025 • 0 new comments -
chore: update image push from docker to ghcr
#4358 commented on
Jun 12, 2025 • 0 new comments -
Auto-update annotation for isvc.
#4342 commented on
Jun 11, 2025 • 0 new comments -
A new DistributedInfereneceService CRD
#4433 commented on
Jun 12, 2025 • 0 new comments -
REST API Support for Creating InferenceService Resources
#4432 commented on
Jun 11, 2025 • 0 new comments -
Harmonizing OCI Image model support
#4083 commented on
Jun 9, 2025 • 0 new comments -
KServe Easy Deploy: Helm-based Onboarding Experience for ML Developers
#4393 commented on
Jun 8, 2025 • 0 new comments -
Remove torchserve from KServe?
#4469 commented on
Jun 8, 2025 • 0 new comments -
Allow more verbose inference client logging
#4479 commented on
Jun 8, 2025 • 0 new comments -
Enable User -Supplied Schemas for Payload Logging
#4484 commented on
Jun 8, 2025 • 0 new comments -
Track code coverage changes for each PR and possibly add a minimum coverage check
#4502 commented on
Jun 8, 2025 • 0 new comments -
MIG support for huggingface runtime
#4505 commented on
Jun 8, 2025 • 0 new comments -
How can you deploy a model artifact from Kubeflow Pipelines with KServe?
#4269 commented on
Jun 7, 2025 • 0 new comments -
KServe CPU Spikes in transformer
#4509 commented on
Jun 6, 2025 • 0 new comments -
Support multiple StorageUri in Inference Service
#3413 commented on
Jun 5, 2025 • 0 new comments -
Extend Model Caching to Serverless with Cloud DataCache Integration
#4408 commented on
Jun 5, 2025 • 0 new comments