Optimizing Load Balancing and Autoscaling
for Large Language Model (LLM) Inference on Kubernetes |
David Gray |
Engaging the KServe Community, The Impact of
Integrating a Solutions with Standardized CNCF Projects |
Adam Tetelman, Taneem Ibrahim, Johnu George, Tessa Pham, Andreea Munteanu |
Advancing Cloud Native AI Innovation Through
Open Collaboration |
Yuan Tang |
Unlocking Potential of Large Models in
Production |
Yuan Tang, Adam Tetelman |
WG Serving: Accelerating AI/ML Inference
Workloads on Kubernetes |
Yuan Tang, Eduardo Arango Gutierrez |
Best Practices for Deploying LLM Inference,
RAG and Fine Tuning Pipelines |
Meenakshi Kaushik, Shiva Krishna Merla |
Production AI at Scale: Cloudera's Journey
Adopting KServe |
Zoram Thanga, Peter Ableda |
From Bash Scripts to Kubeflow and
GitOps: Our Journey to Operationalizing ML at Scale |
Luca Grazioli, Dennis Ohrndorf |
Production-Ready AI Platform on
Kubernetes |
Yuan Tang |
Fortifying AI Security in Kubernetes
with Confidential Containers |
Suraj Deshmukh, Pradipta Banerjee |
Platform Building Blocks: How to Build ML Infrastructure with
CNCF Projects |
Yuzhui Liu, Leon Zhou |
Distributed Machine Learning
Patterns from Manning Publications |
Yuan Tang |
KubeCon 2019: Introducing KFServing:
Serverless Model Serving on Kubernetes |
Dan Sun, Ellis Tarn |
KubeCon 2019: Advanced Model Inferencing
Leveraging KNative, Istio & Kubeflow Serving |
Animesh Singh, Clive Cox |
KubeflowDojo: KFServing - Production Model
Serving Platform |
Animesh Singh, Tommy Li |
NVIDIA: Accelerate and Autoscale
Deep Learning Inference on GPUs with KFServing |
Dan Sun, David Goodwin |
KF Community: KFServing - Enabling
Serverless Workloads Across Model Frameworks |
Ellis Tarn |
KubeflowDojo: Demo - KFServing End to End
through Notebook |
Animesh Singh, Tommy Li |
KubeflowDojo: Demo - KFServing with Kafka
and Kubeflow Pipelines |
Animesh Singh |
Anchor
MLOps Podcast: Serving Models with KFServing |
David Aponte, Demetrios Brinkmann |
Kubeflow 101: What is KFServing? |
Stephanie Wong |
ICML
2020, Workshop on Challenges in Deploying and Monitoring Machine Learning Systems : Serverless
inferencing on Kubernetes |
Clive Cox |
Serverless Practitioners Summit 2020:
Serverless Machine Learning Inference with KFServing |
Clive Cox, Yuzhui Liu |
MLOps Meetup: KServe Live Coding Session
|
Theofilos Papapanagiotou |
KubeCon AI Days 2021: Serving Machine
Learning Models at Scale Using KServe |
Yuzhui Liu |
KubeCon 2021: Serving Machine Learning
Models at Scale Using KServe |
Animesh Singh |
KubeCon China 2021: Accelerate Federated
Learning Model Deployment with KServe |
Fangchi Wang & Jiahao Chen |
KubeCon AI Days 2022: Exploring ML Model
Serving with KServe |
Alexa Nicole Griffith |
KubeCon AI Days 2022: Enhancing the
Performance Testing Process for gRPC Model Inferencing at Scale |
Ted Chang, Paul Van Eck |
KubeCon Edge Days 2022: Model Serving at the
Edge Made Easier |
Paul Van Eck |
KnativeCon 2022: How We Built an ML
inference Platform with Knative |
Dan Sun |
KubeCon EU 2023: The state and future of cloud native model
serving |
Dan Sun, Theofilos Papapanagiotou |
Kubeflow Summit 2023: Scale your Models
to Zero with Knative and KServe |
Jooho Lee |
Kubeflow Summit 2023: What to choose?
ModelMesh vs Model Serving? |
Vaibhav Jain |