How to Build Deep Learning Inference Through Knative Server... - Huamin Chen & Yehuda Sadeh-Weinraub
Building Machine Learning Inference Through Knative Serverless...- Shivay Lamba & Rishit Dagli
What is vLLM? Efficient AI Inference for Large Language Models
Optimize LLM inference with vLLM
How We Built an ML inference Platform with Knative - Dan Sun, Bloomberg LP & Animesh Singh, IBM
Knative Serverless for AI/ML Applications | Ian Lawson
NVidia TensorRT: high-performance deep learning inference accelerator (TensorFlow Meets)
Agentic RAG vs RAGs
I Made The Smallest (And Dumbest) LLM
TensorFlow in 100 Seconds
OpenAI's nightmare: Deepseek R1 on a Raspberry Pi
Build Your Own AI Supercomputer at Home 🤯
How to choose an embedding model
Local AI has a Secret Weakness
Kubernetes Explained in 6 Minutes | k8s Architecture
This is the coolest AI tool to help you generate diagrams (tech or system design ones especially)!
Use Knative When You Can, and Kubernetes When You Must - David Hadas & Michael Maximilien, IBM
PyTorch in 100 Seconds
Exploring ML Model Serving with KServe (with fun drawings) - Alexa Nicole Griffith, Bloomberg
Don't Make an AI LLM - Do This Instead
What is Retrieval-Augmented Generation (RAG)?
EASIEST Way to Fine-Tune a LLM and Use It With Ollama
Accelerate and Autoscale Deep Learning Inference on GPUs with KFServing - Dan Sun
PyTorch vs. TensorFlow
Building a FaaS Platform Using CNCF OSS Projects (Knative, Crossplane, Tekton... Mauricio Salatino