Topics
TensorFlow Serving
toolA production-grade serving system for machine learning models that provides high-performance inference with gRPC and REST APIs, model versioning, and batching capabilities for TensorFlow and other ML frameworks.
9/3/2025
1 pages
machine-learningmodel-servingai-ml-infrastructure+4 more
NVIDIA Triton Inference Server
toolAn open-source inference serving platform that enables deployment of AI models from multiple frameworks with optimized performance for real-time, batched, and streaming inference across cloud, edge, and embedded devices.
8/28/2025
1 pages
ai-ml-infrastructuremodel-servinginference-server+5 more