NVIDIA Triton Inference Server
toolAn open-source inference serving platform that enables deployment of AI models from multiple frameworks with optimized performance for real-time, batched, and streaming inference across cloud, edge, and embedded devices.
ai-ml-infrastructuremodel-servinginference-server+5 more