About TorchServe
An open-source model serving framework for PyTorch models that provides REST/gRPC inference endpoints, model management, and production-oriented features (note: project documentation indicates limited maintenance status).
Key Features
- Model management and multi-model serving with REST and gRPC endpoints.
- Supports batching, custom handlers, and integrations for optimized inference (TorchScript, ONNX, TensorRT).
- Workflows for chaining models and endpoints to create DAGs of inference logic.
- Documentation and tooling for packaging models (model-archiver) and deployment.
Use Cases & Best For
About Model Serving & APIs
Deploy and serve ML models