The most flexible way to serve AI/ML models in production - Build Model ...
A fast, easy-to-use, production-ready inference server for computer visi...
The simplest way to serve AI/ML models in production
Python + Inference - Model Deployment library in Python. Simplest model ...
A REST API for Caffe using Docker and Go
This is a repository for an nocode object detection inference API using ...
This is a repository for an object detection inference API using the Ten...
Orkhon: ML Inference Framework and Server Runtime
Serving AI/ML models in the open standard formats PMML and ONNX with bot...
K3ai is a lightweight, fully automated, AI infrastructure-in-a-box solut...
Deploy DL/ ML inference pipelines with minimal extra code.
ONNX Runtime Server: The ONNX Runtime Server is a server that provides T...
A standalone inference server for trained Rubix ML estimators.
Wingman is the fastest and easiest way to run Llama models on your PC or...