Back to List
Notice:This resource is provided by a third-party author. Please review the code with AI tools or manually before use to ensure security and compatibility.
C++openvinotoolkit/model_server

model_server

A scalable inference server for models optimized with OpenVINO™

85.1/100
859Forks: 249
View on GitHubHomepage →
Loading report...

Similar Projects

serving

92

A flexible, high-performance serving system for machine learning models

C++6.3K

MNN

94

MNN: A blazing-fast, lightweight inference engine battle-tested by Alibaba, powering high-performance on-device LLMs and Edge AI.

C++15.0K

runanywhere-sdks

80

Production ready toolkit to run AI locally

C++10.3K

deeplake

79

Deeplake is AI Data Runtime for Agents. It provides serverless postgres with a multimodal datalake, enabling scalable retrieval and training.

C++9.1K
Back to List