Back to List
Notice:This resource is provided by a third-party author. Please review the code with AI tools or manually before use to ensure security and compatibility.
PythonSemiAnalysisAI/InferenceX

InferenceX

Open Source Continuous Inference Benchmarking Qwen3.5, DeepSeek, GPTOSS - GB200 NVL72 vs MI355X vs B200 vs GB300 NVL72 vs H100 & soon™ TPUv6e/v7/Trainium2/3

68.9/100
638Forks: 97
View on GitHubHomepage →
Loading report...

Similar Projects

LMCache

87

Supercharge Your LLM with the Fastest KV Cache Layer

Python7.6K

vllm

93

A high-throughput and memory-efficient inference and serving engine for LLMs

Python72.4K

ramalama

87

RamaLama is an open-source developer tool that simplifies the local serving of AI models from any source and facilitates their use for inference in production, all through the familiar language of containers.

Python2.6K

ml-engineering

74

Machine Learning Engineering Open Book

Python17.3K
Back to List