STAC Research Report: LLM Model Serving Platform Comparison

A comparison of model-serving platforms

13 February 2025

We recently conducted a study comparing two model-serving platforms - vLLM and Hugging Face’s text-generation-inference (TGI) - for LLM inference using the STAC-AI™ LANG6 (Inference-Only) Test Harness.

The STAC-AI™ benchmark provides rigorous, industry-standard testing for LLM inference infrastructure, helping firms assess performance, efficiency, and reliability in real-world conditions.

This research note examines Inference Rate, Throughput, and Fidelity, offering valuable insights for firms optimizing their LLM infrastructure.

We tested four workloads, combining short- and long-context datasets with both 8B and 70B parameter models. Key areas of analysis include:

  • Differences in Inference Rate across platforms.
  • The impact of platform versions on serving efficiency.
  • Variations in response consistency.
  • Patterns of non-determinism and how they manifest in generated outputs.

Premium subscribers have access to this research note, which includes visualizations of all test results, and configuration information for the solutions tested. Subscribers also can run STAC-AI benchmarks in the privacy of their own labs to test their systems. To learn about subscription options, please contact us.

About STAC News

Read the latest about research, events, and other important news from STAC.

Subscribe to notifications of research, events, and more.

(If you're a human, don't change the following field)
Your first name.
(If you're a human, don't change the following field)
Your first name.

Enter your email above, then click "Sign Up" to join the STAC mail list and (optionally) register to access materials on the site. Click for terms.