Looks like they don't compare to TensorRT-LLM throughput numbers which, last I checked, are SOTA in open source.
Generation benchmark was 5% faster than SGLang.
Looks like they don't compare to TensorRT-LLM throughput numbers which, last I checked, are SOTA in open source.