jp6/cu129/: vllm versions
Because this project isn't in the mirror_whitelist,
no releases from root/pypi are included.
Latest version on stage is: 0.11.2+cu129
A high-throughput and memory-efficient inference and serving engine for LLMs
| Index | Version | Documentation |
|---|---|---|
| jp6/cu129 | 0.11.2+cu129 | |
| jp6/cu129 | 0.10.2+cu129 | |
| jp6/cu129 | 0.9.3+cu129 | |
| jp6/cu129 | 0.8.6 |