vLLM
An open source project that optimizes the speed and affordability of deploying large language models for inference.
Overview
| Category | Ai |
| Self-Hostable | No |
| On-Prem | No |
An open source project that optimizes the speed and affordability of deploying large language models for inference.
| Category | Ai |
| Self-Hostable | No |
| On-Prem | No |