Effortlessly run LLM backends, APIs, frontends, and services with one command.
-
Updated
Sep 19, 2024 - Shell
Effortlessly run LLM backends, APIs, frontends, and services with one command.
A fast, lightweight, parallel inference server for Llama LLMs.
Various image processing scripts.
LLM translation toolkit for subtitle files.
Add a description, image, and links to the exllamav2 topic page so that developers can more easily learn about it.
To associate your repository with the exllamav2 topic, visit your repo's landing page and select "manage topics."