#
TabbyAPI
A FastAPI based application that allows for generating text using an LLM using the Exllamav2 backend, with support for Exl2, GPTQ, and FP16 models.
#
Quickstart
- Follow the installation instructions on the official TabbyAPI GitHub.
- Create your config.yml to set your model path, default model, sequence length, etc. You can ignore most (if not all) of these settings if you want.
- Launch TabbyAPI. If it worked, you should see something like this:
- Under the Text Completion API in SillyTavern, select TabbyAPI.
- Copy your API key from the TabbyAPI terminal into
Tabby API key
and make sure yourAPI URL
is correct (it should behttp://127.0.0.1:5000
by default).
If you did everything correctly, you should see something like this in SillyTavern:
You can now chat using TabbyAPI!
#
TabbyAPI Loader
The developers of TabbyAPI created an official extension to load/unload models directly from SillyTavern. Installation is simple:
- In SillyTavern, click on the Extensions tab and navigate to Download Extensions & Assets.
- Copy
https://raw.githubusercontent.com/theroyallab/ST-repo/main/index.json
into Assets URL and click the plug button to the right. - You should see something like this. Click the download button next to Tabby Loader.
- If the installation was successful, you should see a green pop-up message at the top of your screen. Under the extensions tab, navigate to TabbyAPI Loader and copy your admin key from the TabbyAPI terminal into Admin Key.
- Click the refresh button next to Model Select. When you click on the textbox just below it, you should see all of the models in your model directory.
You can now load and unload your models directly from SillyTavern!
#
Support
Still need help? Visit the TabbyAPI GitHub for a link to the developer's official Discord server and read the wiki.