Home

Awesome

<p align="center">

ExUI

This is a simple, lightweight browser-based UI for running local inference using ExLlamaV2.

Overview of features

Screenshots

chat_screenshot chat_screenshot chat_screenshot chat_screenshot chat_screenshot chat_screenshot chat_screenshot chat_screenshot

Running locally

First, clone this repository and install requirements:

git clone https://github.com/turboderp/exui
cd exui
pip install -r requirements.txt

Then run the web server with the included server.py:

python server.py

Your browser should automatically open on the default IP/port. Config and sessions are stored in ~/exui by default.

Prebuilt wheels for ExLlamaV2 are available here. Installing the latest version of Flash Attention is recommended.

Running in Google Colab

An example Colab notebook is provided here.

Installation

More detailed installation instructions can be found here.

More to come

Stay tuned.

avatar_unicorn.png