A RWKV management and startup tool, full automation, only 8MB. And provides an interface compatible with the OpenAI API. RWKV is a large language model that is fully open source and available for commercial use.
Go to file
2023-05-29 21:34:24 +08:00
.vscode SwitchModelBody.customCuda 2023-05-23 11:51:43 +08:00
backend-golang embed dependencies 2023-05-29 09:39:16 +08:00
backend-python chore 2023-05-29 20:51:20 +08:00
build reduce size 2023-05-29 20:13:06 +08:00
frontend external access to the API Switch 2023-05-29 21:34:24 +08:00
.gitignore embed all core dependencies 2023-05-29 20:14:42 +08:00
exportModelsJson.js update manifest.json 2023-05-07 16:09:16 +08:00
go.mod update 2023-05-17 23:27:52 +08:00
go.sum update 2023-05-17 23:27:52 +08:00
LICENSE navigate card 2023-05-05 13:41:54 +08:00
main.go chore 2023-05-29 20:37:00 +08:00
Makefile update 2023-05-17 21:20:41 +08:00
manifest.json embed all core dependencies 2023-05-29 20:14:42 +08:00
README_ZH.md update readme 2023-05-27 00:38:39 +08:00
README.md update readme 2023-05-27 00:38:39 +08:00
wails.json init 2023-05-03 23:38:54 +08:00

RWKV Runner

This project aims to eliminate the barriers of using large language models by automating everything for you. All you need is a lightweight executable program of just a few megabytes. Additionally, this project provides an interface compatible with the OpenAI API, which means that every ChatGPT client is an RWKV client.

license release

English | 简体中文

FAQs | Preview | Download

Default configs do not enable custom CUDA kernel acceleration, but I strongly recommend that you enable it and run with int8 precision, which is much faster and consumes much less VRAM. Go to the Configs page and turn on Use Custom CUDA kernel to Accelerate.

For different tasks, adjusting API parameters can achieve better results. For example, for translation tasks, you can try setting Temperature to 1 and Top_P to 0.3.

Features

  • RWKV model management and one-click startup
  • Fully compatible with the OpenAI API, making every ChatGPT client an RWKV client. After starting the model, open http://127.0.0.1:8000/docs to view more details.
  • Automatic dependency installation, requiring only a lightweight executable program
  • User-friendly chat interaction interface included
  • Easy-to-understand and operate parameter configuration
  • Built-in model conversion tool
  • Built-in download management and remote model inspection
  • Multilingual localization
  • Theme switching
  • Automatic updates

API Concurrency Stress Testing

ab -p body.json -T application/json -c 20 -n 100 -l http://127.0.0.1:8000/chat/completions

body.json:

{
  "messages": [
    {
      "role": "user",
      "content": "Hello"
    }
  ]
}

Todo

  • Model training functionality
  • CUDA operator int8 acceleration
  • macOS support
  • Linux support

Preview

Homepage

image

Chat

image

Completion

image

Configuration

image

Model Management

image

Download Management

image

Settings

image