Skip to content
This repository has been archived by the owner on Dec 27, 2023. It is now read-only.

cheshire-cat-ai/llama-local

Repository files navigation

llama-local for the Cheshire Cat AI (NVIDIA only)

This is an adaptation of llama-cpp-python (link) to be easily launched from docker-compose and with an NVIDIA GPU.

Clone repo:

git clone https://github.com/cheshire-cat-ai/llama-local.git

Create your .env based on the provided example:

cp .env.example .env

Download the model of your choice (GGML format, many LLAMA versions are available here)

Place your .bin model in the models folder

MODEL_NAME in .env should match the filename of your LLAMA.

Launch the container:

docker compose up

Now go to http://localhost:8000/docs to try out the endpoints

TODO: instructions on how to configure the cat

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •