Skip to content

sirmo/koboldcpp-rocm-docker

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

koboldcpp-rocm-docker

Docker build for running koboldcpp-rocm

Building the image

  • clone the ROCm fork of koboldcpp, inside the local directory of this repo
git clone https://github.com/YellowRoseCx/koboldcpp-rocm.git
  • run the Docker build (this will take awhile)
docker build -t kobold:latest .
  • start the container (assumes the .gguf LLM models are located in ~/models, change accordingly)
docker run --rm -it -p 5001:5001 --device /dev/kfd --device /dev/dri \
--mount type=bind,source="$HOME"/models,target=/models \
kobold:latest
  • run koboldcpp server (inside the container)
./koboldcpp.py --config example_config.json

Additional info

This was tested on 7900xtx

For other architectures modify the ENV HSA_OVERRIDE_GFX_VERSION=11.0.0 line in Dockerfile to the appropriate value.

Some examples:

    gfx900: Vega 64
    gfx906: Radeon VII
    gfx908: Instinct MI100
    gfx90a: Instinct MI200/ MI250
    gfx1030: RX 6800/ 6900/ 6950, also works on many other RDNA2 and RDNA1 cards with the environment variable "HSA_OVERRIDE_GFX_VERSION=10.3.0"
    gfx1100: RX 7900
    gfx1101: Unreleased, probably RX 7700/ 7800?
    gfx1102: RX 7600

About

Docker build for running koboldcpp-rocm

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published