Download Latest Version koboldcpp-1.100.1 source code.zip (45.6 MB)
Email in envelope

Get an email when there's a new version of KoboldCpp

Home / v1.0.8beta
Name Modified Size InfoDownloads / Week
Parent folder
koboldcpp.exe 2023-04-03 16.4 MB
koboldcpp-1.0.8beta source code.tar.gz 2023-04-03 9.2 MB
koboldcpp-1.0.8beta source code.zip 2023-04-03 9.2 MB
README.md 2023-04-03 1.1 kB
Totals: 4 Items   34.7 MB 0

koboldcpp-1.0.8beta

  • Rebranded to koboldcpp (formerly llamacpp-for-kobold). Library file names and references are changed too, Please let me know if anything is broken!
  • Added support for the original GPT4ALL.CPP format!
  • Added support for GPT-J formats, including the original 16bit legacy format as well as the 4bit version from Pygmalion.cpp
  • Switched compiler flag from -O3 to -Ofast. This should increase generation speed even more, but I dunno if anything will break, please let me know if so.
  • Changed default threads to scale according to physical Core counts instead of os.cpu_count(). This will generally result in fewer threads being utilized, but it should provide a better default for slower systems. You can override this manually with --threads parameter.

To use, download and run the koboldcpp.exe Alternatively, drag and drop a compatible quantized model for llamacpp on top of the .exe, or run it and manually select the model in the popup dialog.

and then once loaded, you can connect like this (or use the full koboldai client): http://localhost:5001

Source: README.md, updated 2023-04-03