Download Latest Version koboldcpp-1.100.1 source code.zip (45.6 MB)
Email in envelope

Get an email when there's a new version of KoboldCpp

Home / v1.0.6beta
Name Modified Size InfoDownloads / Week
Parent folder
llamacpp-for-kobold.exe 2023-03-29 16.7 MB
llamacpp-for-kobold-1.0.6-beta source code.tar.gz 2023-03-29 8.8 MB
llamacpp-for-kobold-1.0.6-beta source code.zip 2023-03-29 8.8 MB
README.md 2023-03-29 945 Bytes
Totals: 4 Items   34.3 MB 0

llamacpp-for-kobold-1.0.6-beta

  • This is an experimental release containing new integrations for OpenBLAS, which should increase initial prompt processing speed on compatible systems by over 2 times!
  • Updated Embedded Kobold Lite with the latest version which supports pseudo token streaming. This should make the UI feel much more responsive during prompt generation.
  • Switched to argparse, you can view all command line flags with llamacpp-for-kobold.exe --help
  • To disable OpenBLAS, you can run it with --noblas. Please tell me if you have issues with it, and include which specific OS and platform.

To use, download and run the llamacpp-for-kobold.exe Alternatively, drag and drop a compatible quantized model for llamacpp on top of the .exe, or run it and manually select the model in the popup dialog.

and then once loaded, you can connect like this (or use the full koboldai client): http://localhost:5001

Source: README.md, updated 2023-03-29