Download Latest Version koboldcpp-1.100.1 source code.zip (45.6 MB)
Email in envelope

Get an email when there's a new version of KoboldCpp

Home / v1.0.7
Name Modified Size InfoDownloads / Week
Parent folder
llamacpp-for-kobold.exe 2023-04-01 15.7 MB
llamacpp-for-kobold-1.0.7 source code.tar.gz 2023-04-01 8.9 MB
llamacpp-for-kobold-1.0.7 source code.zip 2023-04-01 8.9 MB
README.md 2023-04-01 795 Bytes
Totals: 4 Items   33.5 MB 0

llamacpp-for-kobold-1.0.7

  • Added support for new version of the ggml llamacpp model format (magic=ggjt, version 3). All old versions will continue to be supported.
  • Integrated speed improvements from parent repo.
  • Fixed an encoding issue with utf-8 in the outputs.
  • Improved console debug information during generation, now shows token progress and time taken directly.
  • Set non-streaming to be the default mode. You can enable streaming with --stream

To use, download and run the llamacpp-for-kobold.exe Alternatively, drag and drop a compatible quantized model for llamacpp on top of the .exe, or run it and manually select the model in the popup dialog.

and then once loaded, you can connect like this (or use the full koboldai client): http://localhost:5001

Source: README.md, updated 2023-04-01