| Name | Modified | Size | Downloads / Week |
|---|---|---|---|
| Parent folder | |||
| koboldcpp.exe | 2024-04-20 | 313.7 MB | |
| koboldcpp_nocuda.exe | 2024-04-20 | 39.7 MB | |
| koboldcpp-linux-x64-nocuda | 2024-04-20 | 58.0 MB | |
| koboldcpp-linux-x64 | 2024-04-20 | 394.8 MB | |
| koboldcpp-1.63 source code.tar.gz | 2024-04-20 | 16.9 MB | |
| koboldcpp-1.63 source code.zip | 2024-04-20 | 17.1 MB | |
| README.md | 2024-04-20 | 1.9 kB | |
| Totals: 7 Items | 840.3 MB | 0 | |
koboldcpp-1.63
Enable Sound, Press Play
https://github.com/LostRuins/koboldcpp/assets/39025047/ef8dd0f5-b634-4bed-8142-7f7da87338e0
- Added support for special tokens in
stop_sequences. Thus, if you set<|eot_id|>as a stop sequence and it can be tokenized into a single token, it will just work and function like the EOS token, allowing multiple EOS-like tokens. - Reworked the Automatic RoPE scaling calculations to support Llama3 (just specify the desired
--contextsizeand it will trigger automatically). - Added a console warning if another program is already using the desired port.
- Improved server handling for bad or empty requests, which fixes a potential flooding vulnerability.
- Fixed a scenario where the BOS token could get lost, potentially resulting in lower quality especially during context-shifting.
- Pulled and merged new model support, improvements and fixes from upstream.
- Updated Kobold Lite: Fixed markdown, reworked memory layout, added a regex replacer feature, added aesthetic background color settings, added more save slots, added usermod saving, added Llama3 prompt template
Edit: Something seems to be flagging the CI built binary on windows defender. Replaced it with a locally built one until I can figure it out.
To use, download and run the koboldcpp.exe, which is a one-file pyinstaller. If you don't need CUDA, you can use koboldcpp_nocuda.exe which is much smaller. If you're using AMD, you can try koboldcpp_rocm at YellowRoseCx's fork here
Run it from the command line with the desired launch parameters (see --help), or manually select the model in the GUI.
and then once loaded, you can connect like this (or use the full koboldai client):
http://localhost:5001
For more information, be sure to run the program from command line with the --help flag.