Skip to content

Commit 7909e85

Browse files
authored
llama-run : improve progress bar (ggml-org#10821)
Set default width to whatever the terminal is. Also fixed a small bug around default n_gpu_layers value. Signed-off-by: Eric Curtin <[email protected]>
1 parent 9177484 commit 7909e85

File tree

3 files changed

+304
-126
lines changed

3 files changed

+304
-126
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -448,7 +448,7 @@ To learn more about model quantization, [read this documentation](examples/quant
448448
449449
</details>
450450
451-
[^3]: [https://github.com/containers/ramalama](RamaLama)
451+
[^3]: [RamaLama](https://github.com/containers/ramalama)
452452
453453
## [`llama-simple`](examples/simple)
454454

examples/run/README.md

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ The purpose of this example is to demonstrate a minimal usage of llama.cpp for r
44

55
```bash
66
llama-run granite-code
7-
...
7+
```
88

99
```bash
1010
llama-run -h
@@ -19,6 +19,8 @@ Options:
1919
Context size (default: 2048)
2020
-n, --ngl <value>
2121
Number of GPU layers (default: 0)
22+
-v, --verbose, --log-verbose
23+
Set verbosity level to infinity (i.e. log all messages, useful for debugging)
2224
-h, --help
2325
Show help message
2426

@@ -42,6 +44,6 @@ Examples:
4244
llama-run https://example.com/some-file1.gguf
4345
llama-run some-file2.gguf
4446
llama-run file://some-file3.gguf
45-
llama-run --ngl 99 some-file4.gguf
46-
llama-run --ngl 99 some-file5.gguf Hello World
47-
...
47+
llama-run --ngl 999 some-file4.gguf
48+
llama-run --ngl 999 some-file5.gguf Hello World
49+
```

0 commit comments

Comments
 (0)