Skip to content

Commit 73f11f6

Browse files
ericcurtinarthw
authored andcommitted
llama-run : improve progress bar (ggml-org#10821)
Set default width to whatever the terminal is. Also fixed a small bug around default n_gpu_layers value. Signed-off-by: Eric Curtin <ecurtin@redhat.com>
1 parent 44792a2 commit 73f11f6

File tree

3 files changed

+304
-126
lines changed

3 files changed

+304
-126
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -482,7 +482,7 @@ To learn more about model quantization, [read this documentation](examples/quant
482482
483483
</details>
484484
485-
[^3]: [https://github.com/containers/ramalama](RamaLama)
485+
[^3]: [RamaLama](https://github.com/containers/ramalama)
486486
487487
## [`llama-simple`](examples/simple)
488488

examples/run/README.md

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ The purpose of this example is to demonstrate a minimal usage of llama.cpp for r
44

55
```bash
66
llama-run granite-code
7-
...
7+
```
88

99
```bash
1010
llama-run -h
@@ -19,6 +19,8 @@ Options:
1919
Context size (default: 2048)
2020
-n, --ngl <value>
2121
Number of GPU layers (default: 0)
22+
-v, --verbose, --log-verbose
23+
Set verbosity level to infinity (i.e. log all messages, useful for debugging)
2224
-h, --help
2325
Show help message
2426

@@ -42,6 +44,6 @@ Examples:
4244
llama-run https://example.com/some-file1.gguf
4345
llama-run some-file2.gguf
4446
llama-run file://some-file3.gguf
45-
llama-run --ngl 99 some-file4.gguf
46-
llama-run --ngl 99 some-file5.gguf Hello World
47-
...
47+
llama-run --ngl 999 some-file4.gguf
48+
llama-run --ngl 999 some-file5.gguf Hello World
49+
```

0 commit comments

Comments
 (0)