Skip to content

Commit f9ee715

Browse files
committed
update configuration examples for multiline yaml commands #133
1 parent 2d00120 commit f9ee715

File tree

2 files changed

+13
-13
lines changed

2 files changed

+13
-13
lines changed

README.md

Lines changed: 9 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -46,14 +46,14 @@ llama-swap's configuration is purposefully simple.
4646
models:
4747
"qwen2.5":
4848
proxy: "http://127.0.0.1:9999"
49-
cmd: >
49+
cmd: |
5050
/app/llama-server
5151
-hf bartowski/Qwen2.5-0.5B-Instruct-GGUF:Q4_K_M
5252
--port 9999
5353
5454
"smollm2":
5555
proxy: "http://127.0.0.1:9999"
56-
cmd: >
56+
cmd: |
5757
/app/llama-server
5858
-hf bartowski/SmolLM2-135M-Instruct-GGUF:Q4_K_M
5959
--port 9999
@@ -82,7 +82,7 @@ startPort: 10001
8282
models:
8383
"llama":
8484
# multiline for readability
85-
cmd: >
85+
cmd: |
8686
llama-server --port 8999
8787
--model path/to/Qwen2.5-1.5B-Instruct-Q4_K_M.gguf
8888
@@ -123,7 +123,7 @@ models:
123123
# Docker Support (v26.1.4+ required!)
124124
"docker-llama":
125125
proxy: "http://127.0.0.1:${PORT}"
126-
cmd: >
126+
cmd: |
127127
docker run --name dockertest
128128
--init --rm -p ${PORT}:8080 -v /mnt/nvme/models:/models
129129
ghcr.io/ggerganov/llama.cpp:server
@@ -247,11 +247,11 @@ Pre-built binaries are available for Linux, FreeBSD and Darwin (OSX). These are
247247
1. Create a configuration file, see [config.example.yaml](config.example.yaml)
248248
1. Download a [release](https://github.com/mostlygeek/llama-swap/releases) appropriate for your OS and architecture.
249249
1. Run the binary with `llama-swap --config path/to/config.yaml`.
250-
Available flags:
251-
- `--config`: Path to the configuration file (default: `config.yaml`).
252-
- `--listen`: Address and port to listen on (default: `:8080`).
253-
- `--version`: Show version information and exit.
254-
- `--watch-config`: Automatically reload the configuration file when it changes. This will wait for in-flight requests to complete then stop all running models (default: `false`).
250+
Available flags:
251+
- `--config`: Path to the configuration file (default: `config.yaml`).
252+
- `--listen`: Address and port to listen on (default: `:8080`).
253+
- `--version`: Show version information and exit.
254+
- `--watch-config`: Automatically reload the configuration file when it changes. This will wait for in-flight requests to complete then stop all running models (default: `false`).
255255

256256
### Building from source
257257

config.example.yaml

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ groups:
1515

1616
models:
1717
"llama":
18-
cmd: >
18+
cmd: |
1919
models/llama-server-osx
2020
--port ${PORT}
2121
-m models/Llama-3.2-1B-Instruct-Q4_0.gguf
@@ -38,7 +38,7 @@ models:
3838
# Embedding example with Nomic
3939
# https://huggingface.co/nomic-ai/nomic-embed-text-v1.5-GGUF
4040
"nomic":
41-
cmd: >
41+
cmd: |
4242
models/llama-server-osx --port ${PORT}
4343
-m models/nomic-embed-text-v1.5.Q8_0.gguf
4444
--ctx-size 8192
@@ -51,15 +51,15 @@ models:
5151
# Reranking example with bge-reranker
5252
# https://huggingface.co/gpustack/bge-reranker-v2-m3-GGUF
5353
"bge-reranker":
54-
cmd: >
54+
cmd: |
5555
models/llama-server-osx --port ${PORT}
5656
-m models/bge-reranker-v2-m3-Q4_K_M.gguf
5757
--ctx-size 8192
5858
--reranking
5959
6060
# Docker Support (v26.1.4+ required!)
6161
"dockertest":
62-
cmd: >
62+
cmd: |
6363
docker run --name dockertest
6464
--init --rm -p ${PORT}:8080 -v /mnt/nvme/models:/models
6565
ghcr.io/ggerganov/llama.cpp:server

0 commit comments

Comments
 (0)