Skip to content

Commit 6041736

Browse files
authored
Update README.md
1 parent 3e481d0 commit 6041736

File tree

1 file changed

+1
-26
lines changed

1 file changed

+1
-26
lines changed

README.md

Lines changed: 1 addition & 26 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ This combines the [LLaMA foundation model](https://github.com/facebookresearch/l
1212

1313
Download the zip file corresponding to your operating system from the [latest release](https://github.com/antimatter15/alpaca.cpp/releases/latest). On Windows, download `alpaca-win.zip`, on Mac (both Intel or ARM) download `alpaca-mac.zip`, and on Linux (x64) download `alpaca-linux.zip`.
1414

15-
Download [ggml-alpaca-7b-q4.bin](https://huggingface.co/Sosaka/Alpaca-native-4bit-ggml/blob/main/ggml-alpaca-7b-q4.bin) and place it in the same folder as the `chat` executable in the zip file. There are several options:
15+
Download `ggml-alpaca-7b-q4.bin` and place it in the same folder as the `chat` executable in the zip file. There are several options:
1616

1717
Once you've downloaded the model weights and placed them into the same directory as the `chat` or `chat.exe` executable, run:
1818

@@ -22,31 +22,6 @@ Once you've downloaded the model weights and placed them into the same directory
2222

2323
The weights are based on the published fine-tunes from `alpaca-lora`, converted back into a pytorch checkpoint with a [modified script](https://github.com/tloen/alpaca-lora/pull/19) and then quantized with llama.cpp the regular way.
2424

25-
## Getting Started (13B)
26-
27-
If you have more than 10GB of RAM, you can use the higher quality 13B `ggml-alpaca-13b-q4.bin` model.
28-
29-
Once you've downloaded the weights, you can run the following command to enter chat
30-
31-
```
32-
./chat -m ggml-alpaca-13b-q4.bin
33-
```
34-
35-
## Getting Started (30B)
36-
37-
If you have more than 32GB of RAM (and a beefy CPU), you can use the higher quality 30B `alpaca-30B-ggml.bin` model. To download the weights, you can use
38-
39-
```
40-
git clone https://huggingface.co/Pi3141/alpaca-30B-ggml
41-
```
42-
43-
Once you've downloaded the weights, you can run the following command to enter chat
44-
45-
```
46-
./chat -m ggml-model-q4_0.bin
47-
```
48-
49-
5025
## Building from Source (MacOS/Linux)
5126

5227

0 commit comments

Comments
 (0)