File tree Expand file tree Collapse file tree 2 files changed +5
-2
lines changed Expand file tree Collapse file tree 2 files changed +5
-2
lines changed Original file line number Diff line number Diff line change 2
2
BUILD_TARGETS = \
3
3
main quantize quantize-stats perplexity embedding vdot q8dot train-text-from-scratch convert-llama2c-to-ggml \
4
4
simple batched batched-bench save-load-state server gguf llama-bench libllava.a llava-cli baby-llama beam-search \
5
- speculative infill benchmark-matmult parallel finetune export-lora tests/test-c.o
5
+ speculative infill tokenize benchmark-matmult parallel finetune export-lora tests/test-c.o
6
6
7
7
# Binaries only useful for tests
8
8
TEST_TARGETS = \
@@ -594,6 +594,9 @@ infill: examples/infill/infill.cpp ggml.o llama.o $(C
594
594
simple : examples/simple/simple.cpp ggml.o llama.o $(COMMON_DEPS ) $(OBJS )
595
595
$(CXX ) $(CXXFLAGS ) $(filter-out % .h,$^ ) -o $@ $(LDFLAGS )
596
596
597
+ tokenize : examples/tokenize/tokenize.cpp ggml.o llama.o $(COMMON_DEPS ) $(OBJS )
598
+ $(CXX ) $(CXXFLAGS ) $(filter-out % .h,$^ ) -o $@ $(LDFLAGS )
599
+
597
600
batched : examples/batched/batched.cpp ggml.o llama.o $(COMMON_DEPS ) $(OBJS )
598
601
$(CXX ) $(CXXFLAGS ) $(filter-out % .h,$^ ) -o $@ $(LDFLAGS )
599
602
Original file line number Diff line number Diff line change @@ -26,7 +26,7 @@ int main(int argc, char ** argv) {
26
26
llama_context_params ctx_params = llama_context_default_params ();
27
27
llama_context * ctx = llama_new_context_with_model (model, ctx_params);
28
28
29
- const bool add_bos = true ;
29
+ const bool add_bos = llama_should_add_bos_token (model) ;
30
30
31
31
std::vector<llama_token> tokens;
32
32
You can’t perform that action at this time.
0 commit comments