Cortex.cpp: model.yaml Format #1123
dan-menlo
started this conversation in
Architecture Specs
Replies: 3 comments 2 replies
-
Questions @nguyenhoangthuan99 @sangjanai @namchuai
See related issues above. |
Beta Was this translation helpful? Give feedback.
0 replies
-
Here is the default setting for a gguf-model:
other information will be infer from the gguf metadata.
|
Beta Was this translation helpful? Give feedback.
0 replies
-
@nguyenhoangthuan99 Can we tame this complexity by defining the following:
name: Llama 3.1
model: llama3.1
version: 1
stop:
- <|end_of_text|>
- <|eot_id|>
- <|eom_id|>
top_p: 0.9
temperature: 0.6
frequency_penalty: 0
presence_penalty: 0
max_tokens: 8192
stream: true
ngl: 33
ctx_len: 8192
engine: cortex.llamacpp
prompt_template: |+
<|begin_of_text|><|start_header_id|>system<|end_header_id|>
{system_message}<|eot_id|><|start_header_id|>user<|end_header_id|>
{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
files:
- /Users/nicolezhu/cortex/models/llama3.1/model.gguf
id: llama3.1
created: 1722850990703
object: model
owned_by: '' i.e.
|
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Goal
GGUF
TBA
TensorRT-LLM
TBA
ONNX
TBA
Key Epics
Related
model.yaml
for Model Downloaded via URL jan#3558Beta Was this translation helpful? Give feedback.
All reactions