Closed
Description
Ref: #8687 (comment)
(cc @ggerganov)
TODO:
- Train some adapters based on stories15M and stories15M_MOE
- Test with
llama-cli -m base_model.gguf --lora lora_adapter.gguf
- Test merging using
llama-export-lora
, then re-run the merged.gguf to verify it outputs the same thing as above
Optionally: make some small stories model with different arch, for example gemma, phi,...