mirror of
https://github.com/mudler/LocalAI
synced 2026-04-21 13:27:21 +00:00
**Description** This PR fixes #1013. It adds `draft_model` and `n_draft` to the model YAML config in order to load models with speculative sampling. This should be compatible as well with grammars. example: ```yaml backend: llama context_size: 1024 name: my-model-name parameters: model: foo-bar n_draft: 16 draft_model: model-name ``` --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io> |
||
|---|---|---|
| .. | ||
| backend | ||
| config | ||
| localai | ||
| openai | ||
| options | ||
| schema | ||
| api.go | ||
| api_test.go | ||
| apt_suite_test.go | ||