mirror of
https://github.com/mudler/LocalAI.git
synced 2025-05-20 10:35:01 +00:00
feat(llama.cpp): add flash_attention
and no_kv_offloading
(#2310)
feat(llama.cpp): add flash_attn and no_kv_offload Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
This commit is contained in:
parent
7123d07456
commit
e49ea0123b
4 changed files with 11 additions and 0 deletions
|
@ -212,6 +212,9 @@ message ModelOptions {
|
|||
float YarnBetaSlow = 47;
|
||||
|
||||
string Type = 49;
|
||||
|
||||
bool FlashAttention = 56;
|
||||
bool NoKVOffload = 57;
|
||||
}
|
||||
|
||||
message Result {
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue