mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2025-01-12 03:31:46 +00:00
llama : increase scratch buffer size for 65B (ref #1152)
Temporary solution
This commit is contained in:
parent
9b0a4d4214
commit
957c8ae21d
@ -54,7 +54,7 @@ static const std::map<e_model, size_t> & MEM_REQ_SCRATCH0()
|
|||||||
{ MODEL_7B, 512ull * MB },
|
{ MODEL_7B, 512ull * MB },
|
||||||
{ MODEL_13B, 512ull * MB },
|
{ MODEL_13B, 512ull * MB },
|
||||||
{ MODEL_30B, 512ull * MB },
|
{ MODEL_30B, 512ull * MB },
|
||||||
{ MODEL_65B, 512ull * MB },
|
{ MODEL_65B, 1024ull * MB },
|
||||||
};
|
};
|
||||||
return _MEM_REQ_SCRATCH0;
|
return _MEM_REQ_SCRATCH0;
|
||||||
}
|
}
|
||||||
@ -65,7 +65,7 @@ static const std::map<e_model, size_t> & MEM_REQ_SCRATCH1()
|
|||||||
{ MODEL_7B, 512ull * MB },
|
{ MODEL_7B, 512ull * MB },
|
||||||
{ MODEL_13B, 512ull * MB },
|
{ MODEL_13B, 512ull * MB },
|
||||||
{ MODEL_30B, 512ull * MB },
|
{ MODEL_30B, 512ull * MB },
|
||||||
{ MODEL_65B, 512ull * MB },
|
{ MODEL_65B, 1024ull * MB },
|
||||||
};
|
};
|
||||||
return _MEM_REQ_SCRATCH1;
|
return _MEM_REQ_SCRATCH1;
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user