llama.cpp/examples/parallel
2023-11-02 17:51:53 -04:00
..
CMakeLists.txt build : link against build info instead of compiling against it (#3879) 2023-11-02 08:50:16 +02:00
parallel.cpp fix includes with help from include-what-you-use 2023-11-02 17:51:53 -04:00
README.md llama : custom attention mask + parallel decoding + no context swaps (#3228) 2023-09-28 19:04:36 +03:00

llama.cpp/example/parallel

Simplified simluation for serving incoming requests in parallel