mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-27 03:44:35 +00:00
3fd62a6b1c
* py : type-check all Python scripts with Pyright * server-tests : use trailing slash in openai base_url * server-tests : add more type annotations * server-tests : strip "chat" from base_url in oai_chat_completions * server-tests : model metadata is a dict * ci : disable pip cache in type-check workflow The cache is not shared between branches, and it's 250MB in size, so it would become quite a big part of the 10GB cache limit of the repo. * py : fix new type errors from master branch * tests : fix test-tokenizer-random.py Apparently, gcc applies optimisations even when pre-processing, which confuses pycparser. * ci : only show warnings and errors in python type-check The "information" level otherwise has entries from 'examples/pydantic_models_to_grammar.py', which could be confusing for someone trying to figure out what failed, considering that these messages can safely be ignored even though they look like errors.
330 lines
10 KiB
Nix
330 lines
10 KiB
Nix
{
|
|
lib,
|
|
glibc,
|
|
config,
|
|
stdenv,
|
|
mkShell,
|
|
runCommand,
|
|
cmake,
|
|
ninja,
|
|
pkg-config,
|
|
git,
|
|
python3,
|
|
mpi,
|
|
blas,
|
|
cudaPackages,
|
|
darwin,
|
|
rocmPackages,
|
|
vulkan-headers,
|
|
vulkan-loader,
|
|
curl,
|
|
useBlas ? builtins.all (x: !x) [
|
|
useCuda
|
|
useMetalKit
|
|
useRocm
|
|
useVulkan
|
|
] && blas.meta.available,
|
|
useCuda ? config.cudaSupport,
|
|
useMetalKit ? stdenv.isAarch64 && stdenv.isDarwin,
|
|
useMpi ? false, # Increases the runtime closure size by ~700M
|
|
useRocm ? config.rocmSupport,
|
|
enableCurl ? true,
|
|
useVulkan ? false,
|
|
llamaVersion ? "0.0.0", # Arbitrary version, substituted by the flake
|
|
|
|
# It's necessary to consistently use backendStdenv when building with CUDA support,
|
|
# otherwise we get libstdc++ errors downstream.
|
|
effectiveStdenv ? if useCuda then cudaPackages.backendStdenv else stdenv,
|
|
enableStatic ? effectiveStdenv.hostPlatform.isStatic,
|
|
precompileMetalShaders ? false
|
|
}@inputs:
|
|
|
|
let
|
|
inherit (lib)
|
|
cmakeBool
|
|
cmakeFeature
|
|
optionals
|
|
strings
|
|
versionOlder
|
|
;
|
|
|
|
stdenv = throw "Use effectiveStdenv instead";
|
|
|
|
suffices =
|
|
lib.optionals useBlas [ "BLAS" ]
|
|
++ lib.optionals useCuda [ "CUDA" ]
|
|
++ lib.optionals useMetalKit [ "MetalKit" ]
|
|
++ lib.optionals useMpi [ "MPI" ]
|
|
++ lib.optionals useRocm [ "ROCm" ]
|
|
++ lib.optionals useVulkan [ "Vulkan" ];
|
|
|
|
pnameSuffix =
|
|
strings.optionalString (suffices != [ ])
|
|
"-${strings.concatMapStringsSep "-" strings.toLower suffices}";
|
|
descriptionSuffix =
|
|
strings.optionalString (suffices != [ ])
|
|
", accelerated with ${strings.concatStringsSep ", " suffices}";
|
|
|
|
executableSuffix = effectiveStdenv.hostPlatform.extensions.executable;
|
|
|
|
# TODO: package the Python in this repository in a Nix-like way.
|
|
# It'd be nice to migrate to buildPythonPackage, as well as ensure this repo
|
|
# is PEP 517-compatible, and ensure the correct .dist-info is generated.
|
|
# https://peps.python.org/pep-0517/
|
|
#
|
|
# TODO: Package up each Python script or service appropriately, by making
|
|
# them into "entrypoints"
|
|
llama-python = python3.withPackages (
|
|
ps: [
|
|
ps.numpy
|
|
ps.sentencepiece
|
|
]
|
|
);
|
|
|
|
# TODO(Green-Sky): find a better way to opt-into the heavy ml python runtime
|
|
llama-python-extra = python3.withPackages (
|
|
ps: [
|
|
ps.numpy
|
|
ps.sentencepiece
|
|
ps.tiktoken
|
|
ps.torchWithoutCuda
|
|
ps.transformers
|
|
|
|
# server bench
|
|
ps.matplotlib
|
|
|
|
# server tests
|
|
ps.openai
|
|
ps.behave
|
|
ps.prometheus-client
|
|
|
|
# for examples/pydantic-models-to-grammar-examples.py
|
|
ps.docstring-parser
|
|
ps.pydantic
|
|
|
|
# for scripts/compare-llama-bench.py
|
|
ps.gitpython
|
|
ps.tabulate
|
|
]
|
|
);
|
|
|
|
xcrunHost = runCommand "xcrunHost" {} ''
|
|
mkdir -p $out/bin
|
|
ln -s /usr/bin/xcrun $out/bin
|
|
'';
|
|
|
|
# apple_sdk is supposed to choose sane defaults, no need to handle isAarch64
|
|
# separately
|
|
darwinBuildInputs =
|
|
with darwin.apple_sdk.frameworks;
|
|
[
|
|
Accelerate
|
|
CoreVideo
|
|
CoreGraphics
|
|
]
|
|
++ optionals useMetalKit [ MetalKit ];
|
|
|
|
cudaBuildInputs = with cudaPackages; [
|
|
cuda_cccl.dev # <nv/target>
|
|
|
|
# A temporary hack for reducing the closure size, remove once cudaPackages
|
|
# have stopped using lndir: https://github.com/NixOS/nixpkgs/issues/271792
|
|
cuda_cudart.dev
|
|
cuda_cudart.lib
|
|
cuda_cudart.static
|
|
libcublas.dev
|
|
libcublas.lib
|
|
libcublas.static
|
|
];
|
|
|
|
rocmBuildInputs = with rocmPackages; [
|
|
clr
|
|
hipblas
|
|
rocblas
|
|
];
|
|
|
|
vulkanBuildInputs = [
|
|
vulkan-headers
|
|
vulkan-loader
|
|
];
|
|
in
|
|
|
|
effectiveStdenv.mkDerivation (
|
|
finalAttrs: {
|
|
pname = "llama-cpp${pnameSuffix}";
|
|
version = llamaVersion;
|
|
|
|
# Note: none of the files discarded here are visible in the sandbox or
|
|
# affect the output hash. This also means they can be modified without
|
|
# triggering a rebuild.
|
|
src = lib.cleanSourceWith {
|
|
filter =
|
|
name: type:
|
|
let
|
|
noneOf = builtins.all (x: !x);
|
|
baseName = baseNameOf name;
|
|
in
|
|
noneOf [
|
|
(lib.hasSuffix ".nix" name) # Ignore *.nix files when computing outPaths
|
|
(lib.hasSuffix ".md" name) # Ignore *.md changes whe computing outPaths
|
|
(lib.hasPrefix "." baseName) # Skip hidden files and directories
|
|
(baseName == "flake.lock")
|
|
];
|
|
src = lib.cleanSource ../../.;
|
|
};
|
|
|
|
postPatch = ''
|
|
substituteInPlace ./ggml/src/ggml-metal.m \
|
|
--replace '[bundle pathForResource:@"ggml-metal" ofType:@"metal"];' "@\"$out/bin/ggml-metal.metal\";"
|
|
substituteInPlace ./ggml/src/ggml-metal.m \
|
|
--replace '[bundle pathForResource:@"default" ofType:@"metallib"];' "@\"$out/bin/default.metallib\";"
|
|
'';
|
|
|
|
# With PR#6015 https://github.com/ggerganov/llama.cpp/pull/6015,
|
|
# `default.metallib` may be compiled with Metal compiler from XCode
|
|
# and we need to escape sandbox on MacOS to access Metal compiler.
|
|
# `xcrun` is used find the path of the Metal compiler, which is varible
|
|
# and not on $PATH
|
|
# see https://github.com/ggerganov/llama.cpp/pull/6118 for discussion
|
|
__noChroot = effectiveStdenv.isDarwin && useMetalKit && precompileMetalShaders;
|
|
|
|
nativeBuildInputs =
|
|
[
|
|
cmake
|
|
ninja
|
|
pkg-config
|
|
git
|
|
]
|
|
++ optionals useCuda [
|
|
cudaPackages.cuda_nvcc
|
|
|
|
# TODO: Replace with autoAddDriverRunpath
|
|
# once https://github.com/NixOS/nixpkgs/pull/275241 has been merged
|
|
cudaPackages.autoAddOpenGLRunpathHook
|
|
]
|
|
++ optionals (effectiveStdenv.hostPlatform.isGnu && enableStatic) [
|
|
glibc.static
|
|
] ++ optionals (effectiveStdenv.isDarwin && useMetalKit && precompileMetalShaders) [
|
|
xcrunHost
|
|
];
|
|
|
|
buildInputs =
|
|
optionals effectiveStdenv.isDarwin darwinBuildInputs
|
|
++ optionals useCuda cudaBuildInputs
|
|
++ optionals useMpi [ mpi ]
|
|
++ optionals useRocm rocmBuildInputs
|
|
++ optionals useBlas [ blas ]
|
|
++ optionals useVulkan vulkanBuildInputs
|
|
++ optionals enableCurl [ curl ];
|
|
|
|
cmakeFlags =
|
|
[
|
|
(cmakeBool "LLAMA_BUILD_SERVER" true)
|
|
(cmakeBool "BUILD_SHARED_LIBS" (!enableStatic))
|
|
(cmakeBool "CMAKE_SKIP_BUILD_RPATH" true)
|
|
(cmakeBool "LLAMA_CURL" enableCurl)
|
|
(cmakeBool "GGML_NATIVE" false)
|
|
(cmakeBool "GGML_BLAS" useBlas)
|
|
(cmakeBool "GGML_CUDA" useCuda)
|
|
(cmakeBool "GGML_HIPBLAS" useRocm)
|
|
(cmakeBool "GGML_METAL" useMetalKit)
|
|
(cmakeBool "GGML_VULKAN" useVulkan)
|
|
(cmakeBool "GGML_STATIC" enableStatic)
|
|
]
|
|
++ optionals useCuda [
|
|
(
|
|
with cudaPackages.flags;
|
|
cmakeFeature "CMAKE_CUDA_ARCHITECTURES" (
|
|
builtins.concatStringsSep ";" (map dropDot cudaCapabilities)
|
|
)
|
|
)
|
|
]
|
|
++ optionals useRocm [
|
|
(cmakeFeature "CMAKE_HIP_COMPILER" "${rocmPackages.llvm.clang}/bin/clang")
|
|
(cmakeFeature "CMAKE_HIP_ARCHITECTURES" (builtins.concatStringsSep ";" rocmPackages.clr.gpuTargets))
|
|
]
|
|
++ optionals useMetalKit [
|
|
(lib.cmakeFeature "CMAKE_C_FLAGS" "-D__ARM_FEATURE_DOTPROD=1")
|
|
(cmakeBool "GGML_METAL_EMBED_LIBRARY" (!precompileMetalShaders))
|
|
];
|
|
|
|
# Environment variables needed for ROCm
|
|
env = optionals useRocm {
|
|
ROCM_PATH = "${rocmPackages.clr}";
|
|
HIP_DEVICE_LIB_PATH = "${rocmPackages.rocm-device-libs}/amdgcn/bitcode";
|
|
};
|
|
|
|
# TODO(SomeoneSerge): It's better to add proper install targets at the CMake level,
|
|
# if they haven't been added yet.
|
|
postInstall = ''
|
|
mkdir -p $out/include
|
|
cp $src/include/llama.h $out/include/
|
|
'';
|
|
|
|
# Define the shells here, but don't add in the inputsFrom to avoid recursion.
|
|
passthru = {
|
|
inherit
|
|
useBlas
|
|
useCuda
|
|
useMetalKit
|
|
useMpi
|
|
useRocm
|
|
useVulkan
|
|
;
|
|
|
|
shell = mkShell {
|
|
name = "shell-${finalAttrs.finalPackage.name}";
|
|
description = "contains numpy and sentencepiece";
|
|
buildInputs = [ llama-python ];
|
|
inputsFrom = [ finalAttrs.finalPackage ];
|
|
shellHook = ''
|
|
addToSearchPath "LD_LIBRARY_PATH" "${lib.getLib effectiveStdenv.cc.cc}/lib"
|
|
'';
|
|
};
|
|
|
|
shell-extra = mkShell {
|
|
name = "shell-extra-${finalAttrs.finalPackage.name}";
|
|
description = "contains numpy, sentencepiece, torchWithoutCuda, and transformers";
|
|
buildInputs = [ llama-python-extra ];
|
|
inputsFrom = [ finalAttrs.finalPackage ];
|
|
};
|
|
};
|
|
|
|
meta = {
|
|
# Configurations we don't want even the CI to evaluate. Results in the
|
|
# "unsupported platform" messages. This is mostly a no-op, because
|
|
# cudaPackages would've refused to evaluate anyway.
|
|
badPlatforms = optionals useCuda lib.platforms.darwin;
|
|
|
|
# Configurations that are known to result in build failures. Can be
|
|
# overridden by importing Nixpkgs with `allowBroken = true`.
|
|
broken = (useMetalKit && !effectiveStdenv.isDarwin);
|
|
|
|
description = "Inference of LLaMA model in pure C/C++${descriptionSuffix}";
|
|
homepage = "https://github.com/ggerganov/llama.cpp/";
|
|
license = lib.licenses.mit;
|
|
|
|
# Accommodates `nix run` and `lib.getExe`
|
|
mainProgram = "llama-cli";
|
|
|
|
# These people might respond, on the best effort basis, if you ping them
|
|
# in case of Nix-specific regressions or for reviewing Nix-specific PRs.
|
|
# Consider adding yourself to this list if you want to ensure this flake
|
|
# stays maintained and you're willing to invest your time. Do not add
|
|
# other people without their consent. Consider removing people after
|
|
# they've been unreachable for long periods of time.
|
|
|
|
# Note that lib.maintainers is defined in Nixpkgs, but you may just add
|
|
# an attrset following the same format as in
|
|
# https://github.com/NixOS/nixpkgs/blob/f36a80e54da29775c78d7eff0e628c2b4e34d1d7/maintainers/maintainer-list.nix
|
|
maintainers = with lib.maintainers; [
|
|
philiptaron
|
|
SomeoneSerge
|
|
];
|
|
|
|
# Extend `badPlatforms` instead
|
|
platforms = lib.platforms.all;
|
|
};
|
|
}
|
|
)
|