mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-27 03:44:35 +00:00
1c641e6aac
* `main`/`server`: rename to `llama` / `llama-server` for consistency w/ homebrew
* server: update refs -> llama-server
gitignore llama-server
* server: simplify nix package
* main: update refs -> llama
fix examples/main ref
* main/server: fix targets
* update more names
* Update build.yml
* rm accidentally checked in bins
* update straggling refs
* Update .gitignore
* Update server-llm.sh
* main: target name -> llama-cli
* Prefix all example bins w/ llama-
* fix main refs
* rename {main->llama}-cmake-pkg binary
* prefix more cmake targets w/ llama-
* add/fix gbnf-validator subfolder to cmake
* sort cmake example subdirs
* rm bin files
* fix llama-lookup-* Makefile rules
* gitignore /llama-*
* rename Dockerfiles
* rename llama|main -> llama-cli; consistent RPM bin prefixes
* fix some missing -cli suffixes
* rename dockerfile w/ llama-cli
* rename(make): llama-baby-llama
* update dockerfile refs
* more llama-cli(.exe)
* fix test-eval-callback
* rename: llama-cli-cmake-pkg(.exe)
* address gbnf-validator unused fread warning (switched to C++ / ifstream)
* add two missing llama- prefixes
* Updating docs for eval-callback binary to use new `llama-` prefix.
* Updating a few lingering doc references for rename of main to llama-cli
* Updating `run-with-preset.py` to use new binary names.
Updating docs around `perplexity` binary rename.
* Updating documentation references for lookup-merge and export-lora
* Updating two small `main` references missed earlier in the finetune docs.
* Update apps.nix
* update grammar/README.md w/ new llama-* names
* update llama-rpc-server bin name + doc
* Revert "update llama-rpc-server bin name + doc"
This reverts commit e474ef1df4
.
* add hot topic notice to README.md
* Update README.md
* Update README.md
* rename gguf-split & quantize bins refs in **/tests.sh
---------
Co-authored-by: HanClinto <hanclinto@gmail.com>
183 lines
7.2 KiB
Nix
183 lines
7.2 KiB
Nix
# The flake interface to llama.cpp's Nix expressions. The flake is used as a
|
||
# more discoverable entry-point, as well as a way to pin the dependencies and
|
||
# expose default outputs, including the outputs built by the CI.
|
||
|
||
# For more serious applications involving some kind of customization you may
|
||
# want to consider consuming the overlay, or instantiating `llamaPackages`
|
||
# directly:
|
||
#
|
||
# ```nix
|
||
# pkgs.callPackage ${llama-cpp-root}/.devops/nix/scope.nix { }`
|
||
# ```
|
||
|
||
# Cf. https://jade.fyi/blog/flakes-arent-real/ for a more detailed exposition
|
||
# of the relation between Nix and the Nix Flakes.
|
||
{
|
||
description = "Port of Facebook's LLaMA model in C/C++";
|
||
|
||
inputs = {
|
||
nixpkgs.url = "github:NixOS/nixpkgs/nixos-unstable";
|
||
flake-parts.url = "github:hercules-ci/flake-parts";
|
||
};
|
||
|
||
# There's an optional binary cache available. The details are below, but they're commented out.
|
||
#
|
||
# Why? The terrible experience of being prompted to accept them on every single Nix command run.
|
||
# Plus, there are warnings shown about not being a trusted user on a default Nix install
|
||
# if you *do* say yes to the prompts.
|
||
#
|
||
# This experience makes having `nixConfig` in a flake a persistent UX problem.
|
||
#
|
||
# To make use of the binary cache, please add the relevant settings to your `nix.conf`.
|
||
# It's located at `/etc/nix/nix.conf` on non-NixOS systems. On NixOS, adjust the `nix.settings`
|
||
# option in your NixOS configuration to add `extra-substituters` and `extra-trusted-public-keys`,
|
||
# as shown below.
|
||
#
|
||
# ```
|
||
# nixConfig = {
|
||
# extra-substituters = [
|
||
# # Populated by the CI in ggerganov/llama.cpp
|
||
# "https://llama-cpp.cachix.org"
|
||
#
|
||
# # A development cache for nixpkgs imported with `config.cudaSupport = true`.
|
||
# # Populated by https://hercules-ci.com/github/SomeoneSerge/nixpkgs-cuda-ci.
|
||
# # This lets one skip building e.g. the CUDA-enabled openmpi.
|
||
# # TODO: Replace once nix-community obtains an official one.
|
||
# "https://cuda-maintainers.cachix.org"
|
||
# ];
|
||
#
|
||
# # Verify these are the same keys as published on
|
||
# # - https://app.cachix.org/cache/llama-cpp
|
||
# # - https://app.cachix.org/cache/cuda-maintainers
|
||
# extra-trusted-public-keys = [
|
||
# "llama-cpp.cachix.org-1:H75X+w83wUKTIPSO1KWy9ADUrzThyGs8P5tmAbkWhQc="
|
||
# "cuda-maintainers.cachix.org-1:0dq3bujKpuEPMCX6U4WylrUDZ9JyUG0VpVZa7CNfq5E="
|
||
# ];
|
||
# };
|
||
# ```
|
||
|
||
# For inspection, use `nix flake show github:ggerganov/llama.cpp` or the nix repl:
|
||
#
|
||
# ```bash
|
||
# ❯ nix repl
|
||
# nix-repl> :lf github:ggerganov/llama.cpp
|
||
# Added 13 variables.
|
||
# nix-repl> outputs.apps.x86_64-linux.quantize
|
||
# { program = "/nix/store/00000000000000000000000000000000-llama.cpp/bin/llama-quantize"; type = "app"; }
|
||
# ```
|
||
outputs =
|
||
{ self, flake-parts, ... }@inputs:
|
||
let
|
||
# We could include the git revisions in the package names but those would
|
||
# needlessly trigger rebuilds:
|
||
# llamaVersion = self.dirtyShortRev or self.shortRev;
|
||
|
||
# Nix already uses cryptographic hashes for versioning, so we'll just fix
|
||
# the fake semver for now:
|
||
llamaVersion = "0.0.0";
|
||
in
|
||
flake-parts.lib.mkFlake { inherit inputs; }
|
||
|
||
{
|
||
|
||
imports = [
|
||
.devops/nix/nixpkgs-instances.nix
|
||
.devops/nix/apps.nix
|
||
.devops/nix/devshells.nix
|
||
.devops/nix/jetson-support.nix
|
||
];
|
||
|
||
# An overlay can be used to have a more granular control over llama-cpp's
|
||
# dependencies and configuration, than that offered by the `.override`
|
||
# mechanism. Cf. https://nixos.org/manual/nixpkgs/stable/#chap-overlays.
|
||
#
|
||
# E.g. in a flake:
|
||
# ```
|
||
# { nixpkgs, llama-cpp, ... }:
|
||
# let pkgs = import nixpkgs {
|
||
# overlays = [ (llama-cpp.overlays.default) ];
|
||
# system = "aarch64-linux";
|
||
# config.allowUnfree = true;
|
||
# config.cudaSupport = true;
|
||
# config.cudaCapabilities = [ "7.2" ];
|
||
# config.cudaEnableForwardCompat = false;
|
||
# }; in {
|
||
# packages.aarch64-linux.llamaJetsonXavier = pkgs.llamaPackages.llama-cpp;
|
||
# }
|
||
# ```
|
||
#
|
||
# Cf. https://nixos.org/manual/nix/unstable/command-ref/new-cli/nix3-flake.html?highlight=flake#flake-format
|
||
flake.overlays.default = (
|
||
final: prev: {
|
||
llamaPackages = final.callPackage .devops/nix/scope.nix { inherit llamaVersion; };
|
||
inherit (final.llamaPackages) llama-cpp;
|
||
}
|
||
);
|
||
|
||
systems = [
|
||
"aarch64-darwin"
|
||
"aarch64-linux"
|
||
"x86_64-darwin" # x86_64-darwin isn't tested (and likely isn't relevant)
|
||
"x86_64-linux"
|
||
];
|
||
|
||
perSystem =
|
||
{
|
||
config,
|
||
lib,
|
||
system,
|
||
pkgs,
|
||
pkgsCuda,
|
||
pkgsRocm,
|
||
...
|
||
}:
|
||
{
|
||
# For standardised reproducible formatting with `nix fmt`
|
||
formatter = pkgs.nixfmt-rfc-style;
|
||
|
||
# Unlike `.#packages`, legacyPackages may contain values of
|
||
# arbitrary types (including nested attrsets) and may even throw
|
||
# exceptions. This attribute isn't recursed into by `nix flake
|
||
# show` either.
|
||
#
|
||
# You can add arbitrary scripts to `.devops/nix/scope.nix` and
|
||
# access them as `nix build .#llamaPackages.${scriptName}` using
|
||
# the same path you would with an overlay.
|
||
legacyPackages = {
|
||
llamaPackages = pkgs.callPackage .devops/nix/scope.nix { inherit llamaVersion; };
|
||
llamaPackagesWindows = pkgs.pkgsCross.mingwW64.callPackage .devops/nix/scope.nix { inherit llamaVersion; };
|
||
llamaPackagesCuda = pkgsCuda.callPackage .devops/nix/scope.nix { inherit llamaVersion; };
|
||
llamaPackagesRocm = pkgsRocm.callPackage .devops/nix/scope.nix { inherit llamaVersion; };
|
||
};
|
||
|
||
# We don't use the overlay here so as to avoid making too many instances of nixpkgs,
|
||
# cf. https://zimbatm.com/notes/1000-instances-of-nixpkgs
|
||
packages =
|
||
{
|
||
default = config.legacyPackages.llamaPackages.llama-cpp;
|
||
vulkan = config.packages.default.override { useVulkan = true; };
|
||
windows = config.legacyPackages.llamaPackagesWindows.llama-cpp;
|
||
}
|
||
// lib.optionalAttrs pkgs.stdenv.isLinux {
|
||
cuda = config.legacyPackages.llamaPackagesCuda.llama-cpp;
|
||
|
||
mpi-cpu = config.packages.default.override { useMpi = true; };
|
||
mpi-cuda = config.packages.default.override { useMpi = true; };
|
||
}
|
||
// lib.optionalAttrs (system == "x86_64-linux") {
|
||
rocm = config.legacyPackages.llamaPackagesRocm.llama-cpp;
|
||
};
|
||
|
||
# Packages exposed in `.#checks` will be built by the CI and by
|
||
# `nix flake check`.
|
||
#
|
||
# We could test all outputs e.g. as `checks = confg.packages`.
|
||
#
|
||
# TODO: Build more once https://github.com/ggerganov/llama.cpp/issues/6346 has been addressed
|
||
checks = {
|
||
inherit (config.packages) default vulkan;
|
||
};
|
||
};
|
||
};
|
||
}
|