From ecc1ae01a415b2930181e35a69a43a056b263159 Mon Sep 17 00:00:00 2001 From: Gaetan Lepage Date: Sun, 8 Dec 2024 10:46:12 +0100 Subject: python312Packages.llama-cpp-python: 0.3.1 -> 0.3.2 Diff: https://github.com/abetlen/llama-cpp-python/compare/refs/tags/v0.3.1...v0.3.2 --- .../python-modules/llama-cpp-python/default.nix | 76 +++++++++++++--------- 1 file changed, 46 insertions(+), 30 deletions(-) (limited to 'pkgs/development/python-modules/llama-cpp-python') diff --git a/pkgs/development/python-modules/llama-cpp-python/default.nix b/pkgs/development/python-modules/llama-cpp-python/default.nix index 4c0391858e4a..579e91866019 100644 --- a/pkgs/development/python-modules/llama-cpp-python/default.nix +++ b/pkgs/development/python-modules/llama-cpp-python/default.nix @@ -2,47 +2,53 @@ lib, stdenv, buildPythonPackage, - cmake, fetchFromGitHub, - gitUpdater, + + # nativeBuildInputs + cmake, ninja, + + # build-system pathspec, pyproject-metadata, - pytestCheckHook, - pythonOlder, scikit-build-core, - llama-cpp-python, - config, - cudaSupport ? config.cudaSupport, - cudaPackages ? { }, + # buildInputs + apple-sdk_11, + # dependencies diskcache, jinja2, numpy, typing-extensions, + + # tests scipy, huggingface-hub, + + # passthru + gitUpdater, + pytestCheckHook, + llama-cpp-python, + + config, + cudaSupport ? config.cudaSupport, + cudaPackages ? { }, + }: -let - version = "0.3.1"; -in -buildPythonPackage { +buildPythonPackage rec { pname = "llama-cpp-python"; - inherit version; + version = "0.3.2"; pyproject = true; - disabled = pythonOlder "3.7"; - - stdenv = if cudaSupport then cudaPackages.backendStdenv else stdenv; - src = fetchFromGitHub { owner = "abetlen"; repo = "llama-cpp-python"; - rev = "refs/tags/v${version}"; - hash = "sha256-eO1zvNJZBE5BCnbgbh00tFIRWBCWor1lIsrLXs/HFds="; + tag = "v${version}"; + hash = "sha256-RtM5acaflUkPOZ/VS3afcMLbsnu6z7D8nXqpN8SR7v8="; fetchSubmodules = true; }; + # src = /home/gaetan/llama-cpp-python; dontUseCmakeConfigure = true; SKBUILD_CMAKE_ARGS = lib.strings.concatStringsSep ";" ( @@ -56,21 +62,28 @@ buildPythonPackage { nativeBuildInputs = [ cmake ninja + ]; + + build-system = [ pathspec pyproject-metadata scikit-build-core ]; - buildInputs = lib.optionals cudaSupport ( - with cudaPackages; - [ - cuda_cudart # cuda_runtime.h - cuda_cccl # - libcublas # cublas_v2.h - ] - ); + buildInputs = + lib.optionals cudaSupport ( + with cudaPackages; + [ + cuda_cudart # cuda_runtime.h + cuda_cccl # + libcublas # cublas_v2.h + ] + ) + ++ lib.optionals stdenv.hostPlatform.isDarwin [ + apple-sdk_11 + ]; - propagatedBuildInputs = [ + dependencies = [ diskcache jinja2 numpy @@ -91,12 +104,15 @@ buildPythonPackage { pythonImportsCheck = [ "llama_cpp" ]; - passthru.updateScript = gitUpdater { rev-prefix = "v"; }; - passthru.tests.llama-cpp-python = llama-cpp-python.override { cudaSupport = true; }; + passthru = { + updateScript = gitUpdater { rev-prefix = "v"; }; + tests.llama-cpp-python = llama-cpp-python.override { cudaSupport = true; }; + }; meta = { description = "Python bindings for llama.cpp"; homepage = "https://github.com/abetlen/llama-cpp-python"; + changelog = "https://github.com/abetlen/llama-cpp-python/blob/v${version}/CHANGELOG.md"; license = lib.licenses.mit; maintainers = with lib.maintainers; [ kirillrdy ]; }; -- cgit v1.2.3